Microsoft Fabric Data Engineer Associate Practice Test (DP-700)
Use the form below to configure your Microsoft Fabric Data Engineer Associate Practice Test (DP-700). The practice test can be configured to only include certain exam objectives and domains. You can choose between 5-100 questions and set a time limit.

Microsoft Fabric Data Engineer Associate DP-700 Information
The Microsoft Fabric Data Engineer Associate (DP-700) exam shows that you know how to work with data in Microsoft Fabric. It tests your ability to collect, organize, and prepare data so it can be used for reports and dashboards. Passing the DP-700 means you can build and manage data pipelines, use tools like Power BI and Azure Synapse, and make sure data is clean and ready for analysis.
This exam is best for people who already have some experience working with data or databases and want to move into a data engineering role. If you enjoy working with numbers, building reports, or using SQL and Python to manage data, this certification can help you stand out to employers. It’s designed for anyone who wants to show their skills in data handling using Microsoft tools.
Before taking the real exam, it’s smart to use DP-700 practice exams, practice tests, and practice questions to prepare. These tools help you get used to the types of questions you’ll see on test day and show which topics you need to study more. By using practice tests often, you can build confidence, improve your score, and walk into the exam knowing what to expect.

Free Microsoft Fabric Data Engineer Associate DP-700 Practice Test
- 20 Questions
- Unlimited
- Implement and manage an analytics solutionIngest and transform dataMonitor and optimize an analytics solution
Your organization uses a Microsoft Fabric data warehouse named ContosoDW. Members of the EuropeanFinance security group must query data in the dbo.Sales table, but:
- They should see only rows where Region = 'EU'.
- They must not be able to read the UnitCost column, while all other columns remain visible.
You need to configure ContosoDW to meet both requirements by using built-in capabilities of Fabric.
Which approach should you implement?
Move the UnitCost column to a separate table protected by workspace item permissions, and use pipeline parameters to pass the required Region value at query time.
Apply dynamic data masking to the UnitCost column and assign the EuropeanFinance group the Viewer role at the workspace level.
Create a view that selects all columns except UnitCost, and give EuropeanFinance SELECT permissions only on the view; do not configure any RLS because workspace roles will enforce the row filter automatically.
Create a table-valued function that filters on Region = 'EU', add it to a row-level security (RLS) security policy for dbo.Sales, and grant SELECT on the allowed columns only, leaving UnitCost ungranted.
Answer Description
In a Microsoft Fabric data warehouse, row-level security is enforced by creating a filter predicate function and binding it to a security policy. The policy filters the rows returned to a user according to logic you define, for example returning only rows where Region = 'EU' for members of the EuropeanFinance group.
To hide a single sensitive column while letting users read the remaining columns, grant SELECT permissions only on the allowed columns and withhold (or explicitly deny) SELECT on the UnitCost column. Fabric's SQL engine supports column-level permissions identical to those in Azure Synapse Analytics and SQL Server. Dynamic data masking would still reveal masked values and does not satisfy a 'no access' requirement, and workspace-level roles or item-level permissions cannot express per-row filters or per-column grants. Therefore, creating a row-level security policy together with explicit column-level SELECT grants is the correct solution.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Row-Level Security (RLS) in Microsoft Fabric?
How does column-level SELECT permission work?
Why is dynamic data masking insufficient for 'no access' requirements?
A Microsoft Fabric workspace contains a dataflow Gen2 that ingests daily CSV files from an Azure Data Lake Storage Gen2 container into a lakehouse table. After a recent refresh, you must verify how many rows were written by the dataflow and whether any source files failed to load, without opening the lakehouse itself. Which Fabric feature should you use to view this information?
The Monitor hub's Dataflows page and its refresh history for the dataflow
The Fabric Capacity Metrics app in Power BI
The workspace's Lineage view for the lakehouse
OneLake Data Explorer for the lakehouse's Files section
Answer Description
The Monitor hub surfaces operational details for Fabric items. Selecting the Dataflows tab and opening the most recent refresh run shows metrics such as total rows written, records processed per entity, file-level status, and any errors encountered. Lineage view only shows dependencies, not run metrics. The Capacity Metrics app provides tenant-level resource consumption, not per-refresh row counts. OneLake Data Explorer lists files in storage but does not display ingestion statistics.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is the Monitor hub in Microsoft Fabric?
How does Azure Data Lake Storage Gen2 work with Microsoft Fabric?
What is a dataflow Gen2 in Microsoft Fabric?
You are developing a PySpark notebook in Microsoft Fabric that joins a 2-TB fact table to three dimension tables, each about 100 MB. Execution metrics show most time is spent on shuffle reads during the joins. Without resizing the Spark pool, you want the dimension tables broadcast to executors to cut shuffle time. Which Spark configuration should you set before running the notebook?
Lower spark.sql.shuffle.partitions to 50 to reduce the number of shuffle partitions.
Enable adaptive query execution by setting spark.sql.adaptive.enabled to true.
Increase the value of spark.sql.autoBroadcastJoinThreshold to 134217728 (128 MB).
Set spark.sql.files.maxPartitionBytes to 134217728 bytes so that fewer input partitions are created.
Answer Description
Spark can avoid expensive shuffle joins by broadcasting small tables to every executor, turning the operation into a more efficient map-side join. Spark decides to broadcast a table only if its size is below the value specified in the spark.sql.autoBroadcastJoinThreshold configuration setting, whose default is 10 MB. Because each dimension table is about 100 MB, you must raise this threshold-setting it to 134,217,728 bytes (128 MB) will allow the three 100 MB dimension tables to qualify for automatic broadcasting and eliminate the large shuffle. Adjusting spark.sql.shuffle.partitions merely changes the number of shuffle partitions and does not enforce broadcasting. Enabling adaptive query execution can improve some plans but will not override the broadcast threshold by itself. Changing spark.sql.files.maxPartitionBytes affects input file partitioning, not join strategy. Therefore, increasing spark.sql.autoBroadcastJoinThreshold is the correct action.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
ELI5: What is spark.sql.autoBroadcastJoinThreshold?
Why does broadcasting reduce shuffle time in Spark joins?
What is the difference between `spark.sql.shuffle.partitions` and `spark.sql.autoBroadcastJoinThreshold`?
You are designing a lakehouse in Microsoft Fabric. To avoid duplicating data files that already reside in an Azure Data Lake Storage Gen2 account managed by another team, you decide to surface the folder /raw/finance in your lakehouse by using a shortcut. After the shortcut is created, analytics engineers will load the data with Spark, but the source team wants to guarantee that their files cannot be modified from your workspace.
Which statement about Fabric shortcuts satisfies the source team's requirement?
Write access is controlled by the lakehouse item role assignment, not by the shortcut type, so you must remove the Engineer role to prevent changes.
Shortcuts that reference Azure Data Lake Storage Gen2 are read-only, so Spark sessions in the lakehouse can read the files but cannot write or delete them.
Any shortcut becomes writable once the workspace owner is granted Contributor rights on the target storage account.
Fabric automatically creates a versioned copy of the target folder; engineers write to the copy while the original files stay untouched.
Answer Description
Shortcuts that point to external storage such as Azure Data Lake Storage Gen2 are mounted in Microsoft Fabric as read-only paths. Spark users in the destination lakehouse can read the data through the shortcut, but any attempt to create, update, or delete files under the shortcut will fail because Fabric blocks write operations to external locations. Only shortcuts that target another OneLake location are writable. Therefore, choosing an external shortcut automatically guarantees that the finance team's files remain unchanged, whereas the other options either describe writable shortcuts, unsupported behaviors, or unrelated features.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a shortcut in Microsoft Fabric?
Why are shortcuts to external storage read-only?
How does Spark interact with shortcuts in Microsoft Fabric?
Your organization uses Microsoft Purview sensitivity labels with Microsoft Fabric. You administer a workspace where engineers frequently create new lakehouses, dataflows, and reports. Compliance requires every new item in the workspace to be labeled "Contoso Confidential" automatically, but creators must still have the option to choose a different label if needed. Which workspace-level information-protection setting should you configure to meet this requirement?
Set a default sensitivity label of Contoso Confidential for the workspace and leave the "Users can change the applied label" option enabled.
Enable the "Inherit sensitivity label from data sources" option so new items copy labels from connected data.
Disable the workspace's default label and rely on tenant-level automatic labeling policies to tag items based on data content.
Enable the "Require users to apply a sensitivity label" option and make Contoso Confidential the only selectable label.
Answer Description
Applying a default sensitivity label at the workspace level causes any newly created or uploaded item to inherit that label automatically, satisfying the requirement that every new item start as Contoso Confidential. Because the default-label feature can be enabled without also enforcing mandatory labeling, authors remain free to replace the default with a different label.
Enabling a mandatory label would not automatically stamp the item; it only blocks saving until a user chooses a label, so the requirement for automatic assignment is unmet. Enabling inheritance from source data applies only when content is created from existing labeled data, not when authors create brand-new items. Relying solely on tenant-level auto-labeling policies does not guarantee that every new workspace item immediately receives the required label, and it ignores the need for authors to change the label on demand. Therefore, setting a default sensitivity label while allowing overrides is the correct choice.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What are Microsoft Purview sensitivity labels?
What does a default sensitivity label do in Microsoft Fabric?
How does inheritance of sensitivity labels from data sources work?
You manage a Microsoft Fabric workspace that uses the built-in deployment pipeline with the default Development, Test, and Production stages. After several iterations, the Development stage contains updated lakehouse tables, a modified data warehouse schema, and new notebooks. You must deploy only the metadata changes to the Test stage without copying any of the underlying data files. Which action should you take before you select Deploy on the Development stage?
Select Deploy; no additional configuration is required because deployment pipelines move only metadata for lakehouses and warehouses by default.
Export the lakehouse and warehouse as .pbip project files and remove their data folders before deployment.
Disconnect the workspace from its Git repository to ensure only metadata is copied when deploying.
Enable data skipping for the lakehouse and warehouse in the Development stage so that data files are excluded during deployment.
Answer Description
When you run a deployment pipeline, Fabric automatically copies the definitions (metadata) of supported items-such as lakehouses, warehouses, and notebooks-from the source stage to the target stage. However, for storage-based items (lakehouses and warehouses), the data itself is not migrated; only the metadata is deployed. Therefore, no special action is needed to prevent data movement-simply selecting Deploy from Development to Test will promote schema and object definitions only, leaving the data in the Test stage unchanged. Snapshot settings, manual data exports, or disabling Git integration do not affect this behavior.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is metadata in the context of Microsoft Fabric deployment?
How does Microsoft Fabric ensure that data is not copied during deployment?
What types of items are supported for deployment in Microsoft Fabric pipelines?
You manage a Microsoft Fabric warehouse that stores sales transactions in a table named Sales.Cards. The table has a varchar(16) column named CardNumber that contains customers' primary account numbers (PANs).
To limit data exposure, you execute the following T-SQL statement:
ALTER TABLE Sales.Cards
ALTER COLUMN CardNumber
ADD MASKED WITH (FUNCTION = 'partial(0,"XXXX-XXXX-XXXX-",4)');
Later, an analyst who has SELECT permission on the table-but does not hold the UNMASK permission-runs the query SELECT TOP 1 CardNumber FROM Sales.Cards;.
Which value will the analyst see if the stored PAN in that row is '1234567812345678'?
5678-XXXX-XXXX-XXXX
XXXX-XXXX-XXXX-5678
0000000000000000
XXXX-XXXX-5678-XXXX
Answer Description
The partial() dynamic data-masking function reveals a specified number of prefix and suffix characters and substitutes the remaining middle characters with a custom padding string. In this case the parameters are:
- Prefix to expose: 0 characters
- Padding string: "XXXX-XXXX-XXXX-"
- Suffix to expose: 4 characters
Therefore, all 16 digits except the final four are replaced by the literal text "XXXX-XXXX-XXXX-". Because the analyst lacks the UNMASK permission, the mask is applied and the query returns:
XXXX-XXXX-XXXX-5678
The other options do not match the behavior of the specified mask: the default() mask would return a string of zeros, while rearranged or partially hidden values are not produced by the given parameters.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Dynamic Data Masking in SQL?
What does the 'UNMASK' permission allow in SQL Server?
How does the 'partial()' masking function work?
You administer a Microsoft Fabric workspace that contains a lakehouse with several Delta tables used by multiple analytics teams. You must ensure that data analysts can query the SalesOrders table but must never see rows where the column TerritoryID equals 60. Business analysts should have unrestricted access. You need to configure this requirement in the most granular and performant way without duplicating data or creating additional views. Which approach should you implement in the lakehouse?
Publish a Power BI semantic model on top of SalesOrders and use that model's row-level security instead of configuring security in the lakehouse.
Grant the data-analysts group the Viewer role on the lakehouse and deny them access to the TerritoryID column by defining column-level security.
Move SalesOrders to a separate folder secured with OneLake folder permissions that exclude the data-analysts group.
Create a row-level security policy on the SalesOrders Delta table that filters out rows with TerritoryID = 60 for the data-analysts security group.
Answer Description
In a lakehouse, Delta tables are queried through Apache Spark or the SQL analytics endpoint. The most granular way to hide specific rows from a subset of users-without copying data or creating separate views-is to create a Row-Level Security (RLS) policy on the Delta table. RLS filters rows at query time based on predicates tied to user identities, so data analysts querying SalesOrders will automatically be prevented from seeing rows where TerritoryID = 60. Workspace roles alone cannot filter individual rows because they apply permissions at the item (object) level only. Column-level security targets specific columns, not rows. Converting the table to a secured Power BI semantic model would introduce an additional object and require separate maintenance, which is unnecessary when RLS can be applied directly on the Delta table.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Row-Level Security (RLS) in a lakehouse?
How do Delta tables support RLS in Microsoft Fabric?
Why can’t column-level security or folder permissions meet the requirement?
You manage a Microsoft Fabric workspace that hosts several lakehouses, pipelines, and Power BI reports. A group of data scientists needs to upload new datasets, create notebooks, and schedule pipeline runs in the workspace. However, they must not be able to change workspace permissions, delete the workspace, or publish the workspace as an app to other users. Which built-in workspace role should you assign to the data scientists to meet these requirements?
Viewer
Admin
Contributor
Member
Answer Description
The Contributor role is designed for users who need to create, edit, and delete content such as datasets, notebooks, lakehouses, and pipelines within a Microsoft Fabric (Power BI) workspace, including setting refresh schedules. Contributors cannot manage workspace permissions, delete the workspace itself, or publish or update an app, which aligns with the restrictions placed on the data scientists.
The Admin role is incorrect because it allows full control, including updating workspace permissions and deleting the workspace.
The Member role is incorrect because, in addition to content creation rights, Members can also add or remove users and publish or update the workspace app-capabilities you want to restrict.
The Viewer role is incorrect because it provides read-only access and does not allow creating or modifying any content.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is the difference between Contributor and Member roles in a Microsoft Fabric workspace?
What is a Microsoft Fabric workspace?
What are the limitations of the Viewer role in Microsoft Fabric workspace?
A semantic model named SalesModel refreshes hourly in Microsoft Fabric. The 02:00 refresh failed, and you must examine the step-by-step execution, including the processing time of each incremental-refresh partition, to locate the failure point. In the Fabric UI, which area should you open first to view this detailed refresh activity?
Open Data pipelines and select Semantic model refresh activity for SalesModel.
Use the Monitoring hub's Workspace Activity view for the workspace that hosts SalesModel.
Launch the Capacity Metrics app and filter by the 02:00 timeframe.
Open the SalesModel settings and view its Refresh history pane.
Answer Description
Detailed, step-level progress-such as the start, duration, and outcome of each partition processed during an incremental refresh-is surfaced in the Semantic model refresh activity page that opens from Data pipelines. The Refresh history pane on the model's settings page lists only a high-level status and overall duration, without partition-by-partition detail. The Capacity Metrics app focuses on capacity health, not individual model refresh steps, and the Monitoring hub's Workspace Activity view targets pipelines and dataflows rather than semantic model partitions.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a semantic model in Microsoft Fabric?
What is incremental refresh in the context of semantic models?
What does the Semantic model refresh activity page in Data pipelines show?
You build a Microsoft Fabric pipeline that copies new rows from an Azure SQL Database table into a destination table in a Fabric lakehouse. The Azure SQL linked service authenticates by using the workspace's managed identity. When you trigger the pipeline, the Copy activity fails immediately and returns error code SqlErrorNumber 262 with the message "Permission denied in database 'Sales' to object 'Orders'." What should you do to resolve the failure so that the pipeline can complete successfully?
Enable staged copy with auto-create table in the Copy activity settings.
Grant the workspace's managed identity SELECT permission on dbo.Orders or add it to the db_datareader role in the Sales database.
Grant the workspace's managed identity INSERT permission on dbo.Orders in the Sales database.
Replace the lakehouse sink with a Data Lake Storage Gen2 linked service.
Answer Description
SqlErrorNumber 262 in Azure SQL Database indicates that the caller lacks the required permission on the referenced object. Because the Copy activity only reads data from the dbo.Orders table, the workspace's managed identity must have SELECT rights (or broader read rights such as membership in the db_datareader role) on the Sales database. Granting INSERT permission, changing the sink, or enabling staged copy does not satisfy the missing authorization check on the source table, so the activity would still fail.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a managed identity in Azure?
What does the db_datareader role do in Azure SQL Database?
What does SqlErrorNumber 262 mean in Azure SQL Database?
You manage a Microsoft Fabric warehouse that contains a table named Customer with an nvarchar(256) column called EmailAddress. Data analysts query the warehouse by using the SQL Endpoint. Analysts must see e-mail addresses in the format j*****@contoso.com, but members of the FabricAdmins security group must see the full, unmasked values. You decide to use dynamic data masking. Which T-SQL actions should you perform to meet the requirements?
Create a view that returns LEFT(EmailAddress,1) + '*****' + SUBSTRING(EmailAddress, CHARINDEX('@', EmailAddress), LEN(EmailAddress)), and deny SELECT on dbo.Customer to the Analysts role.
Run: ALTER TABLE dbo.Customer ALTER COLUMN EmailAddress ADD MASKED WITH (FUNCTION = 'partial(1,"*****",12)'); GRANT UNMASK TO FabricAdmins;
Enable transparent data encryption on the warehouse and add FabricAdmins to the db_accessadmin database role.
Apply the Highly Confidential sensitivity label to EmailAddress and assign FabricAdmins as reviewers for the label.
Answer Description
Apply a dynamic data mask directly to the column and grant the UNMASK permission to the administrative group. The partial() masking function exposes the first character (prefix 1) and the last 12 characters (suffix 12 - the string "@contoso.com"), while padding the middle with the string *****. Granting UNMASK overrides the mask for authorized users.
Enabling transparent data encryption, applying a sensitivity label, or building a masking view do not implement dynamic data masking or satisfy the visibility requirement.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is dynamic data masking (DDM)?
How does the partial() function work in dynamic data masking?
What is the UNMASK permission, and how does it work?
You maintain a database project (.sqlproj) that contains the Data Definition Language (DDL) scripts for a Microsoft Fabric Warehouse. The project is stored in Azure Repos Git. You need to configure an Azure DevOps CI/CD pipeline that will automatically:
- Build the database project to produce a DACPAC file.
- Compare the DACPAC with the target Fabric Warehouse through its T-SQL endpoint.
- Apply only the required schema changes during deployment.
Which pipeline task or command should you add to perform the deployment to the Fabric Warehouse?
Add a Power BI deployment pipeline task that promotes the workspace to the test stage.
Add an Azure CLI task that calls the Microsoft Fabric REST API to import the DACPAC file as a dataflow.
Add a Copy Files task that copies all .sql script files to a blob container linked to the workspace.
Add an Azure SQL Database deployment task that runs the SqlPackage command with the Publish action against the warehouse's T-SQL endpoint.
Answer Description
The Azure SQL Database deployment task (or an equivalent Script task that runs the SqlPackage CLI) can build and publish a DACPAC. During the Publish action SqlPackage connects to the Fabric Warehouse by using the workspace's T-SQL endpoint, compares the DACPAC schema to the target, and then applies only the incremental changes that are required to bring the warehouse in line with the project. Tasks that copy raw .sql files or invoke Power BI or generic Azure CLI commands cannot generate a DACPAC or perform the schema comparison and publish operation that database projects rely on.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a DACPAC file?
What is the role of the T-SQL endpoint in this deployment?
How does SqlPackage facilitate incremental schema updates?
You must transform 200 GB of semi-structured logs in OneLake. Engineers need to use Python libraries like pandas and scikit-learn for feature engineering and visual analysis, then write the result to a Delta Lake table. The code should support iterative development and be runnable on a scheduled Fabric pipeline. Which Fabric transformation option best meets the requirements?
KQL database update policy
Spark notebook
Dataflow Gen2
T-SQL stored procedure in a Fabric warehouse
Answer Description
Spark notebooks in Microsoft Fabric run on Apache Spark, allowing PySpark and native Python packages such as pandas and scikit-learn for complex data preparation and machine-learning tasks. Notebooks provide an interactive environment for iterative development and visualization, can write directly to Delta Lake tables, and are first-class activities that can be scheduled within Fabric Data Factory pipelines. Dataflow Gen2 offers a low-code Power Query experience but cannot execute arbitrary Python or advanced ML libraries. KQL update policies are suited to streaming ingestion and Kusto databases, not large batch processing with external Python packages. T-SQL stored procedures in a warehouse are limited to SQL operations and do not support Python libraries or Spark-based scalability.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Apache Spark and how does it work in Microsoft Fabric?
What is Delta Lake, and why is it suitable for storing transformed data?
How do Fabric pipelines support Spark notebooks for scheduled execution?
You manage a Microsoft Fabric workspace that contains a data warehouse used by several analytics teams. The schema is expected to change frequently, and you must adopt a repeatable, source-controlled process that lets developers design tables offline, validate T-SQL syntax, and generate a deployment artifact that can be promoted by a pipeline to test and production workspaces. Which approach should you recommend to meet these requirements?
Author schema changes in Fabric notebooks and rely on automated retry logic in orchestration pipelines to apply them in downstream workspaces.
Create a SQL Server Database Project in Visual Studio with SQL Server Data Tools (SSDT) and publish directly from the IDE to each workspace.
Create a SQL database project in Visual Studio Code (or Azure Data Studio) by using the SQL Database Projects extension and commit the resulting .sqlproj and generated DACPAC to the Git repository.
Use the Export SQL scripts feature in the Fabric portal after every schema change and store the scripts in the Git repository for manual deployment.
Answer Description
A SQL database project created in Visual Studio Code (or Azure Data Studio) with the SQL Database Projects extension enables developers to design the Fabric warehouse schema offline, use the build process for T-SQL validation, and produce a DACPAC file. The .sqlproj and generated DACPAC can be checked in to Git and deployed to other Fabric workspaces through CI/CD pipelines, satisfying all requirements.
Although a Visual Studio SQL Server Database Project can also target Microsoft Fabric warehouses and generate a DACPAC, the option presented relies on publishing directly from the IDE, bypassing source control and automated pipelines, so it does not meet the requirement. Exporting T-SQL scripts from the Fabric portal or using notebooks lacks built-in validation, structured project artifacts, and automated deployment support, making them unsuitable for robust lifecycle management.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a DACPAC file used for?
What is the SQL Database Projects extension in Visual Studio Code?
How does CI/CD improve database deployment processes?
Your solution must ingest about one million IoT sensor events per second into Microsoft Fabric, auto-scale the compute, refresh operator dashboards in seconds, and let analysts run ad-hoc anomaly-detection queries by writing a familiar declarative language (not notebook code) against both real-time and historical data. Which Fabric streaming engine best meets these requirements?
T-SQL streaming ingestion into a lakehouse SQL endpoint
Eventstream with no downstream analytics engine
KQL streaming in Real-Time Analytics (Azure Data Explorer in Fabric)
Spark Structured Streaming in a Fabric notebook
Answer Description
KQL streaming in Microsoft Fabric's Real-Time Analytics is built on Azure Data Explorer technology. It can ingest well above one million events per second, automatically scales compute, stores the data in OneLake, and lets users run both real-time and historical queries with the same Kusto Query Language (KQL) without writing notebook code. Spark Structured Streaming would require writing code in notebooks, Eventstream alone does not provide a unified query surface for historical analysis, and T-SQL streaming into a lakehouse SQL endpoint cannot currently sustain multi-million-events-per-second ingestion.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is KQL and how is it used in Real-Time Analytics?
How does Microsoft Fabric's Real-Time Analytics handle auto-scaling of compute?
Why is Spark Structured Streaming not ideal for this scenario?
You manage a Microsoft Fabric workspace that contains a lakehouse used by finance analysts. The analysts frequently export tables to Excel and then email the files outside the organization. Management requests that every exported Excel file be automatically encrypted and marked "Confidential - Finance" without requiring the analysts to take any additional steps. Which action will meet this requirement with the current capabilities of Microsoft Fabric?
Set a default sensitivity label for the entire workspace so that all contained items inherit the label and its protection settings automatically.
Apply the "Confidential - Finance" Microsoft Purview sensitivity label directly to the lakehouse item in the Fabric workspace.
This requirement cannot be met because Fabric lakehouse exports do not yet support sensitivity-label-based encryption or content marking.
Enable Azure Information Protection on the OneLake storage account and configure an encryption policy that targets the finance folder.
Answer Description
Microsoft Fabric lakehouses do not presently support sensitivity-label inheritance or protection actions such as encryption and content marking for files exported from the lakehouse. These capabilities are currently limited to Power BI items in specific export scenarios. Because of this limitation, no configuration change in Fabric (including labeling the lakehouse or workspace, changing table properties, or enabling Azure Information Protection on the underlying OneLake storage) will automatically apply encryption or markings to Excel files generated from lakehouse data. Therefore, the requirement cannot be met with the product's current functionality.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Microsoft Fabric and its primary function?
What are sensitivity labels, and how do they work in Microsoft Purview?
What is OneLake storage, and how does Azure Information Protection integrate with it?
You manage a Microsoft Fabric workspace that contains a semantic model named Contoso Sales. The model is configured to refresh every hour. The data-quality team must receive an email whenever a scheduled refresh ends with a Failure status. You need a solution that can be configured directly in the Fabric (Power BI) service without building a custom monitoring pipeline. What should you do?
Pin the semantic model's refresh history to a dashboard and configure a data alert on the visual so that it emails the data-quality team when the value equals Failure.
Create an Azure Monitor metric alert that fires when the semantic model's refresh duration metric exceeds a threshold and add the distribution list to the action group.
Open the semantic model's Scheduled refresh settings, enable refresh failure notifications, and add the data-quality distribution list as email recipients.
Configure a Fabric capacity usage alert for the workspace and specify the distribution list as notification recipients.
Answer Description
In the Fabric (Power BI) service, each semantic model (dataset) has a Scheduled refresh settings pane. In that pane you can turn on refresh failure notifications and specify one or more email addresses in the Email recipients (also surfaced as Email these contacts or Failure notification recipients) box. When a scheduled refresh fails, the service automatically sends an alert email to every address listed. Creating Azure Monitor alerts or capacity-level alerts does not target the semantic model's refresh operation, and dashboard data alerts can only monitor numeric values rendered in a visual-refresh status is not exposed that way. Therefore, enabling the built-in refresh failure notification and adding the distribution list is the correct and simplest solution.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is the Fabric semantic model?
How do you configure scheduled refresh and failure notifications in Fabric?
Why are Azure Monitor or dashboard visual alerts inappropriate for refresh failure notifications?
You are building a daily load from an on-premises SQL Server table into a Delta Lake table in a Microsoft Fabric lakehouse. The source table contains hundreds of millions of rows and has an incrementing bigint column named UpdateSequence that increases each time a row is inserted or modified. Loads must copy only new and changed rows, overwrite any existing versions in the destination, and be fully restartable without creating duplicate rows. Which loading pattern meets these requirements?
Implement a watermark-based incremental load that filters rows where UpdateSequence is greater than the last stored watermark, then upsert the data into the Delta Lake table by using MERGE INTO.
Use an eventstream to capture change events and append them directly to the Delta Lake table without performing updates.
Schedule a nightly pipeline that truncates the Delta Lake table and performs a full bulk INSERT of the entire source table.
Configure two alternating pipelines: the first copies rows whose UpdateSequence modulo 2 equals 0, and the second copies the remaining rows the next day.
Answer Description
A watermark-based incremental load keeps track of the highest UpdateSequence value that was successfully loaded. During each run, the pipeline filters the source by selecting rows with an UpdateSequence greater than this watermark, minimizing the volume of data copied. The pipeline then executes a MERGE INTO statement on the Delta Lake table, which updates existing rows and inserts new ones in an idempotent manner. Because the watermark is updated only after the MERGE finishes, a failed run can be safely retried without introducing duplicates. Truncating and reloading the entire table wastes resources. Simply appending rows or dividing the table by an arbitrary modulo would not update changed records and could still duplicate data.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a watermark-based incremental load?
How does the MERGE INTO statement work in a Delta Lake?
How does using the UpdateSequence column make the process restartable?
You have created a notebook in a Microsoft Fabric workspace that accepts two parameters and transforms raw JSON files into a curated Delta table. You must ensure that the notebook runs automatically whenever a new JSON file is written to the raw container of an Azure Data Lake Storage Gen2 account, and only after the file has been copied to OneLake. You want to design the solution by using Fabric features while writing as little custom code as possible. Which approach should you use?
Create a SQL job in the Lakehouse that listens for CREATE FILE events and, when triggered, uses dynamic SQL to call the notebook through the Fabric REST API.
Configure a scheduled run for the notebook that executes every five minutes and add code to the notebook to poll the raw container for new files and copy them before processing.
Develop a Dataflow Gen2 that copies data from the raw container to OneLake and adds a script step at the end to invoke the notebook.
Create a Data Factory pipeline that uses an Azure Storage event trigger, adds a Copy activity to move the file to OneLake, and then calls the parameterized notebook in a subsequent Notebook activity.
Answer Description
Event-based automation in Microsoft Fabric is available through Data Factory pipelines, not through notebook schedules. A pipeline can be configured with a storage event trigger that fires when a blob is created in a specified ADLS Gen2 path. Inside the pipeline you add a Copy activity to move the blob into OneLake, and then a Notebook activity that executes the existing parameterized notebook. Notebook schedules can only run on a fixed time basis and cannot listen for storage events, while SQL jobs and dataflows cannot orchestrate notebook execution after the copy finishes. Therefore, using a pipeline with an event trigger best meets the requirements and minimises custom code.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a storage event trigger in Data Factory?
What is OneLake in Microsoft Fabric?
How does a Notebook activity work in a Data Factory pipeline?
Nice!
Looks like that's it! You can go back and review your answers or click the button below to grade your test.