Microsoft Fabric Data Engineer Associate Practice Test (DP-700)
Use the form below to configure your Microsoft Fabric Data Engineer Associate Practice Test (DP-700). The practice test can be configured to only include certain exam objectives and domains. You can choose between 5-100 questions and set a time limit.

Microsoft Fabric Data Engineer Associate DP-700 Information
The Microsoft Fabric Data Engineer Associate (DP-700) exam shows that you know how to work with data in Microsoft Fabric. It tests your ability to collect, organize, and prepare data so it can be used for reports and dashboards. Passing the DP-700 means you can build and manage data pipelines, use tools like Power BI and Azure Synapse, and make sure data is clean and ready for analysis.
This exam is best for people who already have some experience working with data or databases and want to move into a data engineering role. If you enjoy working with numbers, building reports, or using SQL and Python to manage data, this certification can help you stand out to employers. It’s designed for anyone who wants to show their skills in data handling using Microsoft tools.
Before taking the real exam, it’s smart to use DP-700 practice exams, practice tests, and practice questions to prepare. These tools help you get used to the types of questions you’ll see on test day and show which topics you need to study more. By using practice tests often, you can build confidence, improve your score, and walk into the exam knowing what to expect.

Free Microsoft Fabric Data Engineer Associate DP-700 Practice Test
- 20 Questions
- Unlimited
- Implement and manage an analytics solutionIngest and transform dataMonitor and optimize an analytics solution
Free Preview
This test is a free preview, no account required.
Subscribe to unlock all content, keep track of your scores, and access AI features!
You manage a Microsoft Fabric workspace that contains a lakehouse used by finance analysts. The analysts frequently export tables to Excel and then email the files outside the organization. Management requests that every exported Excel file be automatically encrypted and marked "Confidential - Finance" without requiring the analysts to take any additional steps. Which action will meet this requirement with the current capabilities of Microsoft Fabric?
Set a default sensitivity label for the entire workspace so that all contained items inherit the label and its protection settings automatically.
Enable Azure Information Protection on the OneLake storage account and configure an encryption policy that targets the finance folder.
Apply the "Confidential - Finance" Microsoft Purview sensitivity label directly to the lakehouse item in the Fabric workspace.
This requirement cannot be met because Fabric lakehouse exports do not yet support sensitivity-label-based encryption or content marking.
Answer Description
Microsoft Fabric lakehouses do not presently support sensitivity-label inheritance or protection actions such as encryption and content marking for files exported from the lakehouse. These capabilities are currently limited to Power BI items in specific export scenarios. Because of this limitation, no configuration change in Fabric (including labeling the lakehouse or workspace, changing table properties, or enabling Azure Information Protection on the underlying OneLake storage) will automatically apply encryption or markings to Excel files generated from lakehouse data. Therefore, the requirement cannot be met with the product's current functionality.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Microsoft Fabric and its primary function?
What are sensitivity labels, and how do they work in Microsoft Purview?
What is OneLake storage, and how does Azure Information Protection integrate with it?
During a scheduled refresh of a Fabric Dataflow Gen2, the operation fails with the message "Expression.Error: The column CustomerID of the table was not found." An upstream SQL view was recently modified and the column was renamed to CustID. You need to restore a successful refresh without recreating the dataflow. What should you do in the dataflow editor?
Recreate the gateway connection using basic authentication credentials.
Enable the dataflow's Enhanced Compute Engine and re-run the refresh.
Edit the Power Query steps, replace references to CustomerID with CustID in the Source or subsequent transformation steps, then save and refresh.
Increase the dataflow refresh timeout value in Settings and retry.
Answer Description
The error indicates that a transformation step still expects a column named CustomerID, but the source schema now exposes the column as CustID. Opening the dataflow in the Power Query editor and updating the Source (or subsequent) step to reference the new column name realigns the query with the current schema. Merely enabling performance features, extending time-outs, or changing gateway settings does not correct the missing column reference, so the refresh would continue to fail until the query steps are fixed.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Power Query and how does it relate to Fabric Dataflows?
How does schema changes in a source affect a Fabric Dataflow Gen2?
What is the role of the Enhanced Compute Engine in Fabric Dataflows?
CSV files land in Azure Blob Storage, exposed in OneLake via a shortcut. You must build a Fabric solution that automatically starts when each new file arrives, passes the file path to an existing PySpark notebook that cleans the data, and loads the output into a Fabric warehouse. The solution must include event triggers, parameterized notebook input, and built-in retry and alerting. Which Fabric component should you create?
Create a Dataflow Gen2 to transform the data and load it into the warehouse.
Create a Data Factory pipeline that contains an Execute Notebook activity followed by a Copy Data activity.
Create a SQL pipeline in the Fabric warehouse to run the PySpark logic and load the data.
Create a Spark notebook and schedule it to run on a frequent interval.
Answer Description
A Data Factory pipeline in Microsoft Fabric can attach a storage event trigger, pass the triggering file path as a parameter to an Execute Notebook activity, and follow it with a Copy Data activity to load the warehouse. Pipelines expose configurable retries and monitoring/alert options without additional custom code. Stand-alone Spark notebooks lack event subscriptions, SQL pipelines cannot run PySpark, and Dataflow Gen2 centers on Power Query rather than notebook orchestration.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Azure Blob Storage and OneLake?
What is a Data Factory pipeline and its role in this solution?
What is the purpose of the Execute Notebook activity in Data Factory pipelines?
Your team develops a Microsoft Fabric lakehouse solution. You must store the definitions of all SQL objects (tables, views, and stored procedures) in a Git repository and automate schema deployments to test and production workspaces by using Azure DevOps pipelines. Which Fabric workspace item should you create to meet these requirements?
Export the lakehouse tables as Delta files and commit the files to the repository.
Enable Git integration for the workspace without adding any additional artifacts.
Author a notebook that contains CREATE TABLE and ALTER statements and store it in Git.
Create a database project in the workspace and connect it to the Git repository.
Answer Description
A SQL database project is the workspace item intended for treating a lakehouse or warehouse schema as source code. After you create the project, you can script tables, views, and stored procedures into .sql files, commit them to Git, and use Azure DevOps or GitHub pipelines to compare environments and deploy incremental changes. Simply enabling Git integration does not extract object definitions, exporting Delta files captures data rather than schema, and notebooks have no built-in schema compare or deployment functionality.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a SQL database project in Microsoft Fabric?
How does Azure DevOps help in automating schema deployments?
Why is Git integration alone insufficient for managing schema in Microsoft Fabric?
In a Microsoft Fabric PySpark notebook, you have a DataFrame named df that contains incremental changes for the Customers dimension. You must write the data to the lakehouse path "Tables/dim_customer" so that it is stored in Delta format, automatically merges any new columns in future loads, and is physically partitioned by the Country column. Which PySpark write command meets all these requirements?
df.repartition("Country").write.format("delta").mode("append").save("Tables/dim_customer")
df.write.format("delta").mode("overwrite").option("overwriteSchema", "true").partitionBy("Country").save("Tables/dim_customer")
df.write.format("parquet").mode("append").option("mergeSchema", "true").partitionBy("Country").save("Tables/dim_customer")
df.write.format("delta").mode("append").option("mergeSchema", "true").partitionBy("Country").save("Tables/dim_customer")
Answer Description
To fulfill all requirements you must:
- Specify Delta format with .format("delta").
- Enable automatic schema evolution using .option("mergeSchema", "true").
- Append, not overwrite, by using .mode("append").
- Create physical partitions with .partitionBy("Country"). Only the correct choice includes all four elements in one statement; the other options miss Delta format, use overwriteSchema instead of mergeSchema, omit partitionBy, or overwrite existing data.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Delta format and how is it different from Parquet?
Why is `mode("append")` required for incremental data loading?
What does `partitionBy("Country")` do in the PySpark write command?
A stored procedure in a Microsoft Fabric warehouse runs this statement to upsert rows from StgSales into DimCustomer:
MERGE dbo.DimCustomer AS tgt
USING dbo.StgSales AS src
ON tgt.CustomerID = src.CustomerID
WHEN MATCHED THEN
UPDATE SET tgt.City = src.City, tgt.Region = src.Region
WHEN NOT MATCHED BY TARGET THEN
INSERT (CustomerID, City, Region)
VALUES (src.CustomerID, src.City, src.Region);
Execution fails with the error: "The MERGE statement attempted to UPDATE or DELETE the same row more than once. A target row matched more than one source row."
You must correct the T-SQL so the procedure succeeds while still performing the required updates and inserts.
Which change should you make to the statement?
Execute SET IDENTITY_INSERT dbo.DimCustomer ON immediately before running the MERGE.
Rewrite the USING clause to select DISTINCT CustomerID, City, Region from dbo.StgSales before the MERGE is executed.
Replace the MERGE with an INSERT statement that uses the ON ERROR clause to ignore conflicts.
Add the table hint WITH (NOLOCK) to dbo.StgSales in the USING clause.
Answer Description
The error appears because more than one row in the source can match a single row in the target on CustomerID, so the MERGE would update the same target row multiple times. Selecting only one row per CustomerID in the source eliminates the conflict. Adding DISTINCT (or similar deduplication logic) in the USING clause guarantees each target row can match at most one source row, allowing the MERGE to complete. Using WITH (NOLOCK) can change the set of rows returned and does not guarantee removal of duplicates; SET IDENTITY_INSERT affects identity columns and is unrelated to this error; and INSERT … ON ERROR is not valid T-SQL syntax.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a MERGE statement in T-SQL?
Why does adding DISTINCT resolve the error?
How does WITH (NOLOCK) affect the query in T-SQL?
You need to give analysts in Workspace A access to a set of parquet files that already reside in Workspace B's lakehouse without copying the data. You plan to create a shortcut in Workspace A that references the existing folder. Which statement about managing shortcuts in Microsoft Fabric meets the requirement?
A shortcut can only be created at the root of the lakehouse; it cannot be placed inside sub-folders.
A shortcut relies on the source folder's permissions, so analysts must already have read access to the folder in Workspace B.
You must schedule a pipeline refresh for the shortcut; otherwise, newly added files will not appear to analysts.
After you create the shortcut, renaming the source folder is automatically reflected and requires no additional action.
Answer Description
Shortcuts in Microsoft Fabric are metadata pointers to an existing location in OneLake. They do not duplicate the data and they rely on the underlying OneLake security model. If a user does not have permission on the source location, the shortcut does not grant it; access will still be denied. Moving or renaming the source path breaks the shortcut, and the data appears immediately without any refresh jobs because no materialization occurs. Shortcuts can be created inside any folder structure within a lakehouse.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a shortcut in Microsoft Fabric?
Can analysts use a shortcut if they do not have permissions to the source folder?
What happens if the source folder for a shortcut is renamed or moved?
You manage a Microsoft Fabric lakehouse that ingests micro-batches into a Delta table named Sales. After several months, the table contains thousands of very small data files, and analysts report that queries filtering on the OrderDate column now take much longer to finish. With minimal code changes, which action should you perform to most effectively improve scan and query performance on the Sales table?
Execute
OPTIMIZE Sales ZORDER BY (OrderDate);to compact small files and cluster rows on the filter column.Run
VACUUM Sales RETAIN 0 HOURS;to delete obsolete data files from the table.Use
COPY INTOto export the data to a single large Parquet file and replace the table.Increase the notebook session's driver memory to provide more Spark cache capacity during queries.
Answer Description
Running the Delta Lake OPTIMIZE command rewrites the table into a smaller number of large files, reducing metadata overhead and improving scan efficiency. Adding the ZORDER BY clause on a frequently filtered column such as OrderDate further clusters related rows in the same data pages, minimizing file-skipping and I/O during predicate pushdown. VACUUM only deletes obsolete files and does not reorganize current data. Converting to plain Parquet would remove Delta functionality and still leave many small files if singleFile cannot handle large data volumes. Increasing notebook driver memory may speed Spark jobs but does not address the root cause-file fragmentation in storage-so it yields little benefit for downstream queries executed by other engines.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Delta Lake's OPTIMIZE command?
What is ZORDER BY in Delta Lake?
Why doesn’t VACUUM improve query performance in this scenario?
You are designing a real-time telemetry solution in Microsoft Fabric. Business analysts must be able to query incoming sensor readings together with the last three months of historical data by using the same syntax that they currently use in Azure Data Explorer. The solution must offer built-in windowing, time-series functions, and automatic data retention without requiring you to write code. Which streaming engine should you choose?
Deploy an Azure Stream Analytics job that writes to a Fabric Data Warehouse.
Develop a Spark Structured Streaming notebook that writes the data to Delta tables in a lakehouse.
Use Eventstream only to capture the data to a lakehouse and query it later with SQL.
Create a KQL database in Real-Time Analytics and ingest the stream directly into a KQL table.
Answer Description
A KQL database in Real-Time Analytics uses the same Kusto Query Language that analysts already know from Azure Data Explorer. It can ingest high-throughput event streams, stores both hot and historical data in the same engine, provides built-in time-series and windowing functions, and lets you configure retention policies without writing code.
Spark Structured Streaming would require notebooks and Python/Scala code and lacks KQL's built-in analytical functions. Eventstream on its own is a data ingestion service; to query data you would still need to land it elsewhere and use another engine. Azure Stream Analytics jobs can output to Fabric but run outside the service and are managed separately, so they do not meet the requirement for a native Fabric engine with KQL querying.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Kusto Query Language (KQL)?
What is Real-Time Analytics in Microsoft Fabric?
How does windowing work in time-series data analysis?
You manage a Microsoft Fabric warehouse. A 4-TB sales fact table is currently stored using the default clustered columnstore index and a ROUND_ROBIN distribution. Most analytical queries join this table with a 15-GB Date dimension and filter on the calendar year. The queries scan many unnecessary rows and exhibit high data movement during joins. Without purchasing additional capacity, which change is most likely to reduce both scan and shuffle costs?
Replicate the fact table instead of distributing it so all compute nodes have a local copy.
Convert the fact table to a clustered rowstore index to improve predicate pushdown on DateKey.
Redistribute the fact table by using HASH on the DateKey column while keeping the clustered columnstore index.
Create a materialized view that filters the fact table to the current calendar year.
Answer Description
Changing the sales fact table to a hash distribution on the surrogate key that joins to the Date dimension will colocate matching rows on the same distribution, eliminating the shuffle that currently occurs with a ROUND_ROBIN table. Because clustered columnstore remains in place, segment elimination will still prune rowgroups for the filtered calendar years, further reducing the amount of data scanned. Simply switching to a clustered index or replicating a 4-TB table is impractical due to size, and materializing the yearly subset would not remove the data-movement overhead of the join.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a clustered columnstore index, and why is it the default for large tables?
What does HASH distribution mean, and how does it reduce shuffle costs in joins?
Why is ROUND_ROBIN distribution less effective for large-table joins?
You are building a Microsoft Fabric Eventstream to ingest sensor readings from Azure IoT Hub. Each event payload contains the fields deviceId (string), ts (epoch milliseconds), temperature (double), and humidity (double).
You must satisfy the following processing requirements:
- Guarantee that statistics are calculated by the timestamp in each event, even if messages arrive out of order.
- Discard any event that arrives more than 2 minutes after its ts value.
- Produce a running 1-minute tumbling-window average of the temperature for each device and store the result in a Real-Time Analytics KQL database table.
Which configuration should you apply to the Eventstream input or query to meet all the requirements?
Use the system column _arrivalTime for windowing, add a WHERE clause that filters events older than 2 minutes, and write results to the KQL table every minute.
Mark ts as the event-time column and use a 1-minute hopping window with a 30-second hop size; do not configure out-of-order tolerance because tumbling windows implicitly drop late data.
Leave event ordering at the default arrival time, and in the query declare a 1-minute session window on ts; set the session timeout to 2 minutes to ignore late events.
Mark ts as the event-time column on the IoT Hub input, set a 2-minute out-of-order tolerance with the late-arrival policy set to Drop, and in the query use a 1-minute tumbling window on ts with GROUP BY deviceId.
Answer Description
Designating ts as the event-time column tells Eventstream to sequence and process data by when the sensor reading actually occurred instead of by its arrival time. You then set a 2-minute out-of-order tolerance on ts and configure the late-arrival events policy to Drop so that any record whose ts timestamp is more than two minutes older than the current watermark is discarded. Once ordering and late-arrival handling are in place, a SQL query that applies a 1-minute tumbling window on ts and groups by deviceId can safely compute AVG(temperature) per device and route the results to a KQL table. Filtering with a WHERE clause or relying on arrival time would not guarantee correct ordering or late-event removal.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is an event-time column in the context of Microsoft Fabric Eventstream?
What is a tumbling window, and how does it work in real-time analytics?
What is the late-arrival policy in Microsoft Fabric Eventstream, and why is it important?
Queries that join the SalesFact and ReturnsFact tables in your Microsoft Fabric warehouse frequently spill into a high-shuffle data-movement step and run for several minutes. Both tables currently use ROUND_ROBIN distribution and each contains over 500 million rows. You must reduce data movement and accelerate the join without replicating either table. Which T-SQL action should you take?
Rebuild both tables as clustered columnstore indexes ordered by the join column.
Create nonclustered B-tree indexes on the join column in both tables.
Enable result-set caching for the warehouse.
Rebuild both tables by using HASH distribution on the shared join column.
Answer Description
Rebuilding both large fact tables with HASH distribution on the common join column causes rows that share the key value to be stored on the same distribution. When two hash-distributed tables are collocated on the same key, the engine can perform a shuffle-free local join, eliminating the costly data-movement step and reducing execution time. Result-set caching only helps when identical queries are rerun and does not remove data movement. B-tree indexes are ignored by the distributed query optimizer when it chooses a data-movement strategy. Ordering a clustered columnstore index may aid segment elimination on range predicates, but it has no effect on which compute node each row resides, so it cannot eliminate the shuffle.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is ROUND_ROBIN distribution in Microsoft Fabric?
What is HASH distribution, and why is it used?
Why are nonclustered B-tree indexes ignored by the distributed query optimizer?
Your organization uses Microsoft Fabric workspaces backed by a capacity in the F64 SKU. As a Fabric administrator, you need to set up a near real-time feed of all workspace-level activities-such as item creation, permission changes, and publish operations-so that your Security Operations Center (SOC) can query the data with Kusto Query Language (KQL) and build custom alerts. Which action should you perform first to meet this requirement?
Download the Power Platform admin audit log and schedule a notebook to upload the file to a Kusto database.
Assign the Log Analytics Contributor role to the SOC analysts on the Fabric workspace.
Create a deployment pipeline and enable its workspace usage metrics dataset for the SOC team.
Create an Azure Monitor diagnostic setting on the Fabric capacity or workspace and send the Fabric Activity log to a Log Analytics workspace.
Answer Description
To stream Fabric workspace activities into a target that supports ad-hoc KQL queries, you must configure an Azure Monitor diagnostic setting for the capacity (or individual workspace) and route the Fabric Activity log to Log Analytics. Enabling the diagnostic setting is the prerequisite that turns on continuous export; once the data reaches the Log Analytics workspace, the SOC team can query it with KQL or connect Sentinel for alerting. Downloading the admin audit log provides only a static file, not a streaming feed. Assigning the Log Analytics Contributor role does not itself activate logging. Creating a deployment pipeline does not generate or stream activity logs.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is an Azure Monitor diagnostic setting?
What is the Fabric Activity log?
What is Kusto Query Language (KQL)?
Your team must put a Microsoft Fabric workspace under source control with an Azure DevOps Git repository named FabricData, which already has main and develop branches. Requirements: workspace commits default to develop, JSON files stored in a folder matching the workspace name, and engineers must pull and push within the Fabric portal. Which first action should you take in the Fabric portal?
Enable workspace-level audit logging and specify the FabricData repository as the log storage location.
Connect the workspace to the FabricData repository in Workspace settings and set the default branch to develop with a folder path that matches the workspace name.
Change the default branch in Azure DevOps from main to develop and then clone the repository locally.
Create a deployment pipeline that uses the FabricData repository as the source environment and assigns the develop branch to the test stage.
Answer Description
In Microsoft Fabric, Git integration is configured at the workspace level. The correct first step is to open Workspace settings, choose Git integration, and connect the workspace to the FabricData repository. During this connection you specify Azure DevOps as the provider, set develop as the default branch, and enter a folder path that matches the workspace name. This immediately enables the Sync and Commit buttons so engineers can pull and push without leaving the portal. Creating a deployment pipeline, changing the default branch in Azure DevOps, or enabling audit logging do not establish the required Git link.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Git integration in Microsoft Fabric?
Why should the folder path match the workspace name when configuring Git integration?
What are the roles of the Sync and Commit buttons in Microsoft Fabric Git integration?
You manage a Microsoft Fabric workspace that contains a data warehouse named SalesDW. Your DevOps team wants every table, view, and stored procedure in the warehouse to be stored in a Git repository so that schema changes can be reviewed and deployed through an Azure DevOps pipeline to test and production workspaces. What is the most appropriate first step to create a deployable artifact that captures the current warehouse schema?
Use Visual Studio Code with the SQL Database Projects extension to import SalesDW and build a .dacpac file.
Export SalesDW from the Fabric portal as a Power BI project (.pbip) and push it to the Git repository.
Generate an Azure Resource Manager (ARM) template for the warehouse item from the Azure portal and store it in Git.
Execute a T-SQL BACKUP DATABASE command in a Fabric notebook and add the backup file to source control.
Answer Description
Microsoft Fabric data warehouses are compatible with SQL database projects. By using the SQL Database Projects extension in Visual Studio Code (or Visual Studio) you can import the existing SalesDW schema and automatically generate a .dacpac file. The .dacpac (Data-tier Application Component Package) is a portable artifact that represents the entire database schema and can be committed to a Git repository and deployed to other Fabric workspaces through CI/CD pipelines. Exporting a Power BI project captures reports and models, not warehouse objects; ARM template exports do not include Fabric warehouse schemas; and BACKUP DATABASE produces a backup, not a schema-focused deployment package suitable for source control and incremental deployments.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a .dacpac file, and why is it useful for databases?
How does the SQL Database Projects extension help in managing database schemas?
What are CI/CD pipelines, and how do they work in deploying database changes?
You are creating a new Microsoft Fabric workspace that will host several lakehouses for a sales analytics project. To comply with organizational policies, you must ensure that:
- All files written to OneLake in this workspace are stored in the West Europe region.
- Workspace consumers must be able to read and query data, but they must NOT be able to download the underlying parquet or CSV files from OneLake.
Which combination of settings should you configure in the OneLake section of the workspace settings to meet both requirements?
Enable OneLake shortcuts and set the Default storage location to West Europe.
Change the default file format to Delta and enable personal workspaces only.
Set Default storage location to West Europe and disable the Download files from OneLake toggle.
Disable the V-Order optimization option and enable item-level security.
Answer Description
In Microsoft Fabric, every workspace has a OneLake section in its settings.
- The Default storage location property is used to choose the Azure region where the files that the workspace writes to OneLake will be stored. Selecting West Europe guarantees that any new lakehouse or other OneLake-backed item created in the workspace stores its data in the West Europe OneLake region.
- The Download files from OneLake toggle controls whether users who have access to the workspace can download the physical files that back datasets, lakehouses, or other items. Setting this toggle to Disabled blocks file download while still allowing data exploration and queries through Fabric experiences (for example, SQL analytics endpoint or Power BI).
Other options-such as turning off the OneLake shortcut feature, changing the default data format, or disabling V-Order optimization-do not affect regional storage or file-download permissions, so they cannot satisfy the requirements.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is OneLake in Microsoft Fabric?
Why is the Default storage location important for compliance?
What does the ‘Download files from OneLake’ toggle do?
In Microsoft Fabric, you need to allow business analysts to mark their own lakehouses as Promoted while ensuring that only members of a central data governance team can apply the Certified badge to any item. Which approach meets these requirements?
Add the governance team to the tenant's certification security group and keep the default endorsement settings that allow any contributor to promote their own items.
Disable item promotion at the tenant level and grant the governance team workspace Admin rights to certify selected items.
Create a custom role in each workspace that alone can apply both Promoted and Certified badges to items.
Make all business analysts workspace Admins and require Admin permission to certify or promote items.
Answer Description
Only users who belong to the security group that the Fabric (Power BI) admin designates for certification can apply the Certified badge. Any user with write access to an item can apply the Promoted badge, provided the admin has not disabled endorsement in tenant settings. Therefore, adding the governance team to the designated certification group and leaving endorsement enabled for everyone else satisfies both conditions. Making analysts workspace admins is unnecessary, and tenant-wide disabling of promotion would prevent analysts from promoting their content.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is the difference between the Certified and Promoted badges in Microsoft Fabric?
How does tenant endorsement settings impact item promotion and certification?
Why is it unnecessary to make business analysts Workspace Admins for item promotion or certification?
A Fabric Eventstream ingests JSON telemetry from Azure Event Hubs and writes the data to a Delta table in a Lakehouse. After a device firmware update, the incoming payload now contains an additional field named "pressure". In the Eventstream monitoring pane you notice that events are being dropped with a "column count mismatch" error on the Lakehouse output. You must allow the new field to be stored without losing events or recreating the table. Which action should you take?
Modify the Eventstream transformation to exclude the "pressure" field from the SELECT statement.
Run a Spark SQL ALTER TABLE command in the Lakehouse to add a "pressure" column with the appropriate data type.
Enable the Auto create table option on the Lakehouse output so the table is regenerated with the new column.
Increase the Lakehouse output batch interval to give the service more time to process larger events.
Answer Description
The Lakehouse Delta table schema must match the list of columns produced by the Eventstream SELECT statement. When a new field appears in the incoming data, the Eventstream continues to forward it, but the Lakehouse write fails if the target table lacks that column. Adding the new column to the existing table with an ALTER TABLE statement brings the schemas back into alignment. Auto-create only works when no table exists, removing the column would discard required data, and changing the batch interval does not affect schema mismatches.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a Delta table in a Lakehouse?
What is a Spark SQL ALTER TABLE command used for?
How does Eventstream interact with a Lakehouse table?
You manage a Microsoft Fabric Warehouse named SalesDW in a workspace. The Orders table contains a SensitiveAmount column that must be completely hidden from all analysts except members of the Azure AD group FinanceLeads. Analysts still need to query every other column in the table. Which approach meets this requirement with minimal ongoing administration?
Create a row-level security policy that filters SensitiveAmount for non-FinanceLeads users.
Create a database role for regular analysts, DENY SELECT on SensitiveAmount to that role, and GRANT SELECT on Orders to FinanceLeads.
Hide SensitiveAmount with object-level security in a semantic model and have analysts query through that model only.
Apply dynamic data masking to SensitiveAmount and grant SELECT on Orders to all analysts.
Answer Description
Column-level security in a Microsoft Fabric Warehouse is implemented with the GRANT and DENY T-SQL statements. The simplest approach is to separate users into two roles:
- Add all analysts who should not see the sensitive data to a database role named AnalystsNoSensitive and DENY SELECT on the SensitiveAmount column to that role.
- Grant SELECT on the entire Orders table (or explicitly on the SensitiveAmount column) to the FinanceLeads group. Because the DENY is scoped only to the AnalystsNoSensitive role-not to the public role-it blocks access for ordinary analysts while FinanceLeads retains full visibility. Dynamic data masking leaves the column visible (only masks its values), row-level security filters rows not columns, and object-level security in a Power BI model cannot protect direct SQL access to the warehouse.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is column-level security in Microsoft Fabric Warehouse?
How is DENY different from GRANT in T-SQL?
What is the purpose of database roles in managing access control?
You manage a Fabric Eventstream that ingests JSON telemetry from Azure IoT Hub and routes the data to an Eventhouse table. After a recent device firmware update, the Eventstream Monitoring dashboard shows a rapid increase in the "Failed to write events" metric for the Eventhouse output, while the "Input events" metric remains steady. Which action should you take first to identify the root cause of the failures?
Examine the rejected events in the Eventhouse destination's error store (dead-letter folder).
Delete and recreate the Eventhouse output with the "Auto create table" option enabled.
Refresh the Eventstream input schema to force automatic column mapping.
Stop and restart the Eventstream to clear transient write errors.
Answer Description
A sudden rise in the "Failed to write events" metric while input volume is unchanged indicates that events reach the Eventstream but are rejected when written to the Eventhouse table. The Eventstream automatically stores such rejected events in the error store (dead-letter folder) of the Eventhouse destination. Inspecting the error store lets you examine sample payloads and error details-such as schema mapping or data-type mismatches-so you can determine why the write operation fails. Refreshing the Eventstream schema, restarting the Eventstream, or recreating the output does not provide the necessary diagnostic information before you understand the rejection reason.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is an Eventhouse table in Microsoft Fabric?
What is a dead-letter folder or error store in an Eventstream?
How does schema mapping affect data ingestion in Eventstreams?
Neat!
Looks like that's it! You can go back and review your answers or click the button below to grade your test.