Microsoft Azure Data Fundamentals Practice Test (DP-900)
Use the form below to configure your Microsoft Azure Data Fundamentals Practice Test (DP-900). The practice test can be configured to only include certain exam objectives and domains. You can choose between 5-100 questions and set a time limit.

Microsoft Azure Data Fundamentals DP-900 Information
The Microsoft Azure Data Fundamentals certification exam is designed to validate foundational knowledge of core data concepts and how they are implemented using Microsoft Azure data services. It is ideal for individuals who are new to data workloads and cloud environments. Using DP-900 practice tests, practice exams, and reviewing many practice questions can help candidates build confidence, familiarise themselves with exam language, and strengthen their grasp of key topics.
The exam covers four major domains: describing core data concepts (25-30%), identifying considerations for relational data on Azure (20-25%), describing non-relational data on Azure (15-20%), and describing analytics workloads on Azure (25-30%). To prepare effectively, leveraging fullālength practice exams and targeted practice questions focused on each domain will help you identify weak areas, improve your timing, and enhance your readiness for the real exam experience.
Practice Exams & Practice Questions
Success on the DP-900 exam isnāt just about recalling facts, you'll need to apply them under timed conditions. Using DP-900 practice tests helps simulate the exam environment, while drilling practice questions for each objective ensures your understanding is solid. Practice exams expose you to question types like case studies, drag-and-drop, multipleāchoice and multipleāresponse, allowing you to manage pacing and reduce surprises on exam day. With consistent work on practice exams and practice questions, youāll go into the exam with increased confidence and reduce the chance of needing a retake.

Free Microsoft Azure Data Fundamentals DP-900 Practice Test
- 20 Questions
- Unlimited
- Describe core data conceptsIdentify considerations for relational data on AzureDescribe considerations for working with non-relational data on AzureDescribe an analytics workload
A company must ingest data from several on-premises SQL Server databases into Azure Data Lake Storage on a nightly schedule. The solution should orchestrate the batch pipeline and allow developers to design data transformations by using a visual, code-free interface. Which Azure service meets these requirements?
Azure Synapse Analytics dedicated SQL pool
Azure Data Factory
Azure Databricks
Azure Event Hubs
Answer Description
Azure Data Factory is Microsoft's cloud-based data integration service. It can connect to on-premises data sources through self-hosted integration runtimes, schedule and orchestrate batch pipelines, and provides Mapping Data Flows so that transformations can be built in a drag-and-drop, code-free designer. Azure Event Hubs is optimized for real-time event streaming, not scheduled batch ETL. Azure Databricks is an analytics platform that requires writing notebooks rather than providing a visual orchestration interface. An Azure Synapse Analytics dedicated SQL pool is an analytical data store, not an ingestion and orchestration service.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Azure Data Factory used for?
How does Azure Data Factory connect to on-premises data sources?
What are Mapping Data Flows in Azure Data Factory?
In Azure Table storage, which combination of properties forms the unique primary key that lets the service locate a single entity in one operation?
The PartitionKey value only
The combination of PartitionKey and RowKey
The combination of Timestamp and ETag
The RowKey value only
Answer Description
Azure Table storage uses a composite primary key made up of the PartitionKey and RowKey values. The PartitionKey groups related entities so the service can route requests to the correct storage node, while the RowKey uniquely identifies the entity within that partition. Together, these two values allow an efficient point lookup. Other properties such as ETag or Timestamp are metadata fields and are not used as part of the primary key, and neither the PartitionKey alone nor the RowKey alone is sufficient to uniquely identify an entity across the table.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is the purpose of the PartitionKey in Azure Table storage?
How does the RowKey function in Azure Table storage?
Why are ETag and Timestamp not part of the primary key in Azure Table storage?
You are designing a solution that will store information about friendships and follower relationships among millions of users. The data is highly interconnected, and your application must run frequent graph traversal queries such as finding the shortest path between two users. In Azure Cosmos DB, which API should you choose to meet these requirements?
Cassandra API
Table API
Gremlin API
Core (SQL) API
Answer Description
The Gremlin API in Azure Cosmos DB provides a property graph data model and supports the Gremlin traversal language, which is designed for working with highly connected data such as social graphs, recommendation engines, and network topologies. It enables efficient graph traversals like shortest-path and neighbor queries.
The Cassandra API focuses on wide-column workloads, the Table API offers a key-value store for schemaless data, and the Core (SQL) API targets JSON document workloads. None of these alternatives provide native graph structures or traversal operators, so they are less suitable for social-network relationship data.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is the Gremlin API used for in Azure Cosmos DB?
How does the Gremlin API differ from the Core (SQL) API in Azure Cosmos DB?
What are some use cases where the Gremlin API is a better choice than other APIs in Azure Cosmos DB?
A company plans to migrate an on-premises SQL Server database to Azure. The database uses Windows-authenticated logins, runs SQL Server Agent jobs, and requires access to the underlying operating system and file system for custom components. The team wants the greatest administrative control over both the database engine and the OS. Which Azure SQL service should they deploy?
Azure SQL Database elastic pool
SQL Server on Azure Virtual Machines
Azure SQL Managed Instance
Azure SQL Database single database
Answer Description
SQL Server on Azure Virtual Machines provides Infrastructure-as-a-Service. Administrators manage the full SQL Server installation and the Windows or Linux OS, so they can use Windows authentication, SQL Server Agent, registry edits, file system access, and other custom components. Azure SQL Database (single database or elastic pool) is Platform-as-a-Service and hides the OS layer; these options do not expose SQL Server Agent or allow Windows logins. Azure SQL Managed Instance offers near-full SQL Server compatibility but still runs as PaaS without giving OS-level access, so custom registry or file system changes are not possible. Therefore, SQL Server on Azure Virtual Machines best fits the requirements.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is SQL Server on Azure Virtual Machines?
How does SQL Server on Azure Virtual Machines differ from Azure SQL Managed Instance?
Why can't Azure SQL Database or elastic pools meet the requirements described?
A team is reviewing the OrderDetails table in an Azure SQL Database. The table stores the columns OrderID, ProductID, ProductName, ProductPrice, and Quantity. Because ProductName and ProductPrice repeat across many rows that share the same ProductID, you are asked to apply normalization principles. Which change best aligns the table with the goals of normalization?
Move ProductName and ProductPrice to a Products table keyed by ProductID and reference that key from OrderDetails.
Replace ProductName with a JSON column that stores both the name and price together in the same field.
Add an additional ProductPrice column to store promotional prices alongside the regular price.
Combine ProductID and OrderID into a single composite column to reduce the number of columns in the table.
Answer Description
Normalization reduces data redundancy and prevents update anomalies by ensuring each fact is stored exactly once. Repeating attributes that describe a product (its name and price) should be kept in a separate Products table that uses ProductID as its primary key. OrderDetails would then hold only the ProductID as a foreign-key reference, eliminating duplicated product information. Creating a composite column, adding extra price columns, or embedding attributes in JSON does not remove redundancy and therefore does not achieve the fundamental purpose of normalization.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is database normalization and why is it important?
What is the difference between a primary key and a foreign key?
Why is redundancy problematic in a database?
Your company is developing a mission-critical retail application that must automatically serve read and write requests from the Azure region nearest each customer and continue operating if any region becomes unavailable. When configuring Azure Cosmos DB, which built-in capability should you enable first to satisfy this requirement?
Analytical store for Azure Synapse Link
Dedicated Gateway
Change feed
Global distribution with multi-region replication
Answer Description
Global distribution lets you associate your Azure Cosmos DB account with one or more Azure regions. The service automatically replicates data to every selected region, supports multi-region writes, and can perform automatic failover if a region goes offline. Change feed, Dedicated Gateway, and the analytical store add other benefits such as event streams, connection isolation, or analytical workloads, but none of them replicate data across regions or provide automatic regional failover.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Azure Cosmos DB Global Distribution?
How does multi-region replication work in Azure Cosmos DB?
What is the difference between Change Feed and Global Distribution?
You are designing a table in a relational database. You must guarantee that every row can be uniquely identified and that the identifying column cannot contain NULL values. Which relational data feature satisfies this requirement?
Primary key constraint
Unique index that allows NULLs
Composite foreign key
Default constraint
Answer Description
A primary key constraint enforces entity integrity by requiring that the selected column or combination of columns contain unique, non-NULL values for every row. Foreign keys enforce referential integrity between tables but do not guarantee uniqueness in the referenced table. A unique index that allows NULLs can still store multiple NULLs and therefore does not fully meet the requirement. A default constraint only supplies a value when none is provided; it does not enforce uniqueness or prevent NULLs on its own.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is the purpose of a primary key in a relational database?
How does a foreign key differ from a primary key in relational databases?
Why can't a unique index that allows NULLs be used instead of a primary key?
You must change the list price of every row in the DimProduct table where ProductCategory is 'Bikes'. Which type of SQL statement should you use to accomplish this task?
UPDATE
SELECT
INSERT
DELETE
Answer Description
The UPDATE statement is designed to modify the values stored in existing rows. It supports a SET clause to specify new column values and a WHERE clause to limit the rows that are altered. SELECT only reads data, INSERT adds new rows, and DELETE removes rows, so none of those statements can change data that is already stored.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is an UPDATE statement in SQL?
How does the WHERE clause work in an UPDATE statement?
What is the difference between UPDATE and INSERT in SQL?
You are designing an Azure analytics solution that periodically ingests data from a transactional SQL database into a data lake. To reduce network and storage costs, you want to transfer only rows that were inserted or updated since the last load instead of copying the entire tables each time. Which ingestion technique should you use?
Database sharding
Change data capture (CDC)
Data masking
Regular full database backups
Answer Description
Change data capture records insert, update, and delete operations on source tables, allowing an ingestion pipeline to move only the rows that have changed since the previous load. This minimizes the amount of data copied and lowers both network usage and storage costs. Sharding spreads data across multiple databases for scale, but it does not identify changed rows. Data masking obfuscates sensitive fields and is unrelated to incremental loads. A full database backup copies all data every time, which does not meet the requirement to ingest only new or modified rows.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Change Data Capture (CDC) in Azure analytics?
How does CDC minimize network and storage costs in Azure solutions?
How is CDC implemented in Azure Data Factory?
You stream sensor events into Azure Stream Analytics and need a query that continuously calculates the average temperature over the most recent five-second period, refreshing the value every second. Which type of temporal window should you use in the query?
Session window
Tumbling window
Hopping window
Sliding window
Answer Description
A hopping window lets you specify both a window length and a hop size. Setting a five-second window length with a one-second hop creates overlapping windows: each new window starts one second after the previous one, but still covers the last five seconds of data. This produces a new, updated average every second. Tumbling windows are non-overlapping, so results would appear only once every five seconds. Sliding windows emit results only when events enter or leave the window, so they do not guarantee output on a fixed schedule. Session windows group bursts of activity separated by idle gaps and are not appropriate for fixed-interval aggregation.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a hopping window in Azure Stream Analytics?
How does a tumbling window differ from a hopping window?
When should you use a sliding window versus a hopping window?
Your company needs to store millions of JSON-formatted telemetry events. The data has no fixed schema, is retrieved by a combination of PartitionKey and RowKey, and must scale while remaining inexpensive. Which Azure storage option should you recommend?
Azure File storage
Azure SQL Database
Azure Blob storage
Azure Table storage
Answer Description
Azure Table storage is a NoSQL key-value store that uses a two-part key consisting of PartitionKey and RowKey for fast retrieval. It is designed for semi-structured data, automatically scales to very large datasets, and costs less than relational services. Blob storage is optimized for unstructured binary objects, File storage exposes SMB shares, and Azure SQL Database is a relational service that requires a predefined schema and incurs higher cost for massive scale.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a PartitionKey and RowKey in Azure Table Storage?
How does Azure Table Storage scale for massive datasets?
Why is Azure Table Storage cost-effective compared to relational databases?
A single relational table stores both project assignments and employee contact details, causing each employee phone number to appear in multiple rows. Which relational design concept should you apply to eliminate this duplication while still allowing projects to reference the correct employees?
Apply horizontal partitioning to split the table by project.
Normalize the schema by placing employee data in a separate table linked with a foreign key.
Denormalize the table to include all employee and project columns together.
Create a nonclustered index on the phone number column.
Answer Description
Normalization restructures data into separate, related tables so that each fact is stored only once. By moving employee details to their own table and referencing that table from the project-assignment table with a foreign key, redundant phone numbers are removed and update anomalies are avoided. Denormalization does the opposite by intentionally introducing redundancy. Indexing and horizontal partitioning can improve performance or manage large datasets, but neither addresses repeated data values or integrity concerns.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is normalization in databases?
What is a foreign key in relational databases?
How does normalization prevent data anomalies?
In the context of data representation options available in Azure solutions, which statement correctly distinguishes semi-structured data from structured data?
It consists purely of binary blobs such as video files with no meaningful metadata or tags.
It has no predefined schema but uses self-describing tags or key-value pairs to organize fields, as in JSON files.
It is stored in fixed-width columns and must adhere to a rigid table schema defined in a relational database.
It stores records as edges and nodes optimized for traversing relationships in a property graph model.
Answer Description
Semi-structured data, such as JSON or XML, does not rely on a rigid table schema. Instead, it carries its own self-describing structure through tags or key-value pairs, allowing records to contain optional or variable attributes. Structured data, by contrast, is organized into relational tables with predefined columns and data types. Unstructured data (for example, video files) lacks an internal organizational model beyond basic metadata, while graph data stores focus on nodes and edges rather than tagged elements. Therefore, the description referring to flexible, self-describing tags is the only accurate characterization of semi-structured data.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is semi-structured data and how is it different from structured data?
Can semi-structured data be stored in Azure and what are some storage options?
What are common use cases of semi-structured data?
Your team is building a mobile app that stores each user profile as a JSON document. The data must be automatically indexed, served with single-digit millisecond latency, and replicated transparently to several Azure regions. Which Azure data storage option best meets these requirements?
Azure Table storage
Azure SQL Database Hyperscale
Azure Cosmos DB using the Core (SQL) API
Azure Blob Storage combined with Azure Cognitive Search
Answer Description
Azure Cosmos DB is a fully managed, globally distributed NoSQL database designed for JSON documents. It automatically indexes every property, offers configurable consistency with single-digit millisecond latency for reads and writes, and can replicate data to any Azure region with the click of a button. Azure SQL Database Hyperscale is a relational store that does not accept schemaless JSON as a native data model and cannot automatically index every property. Azure Blob Storage with Azure Cognitive Search can hold the JSON files but relies on a separate indexer service and does not guarantee low-latency transactional reads or writes. Azure Table storage is a key-attribute store limited to partition-key and row-key indexes and provides eventual consistency without the turnkey global distribution or SLA characteristics required.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Azure Cosmos DB and how does it handle JSON documents?
What is single-digit millisecond latency and why is it important for modern applications?
How does Azure Cosmos DB ensure global data replication across regions?
You are analyzing an existing relational database. Which feature is specifically responsible for enforcing referential integrity between two tables?
Clustered index
Primary key constraint
Stored procedure
Foreign key constraint
Answer Description
In a relational database, a foreign key constraint links a column (or set of columns) in one table to the primary key of another table. This relationship ensures that any value inserted in the child table must first exist in the parent table, thereby maintaining referential integrity. A primary key uniquely identifies rows within its own table but does not enforce relationships across tables. Indexes improve query performance but do not guarantee integrity. Stored procedures contain executable code and likewise do not enforce referential rules automatically.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a foreign key constraint in detail?
How is a primary key different from a foreign key?
Can you explain referential integrity with a practical example?
You are organizing customer sales data in a new Azure SQL Database. You decide to place customer information and order information in separate tables connected by primary and foreign keys. According to database normalization principles, what is the main benefit of structuring the data this way?
It enables the database to natively store large volumes of unstructured JSON documents.
It reduces redundant data and prevents update, insert, and delete anomalies, improving data integrity.
It deliberately duplicates data across tables to improve fault tolerance.
It guarantees faster query performance by eliminating the need for joins.
Answer Description
Normalization splits data into related tables so that each fact is stored only once. This minimizes duplication and the risk of update, insert, and delete anomalies, thereby protecting the consistency and accuracy of the data. Performance gains or support for unstructured data are not the primary objectives, and normalization does not intentionally increase duplication.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What are primary and foreign keys in a database?
What are update, insert, and delete anomalies in database design?
How does normalization improve data integrity?
A data engineer receives device telemetry logs that arrive as JSON files. Each file contains key-value pairs and nested arrays but no fixed columns. You must classify the data type to choose an appropriate Azure storage solution. Which category does this JSON telemetry data fall into?
Structured data
Relational data
Unstructured data
Semi-structured data
Answer Description
JSON documents store information as key-value pairs and often include nested objects or arrays. They have an inherent, flexible schema but do not require a rigid table structure with predefined columns, so they are considered semi-structured data. Structured data fits well into relational tables, while unstructured data (such as images or free-form text) lacks any consistent schema. The term "relational data" refers specifically to structured data managed in relational databases, not to JSON files.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is semi-structured data?
Why is JSON considered semi-structured and not structured?
Which Azure storage solutions work best for JSON data?
A legacy line-of-business application running on several Windows Server virtual machines expects to read and write to a normal SMB network share and relies on NTFS permissions. You need to migrate this workload to Azure without changing the application code. Which Azure storage service best meets the requirement?
Azure Managed Disks (Premium SSD)
Azure Table storage
Azure Blob storage in the hot access tier
Azure File storage
Answer Description
Azure File storage provides fully managed file shares exposed through the SMB protocol and supports Windows access control lists (ACLs), so the application can continue using standard file APIs after migration.
Azure Blob storage is object storage accessed over HTTP rather than SMB, so the app would have to be rewritten. Azure Managed Disks are block-level volumes that attach to virtual machines; although a shared-disk feature allows a disk to be mounted by multiple VMs for clustering scenarios, the disk is still presented as a block device, not an SMB file share. Azure Table storage is a NoSQL key-value store and is unsuitable for file-based workloads.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Azure File Storage?
What are the differences between SMB and HTTP protocols in Azure storage?
Why is Azure Blob Storage not suitable for the SMB-based application?
An IoT project stores device telemetry as JSON objects that can include new properties at any time. Engineers must query individual documents and filter on nested fields without redesigning a fixed schema. Which type of database is most appropriate for this requirement?
Column-family database
Relational database
Graph database
Document database
Answer Description
A document database stores data as self-describing JSON (or BSON) documents, allowing each record to have its own shape. Because the schema is flexible, new properties can be added without altering a central definition, and queries can directly target nested values inside each document. Column-family databases can add columns dynamically but organize data in rows and column families rather than self-contained JSON documents, so querying nested JSON structures is not native. Relational databases rely on a fixed table schema, and graph databases are optimized for relationship traversal rather than flexible JSON storage. Therefore, a document database best meets the requirements.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a document database?
How do document databases query nested fields?
Why are relational databases not suitable for flexible schemas?
An e-commerce company wants to detect potential fraud within seconds of each transaction. The solution must continuously ingest high-volume streaming data from Azure Event Hubs, run SQL-like queries with sub-second latency, and push real-time visuals to Power BI dashboards. Which Azure service should you recommend to meet these requirements?
Azure Data Factory mapping data flows
Azure HDInsight with Apache Hadoop
Azure Stream Analytics
Azure Synapse Analytics dedicated SQL pool
Answer Description
Azure Stream Analytics is designed for real-time event processing. It connects directly to streaming sources such as Event Hubs or IoT Hub, lets users express analytics in a SQL-like language, supports processing with millisecond latency, and can output results to Power BI for live dashboards. Azure Synapse Analytics and Azure Data Factory are optimized for batch workloads, while HDInsight Hadoop clusters focus on large-scale distributed processing rather than low-latency streaming queries.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Azure Stream Analytics?
How does Azure Stream Analytics integrate with Power BI?
What is the difference between Azure Stream Analytics and Azure Synapse Analytics?
That's It!
Looks like that's it! You can go back and review your answers or click the button below to grade your test.