⚡️ Pass with Confidence Sale - 40% off ALL packages! ⚡️

2 days, 6 hours remaining!
00:20:00

CompTIA Data+ Practice Test (DA0-002)

Use the form below to configure your CompTIA Data+ Practice Test (DA0-002). The practice test can be configured to only include certain exam objectives and domains. You can choose between 5-100 questions and set a time limit.

Logo for CompTIA Data+ DA0-002 (V2)
Questions
Number of questions in the practice test
Free users are limited to 20 questions, upgrade to unlimited
Seconds Per Question
Determines how long you have to finish the practice test
Exam Objectives
Which exam objectives should be included in the practice test

CompTIA Data+ DA0-002 (V2) Information

The CompTIA Data+ exam is a test for people who want to show they understand how to work with data. Passing this exam proves that someone can collect, organize, and study information to help businesses make smart choices. It also checks if you know how to create reports, use charts, and follow rules to keep data safe and accurate. CompTIA suggests having about 1 to 2 years of experience working with data, databases, or tools like Excel, SQL, or Power BI before taking the test.

The exam has different parts, called domains. These include learning basic data concepts, preparing data, analyzing it, and creating easy-to-read reports and visualizations. Another important part is data governance, which covers keeping data secure, private, and high quality. Each section of the test has its own percentage of questions, with data analysis being the largest part at 24%.

Overall, the CompTIA Data+ exam is a good way to prove your skills if you want a career in data. It shows employers that you know how to handle data from start to finish, including collecting it, checking it for errors, and sharing results in clear ways. If you enjoy working with numbers and information, this certification can be a great step forward in your career.

Free CompTIA Data+ DA0-002 (V2) Practice Test

Press start when you are ready, or press Change to modify any settings for the practice test.

  • Questions: 20
  • Time: Unlimited
  • Included Topics:
    Data Concepts and Environments
    Data Acquisition and Preparation
    Data Analysis
    Visualization and Reporting
    Data Governance
Question 1 of 20

During the weekly data-load process, a junior data analyst runs a SQL view that casts the column quantity_sold to INT. This week the script fails and returns the runtime error:

Conversion failed when converting the varchar value 'N/A' to data type int.

The schema of the staging and target tables has not changed since the previous successful load. Which action should the analyst take first to troubleshoot the issue and prevent it from happening in future loads?

  • Validate the source file and cleanse any non-numeric values in quantity_sold before loading the staging table.

  • Increase the database server's memory allocation so the CAST operation can complete in memory.

  • Enable detailed query-plan logging on the database server to capture the statement's execution plan.

  • Rewrite the view to use a FULL OUTER JOIN instead of an INNER JOIN to eliminate rows with nulls.

Question 2 of 20

A data analyst needs to gather daily pricing data for a list of products from several e-commerce websites. After confirming that the websites do not offer a data feed or an API, the analyst plans to use a script to automatically visit each product page and extract the price from the site's HTML. Which data acquisition technique is the analyst planning to use?

  • API integration

  • Web scraping

  • Database querying

  • Log file analysis

Question 3 of 20

A data analyst is working with a table of retail sales transactions that includes the store's ZIP code for each transaction. To better understand sales performance, the analyst needs to incorporate local demographic data, such as median household income and population density for each ZIP code, from an external public data source. Which data transformation technique does this scenario describe?

  • Augmentation

  • Creating a derived variable

  • Imputation

  • Standardization

Question 4 of 20

A data-engineering team packages a nightly ETL routine into a Docker container. They want to define the container's vCPU and memory requirements, pay only for those resources while the job is running, and avoid provisioning or patching any EC2 instances or Kubernetes worker nodes. Which AWS service best satisfies these goals?

  • Amazon EKS with self-managed worker nodes

  • AWS Fargate

  • AWS Lambda

  • Amazon EC2 Auto Scaling group using Spot Instances

Question 5 of 20

Your team is building a revenue dashboard for senior leadership. One bar chart compares Year-to-Date revenue for five product lines. Company branding guidelines specify navy (#002B5C) as the primary color and provide four complementary accent colors approved for communications. In the first draft the analyst colored the five bars with progressively lighter tints of navy. During user testing several viewers said the bars looked too similar to distinguish quickly, and one color-blind reviewer reported difficulty seeing any difference at all. The analyst must keep the chart visually on-brand but improve categorical discrimination.

Which design change is the most appropriate?

  • Replace the palette with a red-to-green diverging scale to maximize visual contrast between high and low revenue.

  • Render every bar in neutral gray and rely on direct data labels above each bar for identification.

  • Keep all bars navy, varying only their opacity and add a gradient legend explaining the shades.

  • Apply a qualitative palette that keeps one bar in navy and colors the other four bars with the approved accent hues so each product has a distinct, high-contrast color.

Question 6 of 20

A retail organization is designing a star schema to track daily sales. Analysts discover that a single customer can participate in multiple loyalty programs at the same time, and each loyalty program can include many customers. To keep a single-valued surrogate key in the Customer dimension while still allowing reports to filter sales by loyalty program, which structured data element should the data architect add between the two dimensions?

  • Introduce a bridge table that lists every valid customer-loyalty program pair.

  • Store the loyalty program ID as a degenerate dimension in the fact table.

  • Convert the Customer dimension to a Type 2 slowly changing dimension.

  • Create a junk dimension that combines loyalty program codes with other flags.

Question 7 of 20

A dashboard application sends the following Transact-SQL statement to the OLTP database dozens of times each minute, changing only the literal value in the WHERE clause that comes from the user interface:

SELECT customer_id, order_date, total_amount
FROM sales.orders
WHERE customer_id = 12345;

The DBA observes that every execution causes SQL Server to parse and compile a new plan, and the plan cache is filling up with single-use plans for this query pattern. Without altering table structures or adding new indexes, which basic query optimization technique is most likely to cut CPU time by allowing the same execution plan to be reused across calls?

  • Copy the orders into a temporary table and run the filter against the temp table instead.

  • Submit the statement as a parameterized query that passes the customer_id as a bind variable.

  • Add the WITH (NOLOCK) table hint to the orders table.

  • Append the OPTION (RECOMPILE) hint to force the optimizer to build a fresh plan each time.

Question 8 of 20

A data analyst has been assigned to a new project that uses a legacy customer relationship management (CRM) database. To build accurate reports, the analyst must first understand the definitions, data types, and business rules for each field in the 'Customers' table. Which of the following documents would be the MOST direct resource for this information?

  • Data explainability report

  • Data dictionary

  • Data flow diagram

  • Data lineage report

Question 9 of 20

A retail organization asks its analytics team to deliver weekly sales data to department managers. The managers want to slice, filter, and drill down into product-level details whenever they choose rather than file a request with the BI team. They already sign in to the corporate intranet through single sign-on (SSO), and each manager must see only the rows that belong to their own department. Which delivery method BEST satisfies these requirements?

  • Push real-time sales transaction alerts to managers' mobile devices through a notification service.

  • Export a weekly CSV snapshot of the sales table to a shared network drive for managers to open in Excel.

  • Email a static PDF dashboard to each manager on a weekly schedule.

  • Publish an interactive report to a self-service analytics portal integrated with the intranet's SSO and configured for row-level security.

Question 10 of 20

An investment firm subject to SEC Rule 17a-4 is designing a cloud solution to store daily trading records. Compliance requires that electronic records be retained in a write-once, read-many (WORM) or audit-trail system that prevents alteration before the retention period ends, and that the records remain "easily accessible" for the first 2 years but may be moved to a lower-cost tier for the remainder of the mandatory 6-year retention period.

Which storage strategy BEST satisfies these requirements while controlling long-term cost?

  • Upload the files to object storage with server-side encryption and enable cross-region replication to a bucket in another continent after 2 years.

  • Enable Object Lock in compliance mode on a versioned object-storage bucket, keep each object in the standard tier for 730 days, then use a lifecycle rule to transition it to an immutable archive tier for the remainder of the 6-year retention period.

  • Take block-storage snapshots every day, retain each snapshot for 6 years, and periodically export a monthly snapshot copy to cold storage.

  • Store the files on an encrypted SAN volume replicated to a secondary data center and copy daily incremental backups to tape vaulted off-site for 6 years.

Question 11 of 20

A data analyst is designing a table to store user-profile information for a new global e-commerce platform. The user_bio column must accommodate biographies of variable length that may contain text in many languages and include special characters such as emojis. Which data type is the most appropriate for the user_bio column to ensure broad character compatibility and efficient storage?

  • char

  • CLOB

  • varchar

  • nvarchar

Question 12 of 20

A junior analyst is publishing a Power BI dataset that pulls data from a PostgreSQL database. During the scheduled refresh in the Power BI Service the dataset fails with the message "The remote certificate is invalid according to the validation procedure." Database administrators confirm the PostgreSQL server's SSL certificate is trusted and no recent changes have been made. Internal documentation offers no guidance and the analyst must restore the dashboard before an executive review that starts in two hours. Which action is the MOST appropriate first step to accelerate troubleshooting of this issue?

  • Disable encryption in the connection settings and attempt the refresh without SSL.

  • Search the Microsoft Power BI community forum and knowledge base for posts that describe the same certificate error and recommended fixes.

  • Request that the database team regenerate the server's root certificate and restart PostgreSQL.

  • Reinstall Power BI Desktop and republish the dataset from a fresh installation.

Question 13 of 20

A data analyst for a national retail chain is tasked with identifying promising locations for new stores. The analyst has a dataset containing the exact street addresses of all current stores, along with census tract data that includes median household income and population density across the country. To effectively visualize which geographic areas have high income and are far from existing stores, which of the following visualizations would be most appropriate?

  • Choropleth map

  • Bar chart

  • Pivot table

  • Scatter plot

Question 14 of 20

A reporting analyst is designing a relational table to archive detailed customer feedback notes. Each note can contain up to 200 KB (about 200,000 characters) of plain Unicode text, and analysts will need to run SQL string functions-such as LIKE pattern searches and SUBSTRING extraction-directly against the stored content. Which data type should the analyst assign to the column that stores the feedback notes so the requirement is satisfied without imposing an unnecessary size limit?

  • FLOAT

  • VARCHAR(255)

  • BLOB (Binary large object)

  • CLOB (Character large object)

Question 15 of 20

A data analyst is tasked with analyzing a sales dataset imported from a CSV file. The analyst needs to calculate the total revenue by summing the SaleAmount column. However, the query fails with an error indicating a data type mismatch. Upon inspection, the SaleAmount column contains values formatted as strings with a currency symbol, such as '$1,250.75'. To perform the required calculation, which data transformation technique must be applied to the SaleAmount column first?

  • Standardization

  • Conversion

  • Aggregation

  • Parsing

Question 16 of 20

While troubleshooting a slow production database, a data analyst opens MySQL Workbench and wants to immediately see which SQL statements have performed full table scans, relied on temporary tables, or fall into the slowest five-percent of runtimes-all without writing any SQL. Which built-in Workbench tool should the analyst open to generate this report?

  • Visual Explain Plan panel

  • Performance Schema Reports

  • SQL Data Export wizard

  • Performance Dashboard graphs

Question 17 of 20

A data analyst needs to make a quick modification to a Python script on a remote server through a command-line interface. The task does not require debugging, code completion, or project management features. Which of the following tool categories provides the most direct and lightweight solution for this task?

  • Text editor

  • Business Intelligence (BI) software

  • Integrated Development Environment (IDE)

  • Database Management System (DBMS)

Question 18 of 20

A data analyst creates a sales dashboard that performs well with a small, local test dataset. After deploying the dashboard and connecting it to the production database, which contains millions of rows of historical data, users report that it takes several minutes to load. The analyst confirms other reports on the same server are performing normally. Which of the following is the MOST effective initial step for the analyst to take to troubleshoot the excessive load time?

  • Apply data filtering at the source to limit the query to only the necessary data, such as the most recent fiscal year.

  • Archive the historical data in the production database that is more than two years old.

  • Redesign the dashboard to use only simple tables and remove all complex charts and graphs.

  • Escalate the issue to the database administration team to investigate server processing and memory usage.

Question 19 of 20

A data analyst is working with a data extract from a legacy system saved as product_list.txt. Which statement accurately describes a primary characteristic of a .txt file that the analyst must account for when preparing to load it into a database?

  • The file format itself does not store metadata or formatting, so the analyst must infer the data's structure, such as the delimiter and character encoding.

  • The file is a binary format that requires a specific database driver or proprietary software to be read correctly.

  • The file organizes data into a nested, hierarchical structure of objects and arrays.

  • The file inherently contains schema information, including data types and column headers, in an unformatted header block.

Question 20 of 20

A data analyst at an e-commerce company discovers that the marketing team's customer report shows 150,000 active customers, while the sales team's CRM reports 145,000. This discrepancy is causing confusion in strategic planning. To resolve this and ensure consistent reporting across the organization, which of the following data management concepts is MOST critical to establish?

  • Data lineage documentation

  • Continuous data integration

  • A comprehensive data dictionary

  • A single source of truth (SSOT)