CompTIA DataX DY0-001 (V1) Practice Question

You are building a text-clustering workflow that starts with an extremely sparse 1 000 000 × 50 000 term-document matrix X. Because the matrix will not fit in memory when densified, constructing the covariance matrix XᵀX for a standard principal component analysis (PCA) is not an option. Instead, you choose to apply a truncated singular value decomposition (t-SVD) to reduce the dimensionality of X prior to clustering.

Which statement best explains why t-SVD is generally preferred over covariance-based PCA for this scenario?

  • t-SVD guarantees that the resulting singular vectors are both orthogonal and sparse, making clusters easier to interpret than those obtained from PCA.

  • t-SVD forces all components of the lower-dimensional representation to be non-negative, so the projected features can be read as probabilities without any post-processing.

  • t-SVD automatically scales every column of X to unit variance, eliminating the need for TF-IDF or other term-weighting schemes.

  • t-SVD can be computed with iterative methods (e.g., randomized SVD or Lanczos) that multiply X by vectors without ever materializing XᵀX, allowing the decomposition to run efficiently on the sparse matrix.

CompTIA DataX DY0-001 (V1)
Machine Learning
Your Score:
Settings & Objectives
Random Mixed
Questions are selected randomly from all chosen topics, with a preference for those you haven’t seen before. You may see several questions from the same objective or domain in a row.
Rotate by Objective
Questions cycle through each objective or domain in turn, helping you avoid long streaks of questions from the same area. You may see some repeat questions, but the distribution will be more balanced across topics.

Check or uncheck an objective to set which questions you will receive.

SAVE $64
$529.00 $465.00
Bash, the Crucial Exams Chat Bot
AI Bot