AWS Certified AI Practitioner AIF-C01 Practice Question
Transformer-based large language models process an entire sequence of tokens at once instead of one token at a time. Which architectural feature enables this parallel computation?
Transformers replace the recurrence found in RNNs with self-attention layers. Self-attention lets the model compute relationships between every pair of tokens simultaneously, so the whole sequence can be processed in parallel on GPU hardware. Gradient clipping is a training safeguard, LSTM cells belong to recurrent networks, and data parallelism is a high-level training strategy rather than an internal model component. None of these alternatives provide the core parallel-sequence capability that distinguishes transformer LLMs.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is the self-attention mechanism in transformers?
Open an interactive chat with Bash
How does self-attention enable parallel processing?
Open an interactive chat with Bash
Why don’t LSTM cells or data parallelism offer parallel token computation like transformers?
Open an interactive chat with Bash
AWS Certified AI Practitioner AIF-C01
Fundamentals of Generative AI
Your Score:
Report Issue
Bash, the Crucial Exams Chat Bot
AI Bot
Loading...
Loading...
Loading...
Pass with Confidence.
IT & Cybersecurity Package
You have hit the limits of our free tier, become a Premium Member today for unlimited access.
Military, Healthcare worker, Gov. employee or Teacher? See if you qualify for a Community Discount.
Monthly
$19.99
$19.99/mo
Billed monthly, Cancel any time.
3 Month Pass
$44.99
$14.99/mo
One time purchase of $44.99, Does not auto-renew.
MOST POPULAR
Annual Pass
$119.99
$9.99/mo
One time purchase of $119.99, Does not auto-renew.
BEST DEAL
Lifetime Pass
$189.99
One time purchase, Good for life.
What You Get
All IT & Cybersecurity Package plans include the following perks and exams .