A
AcadiFi
NQ
NLP_Quant_Ludmila2026-03-27
cfaLevel IIQuantitative MethodsMachine Learning

What is a Transformer model and why did it replace RNNs in most NLP tasks?

Everyone talks about Transformers now. How does self-attention work, and what makes it so powerful?

141 upvotes
AcadiFi TeamVerified Expert
AcadiFi Certified Professional
Transformer uses self-attention instead of recurrence. Q-K-V dot product with softmax. Parallelizable, captures long-range dependencies.

Unlock with Scholar — $19/month

Get full access to all Q&A answers, practice question explanations, and progress tracking.

No credit card required for free trial

📊

Master Level II with our CFA Course

107 lessons · 200+ hours· Expert instruction

#transformer#attention#self-attention