top of page

Search


The Architecture That Redefined AI
This article offers a deep dive into the seminal paper Attention Is All You Need, which introduced the Transformer architecture. It explores the limitations of recurrent models, the mechanics of self-attention, training strategies, and the Transformer’s groundbreaking performance on machine translation tasks. The article also highlights the architecture’s enduring legacy as the foundation for modern NLP systems like BERT and GPT.

Juan Manuel Ortiz de Zarate
May 279 min read
3 views


AI That Thinks Before It Speaks
Optimizing AI reasoning with adaptive test-time computation using recurrent depth transformers for smarter, efficient problem-solving.

Juan Manuel Ortiz de Zarate
Feb 199 min read
2 views
bottom of page