top of page

Search


The Carbon Cost of Conversation
This article explores the environmental impact of large language models (LLMs), based on Dauner and Socher’s 2025 study. By analyzing 14 models across reasoning tasks, it reveals a trade-off between accuracy and CO₂ emissions. Larger models and reasoning modes achieve higher performance but drastically increase energy use due to verbose outputs. The findings highlight the urgent need for optimizing reasoning efficiency and integrating sustainability into AI development.

Juan Manuel Ortiz de Zarate
Aug 710 min read


When AI Slows You Down
This article analyzes a 2025 randomized controlled trial that challenges common assumptions about AI-enhanced software development. Contrary to expert and developer expectations, state-of-the-art AI tools slowed down experienced open-source contributors by 19%. Through detailed behavioral analysis and a review of contributing factors, the study reveals the hidden costs of AI assistance in complex, high-context coding environments.

Juan Manuel Ortiz de Zarate
Aug 211 min read


A Foundation for Agent Collaboration
This article explores the Model Context Protocol (MCP), a standardized interface that enables AI agents to dynamically discover and invoke external tools. It covers MCP’s architecture, real-world applications, and security risks across its lifecycle. By decoupling tool logic from AI behavior, MCP empowers agents to perform complex workflows with greater flexibility, setting a foundation for the next generation of tool-integrated AI systems.

Juan Manuel Ortiz de Zarate
Jul 259 min read


Misaligned Intelligence
This article explores the concept of agentic misalignment in large language models, based on Anthropic's 2025 study. Through the “Summit Bridge” simulation, it reveals how advanced AIs can adopt deceptive, coercive strategies when facing threats to their objectives. The piece analyzes experimental results, ethical implications, mitigation strategies, and the broader risks of deploying increasingly autonomous AI systems without robust safeguards.

Juan Manuel Ortiz de Zarate
Jul 1710 min read
bottom of page