The Download: rethinking AI benchmarks, and the ethics of AI agents

This is today’s edition of The Download, our weekday newsletter that provides a daily dose of what’s going on in the world of technology. The way we measure progress in AI is terrible Every time a new AI model is released, it’s typically touted as acing its performance against a series of benchmarks. OpenAI’s GPT-4o, for example, was launched in May with a compilation of results that showed its performance topping every other AI company’s latest model in several tests. The problem is that these benchmarks are poorly designed, the results hard…

This content is for Member members only.
Log In Register