Leading artificial intelligence research organizations, often referred to as "frontier labs," have largely ceased publishing extensive model evaluations as these processes become fully integrated into their core training ecosystems. This observation was shared by Alexander Doria, a Partner at Founders Fund with a Ph.D. in Computer Science from Stanford University specializing in machine learning and natural language processing. Doria's "hot take" suggests a significant shift in how advanced AI capabilities are assessed and disclosed.
The trend reflects a deeper embedding of evaluation methodologies directly within the AI development pipeline. Evaluations are now considered crucial for the iterative improvement and fine-tuning of large language models (LLMs), with automated metrics providing rapid feedback during development. This integration allows labs to continuously assess performance, identify weaknesses, and enhance models throughout their lifecycle, making evaluations an intrinsic part of the training process rather than a separate, post-development public exercise.
Despite this observed shift towards internalization, the landscape of AI evaluation remains dynamic. Major players like OpenAI and Anthropic, for instance, collaboratively published a joint AI safety evaluation of their GPT and Claude models in August 2025. This collaboration underscores an ongoing, albeit perhaps more focused, commitment to public transparency, particularly concerning safety and alignment aspects of advanced AI. The AI Safety Institute (AISI), which rebranded to the AI Security Institute in February 2025, also continues its work on independent evaluations, publishing lessons and methodologies from its assessments of frontier AI systems.
The evolving practices highlight a broader "AI safety dilemma" within the industry, balancing the benefits of transparency with concerns over dual-use risks and competitive advantage. While some advocate for greater openness to foster trust and accountability, others argue for more controlled disclosures to prevent misuse of powerful AI models. This tension influences the extent and nature of publicly shared evaluation data.
As AI capabilities continue to advance, the methods and transparency of model evaluations will remain a critical point of discussion. The integration of evaluations into the training ecosystem signifies a maturation of AI development practices, yet it also raises questions about how external stakeholders and the public will gain comprehensive insights into the capabilities and safety of the most powerful AI systems.