- ML News
- Posts
- AI Roundup: Models, Agents, & Optimization
AI Roundup: Models, Agents, & Optimization
This week’s AI highlights include model developments at Meta and OpenAI, research on cost-effective prompt engineering, significant infrastructure updates for Stargate, and a paper exploring the potential bias introduced by activation functions.
Meta Abandons Llama in Favor of Claude Sonnet…
📝Meta switching from Llama to Claude Sonnet for internal coding is a significant shift in AI strategy and indicates that Claude Sonnet is more efficient for internal coding purposes.
Introducing ChatGPT agent
📝ChatGPT Agent introduces a new paradigm for AI interaction, enabling it to think, act, and use tools to complete complex tasks under user guidance, potentially revolutionizing workflows.
[R] Interesting paper on cost-aware prompt optimization (CAPO)
📝This paper presents a cost-aware prompt optimization technique using evolutionary algorithms, which can significantly improve the efficiency and performance of prompts, especially relevant in the context of rising API costs.
Stargate advances with 4.5 GW partnership with Oracle
📝OpenAI’s partnership with Oracle to develop 4.5 gigawatts of data center capacity demonstrates significant infrastructure investments to support AI development and deployment, indicating a focus on scaling AI capabilities.
[R][D] Interpretability as a Side Effect? Are Activation Functions Biasing Your Models?
📝This paper challenges the notion that interpretability is inherent in deep learning, suggesting it may be a byproduct of design choices in activation functions, potentially reframing interpretability efforts and model expressiveness.