Higher Reasoning: OpenAI debuts o1 and pro mode for $200/month
OpenAI launched not only its highly anticipated o1 model but also an operating mode that enables the model to deliver higher performance — at a hefty price.
AI Power Couple Recommits: Amazon deepens Anthropic partnership with $4 billion investment
Amazon and Anthropic expanded their partnership, potentially strengthening Amazon Web Services’ AI infrastructure and lengthening the high-flying startup’s runway.
Reasoning Revealed: DeepSeek-R1, a transparent challenger to OpenAI o1
An up-and-coming Hangzhou AI lab unveiled a model that implements run-time reasoning similar to OpenAI o1 and delivers competitive performance. Unlike o1, it displays its reasoning steps.
Next-Gen Models Show Limited Gains: AI giants rethink model training strategy as scaling laws break down
Builders of large AI models have relied on the idea that bigger neural networks trained on more data and given more processing power would show steady improvements. Recent developments are challenging that idea.
Claude Controls Computers: Anthropic empowers Claude Sonnet 3.5 to operate desktop apps, but cautions remain
API commands for Claude Sonnet 3.5 enable Anthropic’s large language model to operate desktop apps much like humans do. Be cautious, though: It’s a work in progress.
AI Bromance Turns Turbulent: Microsoft and OpenAI partnership faces strain as both seek less dependence
Once hailed by OpenAI chief Sam Altman as the “best bromance in tech,” the partnership between Microsoft and OpenAI is facing challenges as both companies seek greater independence.
Enabling LLMs to Read Spreadsheets: A method to process large spreadsheets for accurate question answering
Large language models can process small spreadsheets, but very large spreadsheets often exceed their limits for input length. Researchers devised a method that processes large spreadsheets so LLMs can answer questions about them.
AI Agents for AI Research: Agentic workflow generates novel scientific research papers
While some observers argue that large language models can’t produce truly original output, new work prompted them to generate novel scientific research.