Machine Learning Research
OpenAI’s GPT-4.5 Goes Big: OpenAI releases GPT-4.5, its most powerful non-reasoning model and maybe its last
OpenAI launched GPT-4.5, which may be its last non-reasoning model.
Machine Learning Research
OpenAI launched GPT-4.5, which may be its last non-reasoning model.
Machine Learning Research
Typical large language models are autoregressive, predicting the next token, one at a time, from left to right. A new model hones all text tokens at once.
Machine Learning Research
Although large language models can improve their performance by generating a chain of thought (CoT) — intermediate text tokens that break down the process of responding to a prompt into a series of steps.
Business
Elon Musk and a group of investors made an unsolicited bid to buy the assets of the nonprofit that controls OpenAI, complicating the AI powerhouse’s future plans.
Machine Learning Research
Replit, an AI-driven integrated development environment, updated its mobile app to generate further mobile apps to order.
Machine Learning Research
xAI’s new model family suggests that devoting more computation to training remains a viable path to building more capable AI.
Machine Learning Research
While Hangzhou’s DeepSeek flexed its muscles, Chinese tech giant Alibaba vied for the spotlight with new open vision-language models.
Machine Learning Research
OpenAI introduced a state-of-the-art agent that produces research reports by scouring the web and reasoning over what it finds.
Machine Learning Research
Google updated the December-vintage reasoning model Gemini 2.0 Flash Thinking and other Flash models, gaining ground on OpenAI o1 and DeepSeek-R1.
Machine Learning Research
As Anthropic, Google, OpenAI, and others roll out agents that are capable of computer use, new work shows how underlying models can be trained to do this.
Machine Learning Research
OpenAI introduced a successor to its o1 models that’s faster, less expensive, and especially strong in coding, math, and science.
Machine Learning Research
The practice of fine-tuning models on synthetic data is becoming well established. But synthetic training data, even if it represents the training task well, may include characteristics like toxicity that impart unwelcome properties in the trained model’s output...