That moment when I told ChatGPT it needed a history lesson, and it agreed with me
I had an experience this week which forcefully reminded me that ChatGPT and Google’s Gemini were great but not perfect. And to be clear, I have jumped
Even as OpenAI continues clinging to its assertion that the only path to AGI lies through massive financial and energy expenditures, independent researchers are leveraging open-source technologies to match the performance of its most powerful models — and do so at a fraction of the price.
Last Friday, a unified team from Stanford University and the University of Washington announced that they had trained a math and coding-focused large language model that performs as well as OpenAI’s o1 and DeepSeek’s R1 reasoning models. It cost just $50 in cloud compute credits to build. The team reportedly used an off-the-shelf base model, then distilled Google’s Gemini 2.0 Flash Thinking Experimental model into it. The process of distilling AIs involves pulling the relevant information to complete a specific task from a larger AI model and transferring it to a smaller one.
What’s more, on Tuesday, researchers from Hugging Face released a competitor to OpenAI’s Deep Research and Google Gemini’s (also) Deep Research tools, dubbed Open Deep Research, which they developed in just 24 hours. “While powerful LLMs are now freely available in open-source, OpenAI didn’t disclose much about the agentic framework underlying Deep Research,” Hugging Face wrote in its announcement post. “So we decided to embark on a 24-hour mission to reproduce their results and open-source the needed framework along the way!” It reportedly costs an estimated $20 in cloud compute credits, and would require less than 30 minutes, to train.
Hugging Face’s model subsequently notched a 55% accuracy on the General AI Assistants (GAIA) benchmark, which is used to test the capacities of agentic AI systems. By comparison, OpenAI’s Deep Research scored between 67 – 73% accuracy, depending on the response methodologies. Granted, the 24-hour model doesn’t perform quite as well as OpenAI’s offering, but it also didn’t take billions of dollars and the energy generation capacity of a mid-sized European nation to train.
These efforts follow news from January that a team out of University of California, Berkeley’s Sky Computing Lab managed to train their Sky T1 reasoning model for around $450 in cloud compute credits. The team’s Sky-T1-32B-Preview model proved the equal of early o1-preview reasoning model release. As more of these open-source competitors to OpenAI’s industry dominance emerge, their mere existence calls into question whether the company’s plan of spending half a trillion dollars to build AI data centers and energy production facilities is really the answer.
I had an experience this week which forcefully reminded me that ChatGPT and Google’s Gemini were great but not perfect. And to be clear, I have jumped
OpenAI’s much-anticipated ChatGPT 5.0 is expected to arrive August 7 —and it could change how everyday users interact with AI. While the leap from GPT
OpenAI has introduced its 4o model into ChatGPT to enable native image generation within the chatbot atmosphere. This upgrade makes it so you don’t ha
MediaTek has today launched a new silicon for Chrome OS devices, one that puts it roughly in the same performance league as the Copilot+ PCs hawked by
ChatGPT didn’t just emerge onto the AI scene, it birthed an entire revolution of AI assistants and agents and made them accessible to consumers who we
OpenAIFollowing its U.S. debut in January, OpenAI’s Operator AI agent will soon be expanding to eight new nations, the company announced on Friday.“O
You no longer need to sign in to use ChatGPT Search.“ChatGPT search is now available to everyone on chatgpt.com,” OpenAI said in a post on X announcin
OpenAIOpenAI CEO Sam Altman announced via an X post Wednesday that the company’s o3 model is being effectively sidelined in favor of a “simplified” GP
We are a comprehensive and trusted information platform dedicated to delivering high-quality content across a wide range of topics, including society, technology, business, health, culture, and entertainment.
From breaking news to in-depth reports, we adhere to the principles of accuracy and diverse perspectives, helping readers find clarity and reliability in today’s fast-paced information landscape.
Our goal is to be a dependable source of knowledge for every reader—making information not only accessible but truly trustworthy. Looking ahead, we will continue to enhance our content and services, connecting the world and delivering value.