Perplexity has announced that they are about to release Comet, a brand new browser experience that they refer to as an ‘agentic browser’. This is a big bold move from Perplexity. They’ve already had some impressive search tools available in their models, but this is something different, a standalone offering that is looking to disrupt the search market in a meaningful way.
Elon Musk was feeling bold with his predictions today when he posted on X “We are on the event horizon of the Singularity” Obviously no one really knows what this is in reference to, but Grok 3 has been on fire recently, so maybe he’s seeing something in their research.
Deepseek R2 had been slated for launch in May, but the success of some of their competitors recent models has put the pressure on. The release has now moved from May to “as soon as possible”. Will DeepSeek R2 have the same huge impact as their R1 model did?
AI startup Convergence made headlines in recent weeks by launching a rival to Operator called Proxy. It was very well received and made some instant fans. Now they’re back with Proxy Lite, a 3 billion parameter vision language model that’s open source and can run on your local machine.
OpenAI announced that Deep Research would now be available to all of its Plus subscribers, and that a version of Advanced Voice powered by GPT4o-mini would be available to all ChatGPT users, even to those on the free tier. OpenAI are definitely feeling increased pressure to ship more often, and that’s good for the users.
Perplexity has updated their iOS app with a new voice mode. Perplexity claims that this is the only model that can reliably incorporate real time voice and information across several languages simultaneously.
Figure have revealed a new video of their figure 02 robots working together with their new Helix Neural Network. Last time we saw two robots coordinating to carry out some simple home tasks. In this video there are many robots working in an industrial setting carrying out some picking and sorting tasks that would usually be done by human hand. Figure claim that training their second commercial use case took just 30 days.
The little known AI company Inception Labs has unveiled a whole new kind of AI model. Their Mercury model is being called a DLLM or Diffusion Large Language model. This means that the output is created and then incrementally refined from low to high fidelity. The result is that Mercury can generate over 1,000 tokens per second, far higher than any current model. This feels like a major breakthrough in model architecture and it will be very interesting to see if other companies adopt this approach.
Pika has released version 2.2 of their video model. It can generate 10 second clips with a 1080p resolution and it now includes Pikaframes where the user can select the start and end frames of a video for far greater control.
Prior to the release of ChatGPT 4.5, we predicted that OpenAI would have to launch something pretty special in order to stave off the criticism that it has faced recently. Either they would need to launch something groundbreaking at the $200 pro level to justify the price tag, or they would need to launch a great model that provided a decent step up at the free or plus tier. It would seem that OpenAI delivered neither of these things.
What we got was a model that struggles to stand out from the crowd. We’ve seen very few benchmarks where it’s managed to do anything significant, and yet it’s only available to users paying the hefty Pro subscription. This has left most users feeling underwhelmed and this really feels like the first time where OpenAI have not been able to display their quality and maintain a dominant position in the market.
Grok, on the other hand, does continue to make some headlines. There are strong rumours that XAI will introduce artefacts to the interface very soon so that users can examine and adjust the code outputs that Grok generates. This will add a whole new dimension to the usability of Grok and is likely to be very well received.
Google’s NotebookLM is said to have a mind map feature coming soon. This will allow users to see a mind map style output based on the sources being given to the notebook and the user will be able to interact and prompt questions on item in the map. We’re fans of NotebookLM so we’re looking forward to this one.
Tuesday February 25
DeepSeek, a Chinese AI startup, has decided to expedite the release of its R2 artificial intelligence model from a planned May launch to as soon as possible, driven by competitive pressures within the AI sector. This decision reflects a broader trend where major Chinese tech companies like Alibaba, Tencent, and ByteDance are increasing their orders for Nvidia’s H20 AI chip to leverage DeepSeek’s cost-effective AI models. The R2 model is expected to enhance capabilities in coding and multilingual reasoning, showcasing the intense pace of AI development and market competition in China.
DeepSeek triggered a $1 trillion-plus sell-off in global equities markets last month with a cut-price AI reasoning model that outperformed Western competitors. The firm is now accelerating the launch of January's R1 model's successor, according to sources https://t.co/aw2mGa3DeG
— Reuters (@Reuters) February 25, 2025
Convergence AI has released Proxy Lite, a 3B parameter Vision Language Model, which is claimed to be the first small open-weights model excelling in UI navigation tasks. This model is open-source and designed to operate efficiently on local machines, potentially broadening access to advanced AI technologies. The release has been met with excitement on social media, with various users highlighting its capabilities and ease of use, although these claims require further validation for complete accuracy.
Today we are pleased to announce the release of proxy-lite-3b, the first in a series of open source releases happening over the coming weeks.
— Convergence (@convergence_ai_) February 25, 2025
proxy-lite-3b is the first small open weights model that performs well on ui navigation tasks – beating all other open source models so… pic.twitter.com/LVNIXKo9Fg
Google has launched a free version of Gemini Code Assist, an AI-powered coding tool, for individual developers worldwide. This tool provides up to 180,000 code completions per month, supports all programming languages in the public domain, and features a 128,000-token context window. It integrates with popular IDEs like Visual Studio Code, offering functionalities such as code assistance and review. Gemini Code Assist is designed to compete with GitHub Copilot, providing significantly higher usage limits for free users.
https://twitter.com/GoogleDeepMind/status/1894349711160578093
https://blog.google/technology/developers/gemini-code-assist-free
Monday February 24
Perplexity has announced that they are about to release ‘Comet’, a brand new browser experience that they refer to as an ‘agentic browser’.
Comet: A Browser for Agentic Search by Perplexity
— Perplexity (@perplexity_ai) February 24, 2025
Coming soon. pic.twitter.com/SwVSwudgtN
Anthropic has released their latest model, Claude 3.7 Sonnet, their most intelligent model to date and the first Claude model to offer extended thinking this year.
Introducing Claude 3.7 Sonnet: our most intelligent model to date. It's a hybrid reasoning model, producing near-instant responses or extended, step-by-step thinking.
— Anthropic (@AnthropicAI) February 24, 2025
One model, two ways to think.
We’re also releasing an agentic coding tool: Claude Code. pic.twitter.com/jt7qQmFWuC
Luma Labs are the latest generative AI player to build sound creation directly into their video creation workflows. If we know Luma Labs, then we know this is likely to be a very high quality offering.
Video to Audio is now here in #DreamMachine. To generate sound for your video generations, just select the new "Audio" button. Create with a single click or describe with prompts for more customized direction. Audio is available now in beta for free to all users. pic.twitter.com/DyHt8meeO8
— Luma AI (@LumaLabsAI) February 24, 2025
Elon Musk was feeling bold with his predictions today when he posted on X “We are on the event horizon of the singularity.”
We are on the event horizon of the singularity
— Elon Musk (@elonmusk) February 23, 2025