The Gemini’s multimodality feature allows remixing information into any format — audio, video, images, and text.
Another Google tool that has been enhanced is NotebookLM, which enables users to upload sources for research and makes the system an expert, presenting the outcome as audio Audio Overviews and Mind Maps.
Additionally, Google announced that it will soon introduce a feature that enables users to convert the content of their notebooks into educational videos.
Google continues to enhance its new search modality, called AI Mode, with advanced reasoning, multimodality, web links, the ability to ask follow-up questions, and soon, Deep Search.
In April, Google gave U.S. college students a free Gemini upgrade through 2026 final exams. The company is now expanding its offerings to students in Brazil, Indonesia, Japan, and the United Kingdom. Students in these countries will receive free access to the Google AI Pro plan for 15 months, helping them fine-tune their writing, study for exams, and get homework help, along with 2 TB of free storage, NotebookLM, and more.
Students globally will also have the ability to create custom quizzes to help them prepare for exams by simply asking Gemini to “create a practice quiz…” on any topic, or base them on uploaded documents such as class notes.
The quiz experience provides hints, offers explanations for both right and wrong answers, and provides a helpful summary at the end, highlighting areas of strength as well as those that may benefit from further study.
Later this year, the search giant will introduce Sparkify, which will turn users’ questions or ideas into short animated videos through the latest Gemini and Veo models.
With Project Astra, Google is prototyping a conversational, personalized tutor that can help with homework. The tools walk users through problems step-by-step, identify mistakes, and even generate diagrams to help explain concepts if they get stuck. This research project will be coming to Google products later this year. Android Trusted Testers can sign up for the waitlist to see a preview.
Anthropic last week launched Claude Opus 4 and Claude Sonnet 4, which offer advanced features in coding, reasoning, and AI agents.
“Claude Opus 4 is the world’s best coding model, with sustained performance on complex, long-running tasks and agent workflows. Claude Sonnet 4 is a significant upgrade to Claude Sonnet 3.7, delivering superior coding and reasoning while responding more precisely to your instructions,” advertised the company.
Both models can use tools like web search and demonstrate improved memory capabilities.
In addition, the company announced that Claude Code was generally available. It supports background tasks via GitHub Actions and native integrations with VS Code and JetBrains, displaying edits directly in users’ files for pair programming.
Anthropic, a start-up founded by ex-OpenAI researchers, released four new capabilities on the Anthropic API, enabling developers to build more powerful code execution tools, the MCP connector, Files API, and the ability to cache prompts for up to one hour.
Claude Opus 4 powers known frontier agent products like Cursor, Replit, Block, Rakuten, and Cognition.
Anthropic’s Claude 4 models arrived as the company looks to substantially grow revenue. Reportedly, the organization aims to reach $2.2 billion in earnings this year.
— Julian Goldie SEO (@JulianGoldieSEO) May 24, 2025
Still manually searching for flights and leads?
Claude 4 just changed everything.
I just turned Claude Sonnet 4 into a browser-controlling AI super agent that finds flights, scrapes LinkedIn for leads, and navigates websites automatically.
Microsoft CEO Satya Nadella unveiled the company’s biggest product moves, including Copilot and Azure updates, developer tools, and more, during this week’s Build conference in Seattle.
“We’ve entered the era of AI agents. Thanks to groundbreaking advancements in reasoning and memory, AI models are now more capable and efficient, and we’re seeing how AI systems can help us all solve problems in new ways,” said Satya Nadella.
As an example, the company mentioned that 15 million developers are already using GitHub Copilot, streamlining the way they code, check, deploy, and troubleshoot.
Hundreds of thousands of customers are using Microsoft 365 Copilot to help research, brainstorm and develop solutions, and more than 230,000 organizations — including 90% of the Fortune 500 — have already used Copilot Studio to build AI agents and automations.
The Multi-Agent Orchestration, or how multiple specialized AI agents work together to accelerate work, was one of the notorious announcements.
Today, at Build we showed you how we are building the open agentic web. It is reshaping every layer of the stack, and our goal is to help every dev build apps and agents that empower people and orgs everywhere. Here are 5 big things we announced today:
One example was an onboarding process handled by several agents: one managed human resources paperwork, another set up IT accounts, and a third scheduled training sessions. Each agent handled its part and reported back, with the system summarizing the results for the user.
In sales, one agent pulled CRM data, another drafted a proposal, and a third scheduled follow-ups in Outlook.
Microsoft announced that it is moving toward an open, connected agent ecosystem, with Model Context Protocol (MCP), a protocol that lets agents securely access data and services across platforms, and NLWeb, a system that helps websites become accessible to AI agents via natural language.
The goal is to create an environment where agents can collaborate and interact freely across the web and different apps, no matter who built them or where they run.
Also, the company presented its Agent Store, a platform where users can access and deploy various AI agents, including those built by Microsoft and third-party developers, within the Microsoft 365 Copilot and other Microsoft services. These agents are designed to perform specific tasks, such as summarizing documents, generating code, or managing meetings, and can be customized for individual needs or business requirements.
Thanks for joining us at Build, @sama! Enjoyed our conversation about SWE agents and how the role of developers is evolving. pic.twitter.com/piUVxNqTht
Great to have Jensen at Build to talk about our partnership and how we are building and scaling the largest AI supercomputer in the world on Azure. pic.twitter.com/hNgXJkVfJL
Figma, a design company, announced multiple AI features this month, during its Config annual conference. These include an AI creator for powering website and app prototypes, a way to create assets in bulk, a new drawing tool, and a blog post generator.
With this launch, the company is trying to compete with Canva, Adobe, WordPress, Squarespace, Wix, Replit, Lovable, and Hostinger, among other firms.
However, the company denies directly competing with these creative tools by saying that it is in the business of building digital products, with a third of the company’s users being developers.
Figma also announced a new plan, a content seat, starting at $8 monthly. This plan will give users access to Figma Buzz, Slides, FigJam, and Sites CMS.
Google Gemini’s Gemma open models reached a milestone of 150 million downloads this month.
Omar Sanseviero, a developer relations engineer at Google DeepMind, also revealed that developers have created more than 70,000 variants of Gemma on the AI dev platform Hugging Face.
This collection of lightweight models, built from the same research and technology that powers Gemini 2.0, was launched in February 2024 to compete with another open model like Meta’s Llama. In late April, it exceeded 1.2 billion downloads.
The latest Gemma releases are multimodal, enabling users to work with images and text in 100 languages. Versions are also fine-tuned for particular applications, like drug discovery.
Gemma and Llama have non-standard licensing terms, which some developers say they might be risky for commercial use of the models.
Gemma just passed 150 million downloads and over 70k variants on Hugging Face🚀🚀🚀
What would you like to see in the next Gemma versions?
Vibe coding tools have surged in popularity in recent months, and OpenAI is attempting to capture a share of that market segment. In early May, the maker of ChatGPT reportedly closed a deal to acquire Windsurf, the developer behind another popular AI coding platform, for $3 billion. The launch of Codex shows that OpenAI is also building its own AI coding tools.
Issued as a research preview, Codex can work on many tasks in parallel, write features, answer questions about the user’s codebase, fix bugs, and propose pull requests for review.
Each task runs on a sandboxed virtual computer in the cloud, preloaded with the repository.
According to OpenAI, it generates code that closely mirrors human style and PR preferences, adheres precisely to instructions, and can iteratively run tests until it receives a passing result.
Codex, accessible through the sidebar in ChatGPT, is powered by Codex-1, a version of OpenAI o3 AI reasoning model optimized for software engineering tasks.
OpenAI is also updating Codex CLI, the company’s recently launched open-source coding agent that runs in the user’s terminal, with a version of its o4-mini model that’s optimized for software engineering.
In OpenAI’s API, it costs $1.50 per 1M input tokens (750,000 words) and $6 per 1M output tokens.
This possibility of linking payments outside the apps, dodging Apple’s commission fees, has unexpectedly come after a landmark ruling on Wednesday in the Epic v. Apple antitrust case. Apple can no longer charge developers 27 percent fees on purchases made outside apps, and it cannot dictate how developers point users to buy things elsewhere.
A guide showed developers how to accept transactions outside an app using Stripe Checkout. Instead of showing a transaction page inside the app, it launches Safari and takes users straight to a “buy now” page.
Other apps are also beginning to launch new updates that link to the web for in-app purchases, allowing the developers to skirt Apple’s commissions.
Patreon announced yesterday that it will update its app so creators can accept payments outside Apple’s payment system.
Also, Spotify said it’s letting users subscribe to premium plans from links inside its iOS app.
Stripe 🤝iOS developers
Big news iOS app developers! You can now accept payments with @stripe outside of your app, with no iOS app store commissions.
The Stripe team 🧑🍳cooked up a quick guide walking you through how. Go live now!
Anthropic launchedIntegrations this month, a new way to connect apps and tools to its AI assistant Claude. It also announced that it has expanded its Advanced Research capability to allow this chatbot to search the web, users’ Google Workspace, and its Integrations. These tools, available in beta for paid users, are part of the company’s effort to compete with Google’s Gemini and OpenAI’s ChatGPT.
In November, Anthropic launched its MCP (Model Context Protocol), an open standard connecting AI apps to tools and data. With Integrations, Claude can work seamlessly with remote MCP servers across web and desktop apps.
Several Integrations from Anthropopic partners, such as Atlassian, Zapier, Cloudflare, Intercom, Square, and PayPal, expand what Claude can do.
For example, the Atlassian Integration lets Claude summarize and create pages in Atlassian’s Confluence workplace software.