Google’s Astra is its first AI-for-everything agent (2024)

Google is set to introduce a new system called Astra later this year and promises that it will be the most powerful, advanced type of AI assistant it’s ever launched.

The current generation of AI assistants, such as ChatGPT, can retrieve information and offer answers, but that is about it. But this year, Google is rebranding its assistants as more advanced “agents,” which it says could show reasoning, planning, and memory skills and are able to take multiple steps to execute tasks.

People will be able to use Astra through their smartphones and possibly desktop computers, but the company is exploring other options too, such as embedding it into smart glasses or other devices, Oriol Vinyals, vice president of research at Google DeepMind, told MIT Technology Review.

“We are in very early days [of AI agent development],” Google CEO Sundar Pichai said on a call ahead of Google’s I/O conference today.

“We’ve always wanted to build a universal agent that will be useful in everyday life,” said Demis Hassabis, the CEO and cofounder of Google DeepMind. “Imagine agents that can see and hear what we do, better understand the context we’re in, and respond quickly in conversation, making the pace and quality of interaction feel much more natural.” That, he says, is what Astra will be.

Google’s announcement comes a day after competitor OpenAI unveiled its own supercharged AI assistant, GPT-4o. Google DeepMind’s Astra responds to audio and video inputs, much in the same way as GPT-4o (albeit it less flirtatiously).

In a press demo, a user pointed a smartphone camera and smart glasses at things and asked Astra to explain what they were. When the person pointed the device out the window and asked “What neighborhood do you think I’m in?” the AI system was able to identify King’s Cross, London, site of Google DeepMind’s headquarters. It was also able to say that the person’s glasses were on a desk, having recorded them earlier in the interaction.

The demo showcases Google DeepMind’s vision of multimodal AI (which can handle multiple types of input—voice, video, text, and so on) working in real time, Vinyals says.

“We are very excited about, in the future, to be able to really just get closer to the user, assist the user with anything that they want,” he says. Google recently upgraded its artificial-intelligence model Gemini to process even larger amounts of data, an upgrade which helps it handle bigger documents and videos, and have longer conversations.

Tech companies are in the middle of a fierce competition over AI supremacy, and AI agents are the latest effort from Big Tech firms to show they are pushing the frontier of development. Agents also play into a narrative by many tech companies, including OpenAI and Google DeepMind, that aim to build artificial general intelligence, a highly hypothetical idea of superintelligent AI systems.

“Eventually, you’ll have this one agent that really knows you well, can do lots of things for you, and can work across multiple tasks and domains,” says Chirag Shah, a professor at the University of Washington who specializes in online search.

This vision is still aspirational. But today’s announcement should be seen as Google’s attempt to keep up with competitors. And by rushing these products out, Google can collect even more data from its over a billion users on how they are using their models and what works, Shah says.

Google is unveiling many more new AI capabilities beyond agents today. It’s going to integrate AI more deeply into Search through a new feature called AI overviews, which gather information from the internet and package them into short summaries in response to search queries. The feature, which launches today, will initially be available only in the US, with more countries to gain access later.

This will help speed up the search process and get users more specific answers to more complex, niche questions, says Felix Simon, a research fellow in AI and digital news at the Reuters Institute for Journalism. “I think that’s where Search has always struggled,” he says.

Another new feature of Google's AI Search offering is better planning. People will soon be able to ask Search to make meal and travel suggestions, for example, much like asking a travel agent to suggest restaurants and hotels. Gemini will be able to help them plan what they need to do or buy to cook recipes, and they will also be able to have conversations with the AI system, asking it to do anything from relatively mundane tasks, such as informing them about the weather forecast, to highly complex ones like helping them prepare for a job interview or an important speech.

People will also be able to interrupt Gemini midsentence and ask clarifying questions, much as in a real conversation.

In another move to one-up competitor OpenAI, Google also unveiled Veo, a new video-generating AI system. Veo is able to generate short videos and allows users more control over cinematic styles by understanding prompts like “time lapse” or “aerial shots of a landscape.”

Google has a significant advantage when it comes to training generative video models, because it owns YouTube. It’s already announced collaborations with artists such as Donald Glover and Wycleaf Jean, who are using its technology to produce their work.

Earlier this year, OpenA’s CTO, Mira Murati, fumbled when asked about whether the company’s model was trained on YouTube data. Douglas Eck, senior research director at Google DeepMind, was also vague about the training data used to create Veo when asked about by MIT Technology Review, but he said that it “may be trained on some YouTube content in accordance with our agreements with YouTube creators.”

On one hand, Google is presenting its generative AI as a tool artists can use to make stuff, but the tools likely get their ability to create that stuff by using material from existing artists, says Shah. AI companies such as Google and OpenAI have faced a slew of lawsuits by writers and artists claiming that their intellectual property has been used without consent or compensation.

“For artists it’s a double-edged sword,” says Shah.

Google’s Astra is its first AI-for-everything agent (2024)

FAQs

Google’s Astra is its first AI-for-everything agent? ›

A universal AI agent that is helpful in everyday life. Building on our Gemini models, Project Astra explores the future of AI assistants that can process multimodal information, understand the context you're in, and respond naturally in conversation.

What is Google Astra? ›

A universal AI agent that is helpful in everyday life. Building on our Gemini models, Project Astra explores the future of AI assistants that can process multimodal information, understand the context you're in, and respond naturally in conversation.

Who is the godfather of AI Google? ›

Geoffrey Hinton, known as the “godfather of AI,” is leaving his role at Google and plans to warn of the risks of the technology he's long promoted.

What AI does Google use? ›

Gemini, a multimodal model from Google DeepMind, is capable of understanding virtually any input, combining different types of information, and generating almost any output. Prompt and test Gemini in Vertex AI using text, images, video, or code.

What is the Google equivalent of ChatGPT? ›

ChatGPT took early lead among AI-generated chatbots before Google answered with Gemini, formerly Bard. While ChatGPT and Gemini perform similar tasks, there are differences.

What is the Google AI assistant? ›

Interaction. Google Assistant, in the nature and manner of Google Now, can search the Internet, schedule events and alarms, adjust hardware settings on the user's device, and show information from the user's Google account.

What does Astra Space do? ›

Astra empowers its customers to Improve Life on Earth from Space® by providing dedicated small launches and mission-critical on-orbit mobility to smallsat builders and operators around the world.

Which AI is better than Google? ›

GenAI has enabled new search engine platforms with unique features and advantages, challenging Google's dominance.

What's the AI everyone is using? ›

As a leader in the AI space, Google Assistant is considered to be one of the most advanced virtual assistants of its kind on the market. Using natural language processing, it supports both voice and text commands, and can handle everything from internet searches to voice-activated control of other devices.

What is the new AI version of Google? ›

Now, with generative AI, Search can do more than you ever imagined. So you can ask whatever's on your mind or whatever you need to get done — from researching to planning to brainstorming — and Google will take care of the legwork. This is all made possible by a new Gemini model customized for Google Search.

Which AI is better than ChatGPT? ›

The 15 Best ChatGPT Alternatives to Try in 2024
AI assistantBest for
ClaudeNuanced AI conversations
GitHub CopilotCoding assistance
CodeiumCode completion and generation
Cohere GenerateCustomizable AI models
11 more rows
May 13, 2024

Is Google Bard better than ChatGPT? ›

ChatGPT is better than Bard at generating code

ChatGPT. is trained on a massive dataset of text and code, including a large portion of code from GitHub repositories. While also trained on a massive dataset, Bard's focus is broader, incorporating a wider range of text and information.

Which is the best AI chatbot? ›

Best AI chatbot for content marketing

Jasper software is an AI copilot designed for marketing teams to have a central system for all content creation. Its own AI chatbot, Jasper Chat, quickly generates useful, applicable, and unique content.

What virtual assistant does Google use? ›

Google Assistant, your own personal Google.

What is Astra Finance? ›

About us. The all-in-one platform for lightning-fast payments. Astra provides the infrastructure for moving money instantly and securely. Astra's proprietary, vertically integrated platform allows developers to easily embed real-time payments into their products.

What does Google app Launcher do? ›

Using this app launcher you can have access to all your installed applications from within a toolbar popup window. The panel enables you to sort your apps the way you want them, and you can directly access the following services: 1. Launch an application directly from the popup view 2.

What does Google services for AR do? ›

Google Play Services for AR is automatically installed and updated on supported devices. This service unlocks augmented reality (AR) experiences built using ARCore. Automatic updates ensure that apps with AR functionality work without requiring an additional download. This service was previously known as ARCore.

References

Top Articles
Latest Posts
Article information

Author: Eusebia Nader

Last Updated:

Views: 6021

Rating: 5 / 5 (60 voted)

Reviews: 91% of readers found this page helpful

Author information

Name: Eusebia Nader

Birthday: 1994-11-11

Address: Apt. 721 977 Ebert Meadows, Jereville, GA 73618-6603

Phone: +2316203969400

Job: International Farming Consultant

Hobby: Reading, Photography, Shooting, Singing, Magic, Kayaking, Mushroom hunting

Introduction: My name is Eusebia Nader, I am a encouraging, brainy, lively, nice, famous, healthy, clever person who loves writing and wants to share my knowledge and understanding with you.