Google I/O 2024 announcements — all the big Gemini AI and Android 15 news


a photo of the Apple Watch 9 and Google Pixel Watch 2

(Image credit: Future)

The Wear OS 5 announcement by Google today details some of the new health metrics coming to Google’s smartwatch software. They’ve also been things that rival devices — namely the Apple Watch — have already been able to track. Senior writer for fitness Dan Bracaglia takes a look at how Google is trying to catch up to Apple and Garmin with Wear OS 5. 

Watching google cast of NBA app on Rivian R1S infotainment screen

(Image credit: Future)

While I was at Google I/O on Tuesday, I got the chance to sit in a Rivian R1S and watch videos on the EV’s 15.6-inch infotainment system. No, I wasn’t slacking off — rather, it was a demo of a newly announced Android feature that’s going to add Google Cast support to cars with Android Automotive.

The idea is that when you’re at a charging station and topping off the battery for your Rivian vehicle, you’ve got a lot of sitting around and waiting to do. So why not spend the time watching something? This Android update means you can now beam any of the 3,000 or so cast-enabled apps to the screen in your car.

Rivian will be the first automaker to gain this support, but others are coming, Google says.

AI created video. Blue balloon man dancing in a parking lot

(Image credit: Google)

One of the multitude of AI announcements from Google this week was Veo, an AI model that can generate high-def video using text prompts. A standout feature in Veo is its apparent ability to understand cinematic terms, so that you can tell it to do a time-lapse video or use an aerial shot and Veo will comply.

Google says that Veo can generate more than 1 minute of video footage, though how much more than 1 minute is unclear. We do have some Google Veo samples, though, to give us some idea of what the videos look like and how they compare to OpenAI’s Sora video generator.

Google Pixel Watch 2

(Image credit: Future)

Wednesday’s other big software announcement from Google involved Wear OS 5, the software that supports a number of smartwatches, including Google’s own Pixel Watches. The story here is power efficiency — as in, that’s what Wear OS 5 offers. Run a marathon with the new software, Google says, and you’ll have 20% more battery left compared to if you were using a Wear OS 4-powered watch.

Wear OS 5 also adds some metrics including Ground Contact Time, Stride Length, Vertical Oscillation, and Vertical Ratio. This is not exactly ground-breaking since, as my colleague Kate Kozuch points out, the Apple Watch has been able to monitor such things since 2022.

There’s one other change to Wear OS 5 — it introduces a new version of Watch Face Format, which you use for customization purposes when creating a watch face.

Android mascot in a triangle flexing his arms as an android 15 promo

(Image credit: Googlw)

Philip Michaels here, back from a day spent in Google’s company at I/O. Day 2 is a much more low-key affair, so I’m watching from afar. And one of the first things Google did today was release Android 15 beta 2.

What are the highlights of this new version? My colleague Tom Pritchard found 7 Android 15 changes to get excited about, such as Private Spaces for tucking away sensitive apps and the ability to added passes to Google Wallet with just a photo. There’s also some security features including Theft Detection Lock, which taps into AI to determine if someone’s swiped your phone. Because let no update at Google take place these days if it doesn’t include AI.

What the competition is doing…

Sam Altman CEO of OpenAI

(Image credit: Getty Images)

The reason Google is going so hard on AI surely has something to do with the looming spectre of OpenAI and ChatGPT. In case you missed it, OpenAI held its own event earlier in the week to reveal what it’s been working on.

Headline news is the introduction of GPT-4o, a new model that’s both much faster and more efficient than the current GPT-4. What’s more, OpenAI is rolling it out to its free tier of users, meaning they’ll get custom chatbots and access to the ChatGPT store which has models and tools built by users. All that without having to hand over $20 for the premium tier.

Given that Google built much of its empire on “free” tools like Photos, Docs and Drive — not to mention Search and YouTube — the news from OpenAI must have rattled some cages over at Mountain View.

Gemini Live vs ChatGPT-4o

Google Gemini vs GPT 4o

(Image credit: Google / Future)

Gemini Live was a big deal during the I/O keynote and clearly meant as a rival to OpenAI’s ChatGPT Voice. So, how do the two compare?

Both offer a conversational, natural language voice interface, both offer the potential for live video analysis through a smartphone camera and both seem to be fast enough for a truly natural conversation where you can interrupt the AI mid-flow.

However, ChatGPT Voice sounds more natural, can detect and respond to emotion and vocal tones and even adapt in real-time to how you ask it to speak. Our early verdict is that Google has a bit of catching up to do.

Soccer coaching courtesy of Gemini

AI-based soccer instruction demo at Google I/O 2024

(Image credit: Future)

My colleague, and Tom’s Guide phone editor, Philip Michaels was at the I/O keynote yesterday. And while there wasn’t any actual phone hardware to get his hands on, Phil did get a chance to enjoy a spot of AI-based coaching when it came to his soccer skills.

As Phil tells it: “My favorite demo at Google I/O allowed me to kick a soccer ball in anger and use AI analysis to pinpoint where my wayward penalty kicks had gone so wrong. I had three attempts to kick the ball at goal while cameras measured my form to produce scores for power, accuracy and style.”

The impressive part was the data wasn’t just fed back to him in numbers and graphs. Google used Gemini 1.5 Pro to interpret the data so an AI coach could vocalise the feedback directly, letting Phil know he probably shouldn’t give up the day job.

Can AI save us from email?

Gmail app on iPhone in woman's hand

(Image credit: Shutterstock)

It’s fair to say, I rather dislike dealing with emails so any help I can get in sorting through the vast amounts of the things I receive on a daily basis is a win. The AI upgrades Google has planned for Gmail are, therefore, my cup of tea.

During the keynote, Google showed a new Summarize Emails feature that I can’t wait to get using. It lets you ask for something specific within the context of a long and probably boring email and, as the name suggests, will summarize those in neat and tidy bullet points.


♬ original sound – Tom’s Guide

Get Google’s “most capable AI model” for free

Google Gemini

(Image credit: Google)

Everyone likes something for free, right? Well, Google is currently offering a two-month free trial of Gemini Advanced, the premium version of its Gemini chatbot. 

This gives you access to the newly-announced Gemini 1.5 Pro model as well as a few extra goodies like 2TB of storage for Google One and access to Gemini in Workspace. The latter means Gemini can help you craft emails or give your written reports in Docs a little extra polish.

If you’re interested in giving it a try, the sign-up page is here. Although it’s not yet available beyond the U.S. — so it’s no bueno if you attempt to sign up from London, like I just tried to do. 

Playing with Project Astra

Project Astra

(Image credit: Future)

Project Astra was one of the eye-catching announcements from I/O and is, essentially, a real-time, camera-based AI that can do anything from identify an object in frame to craft an fictional story about said object.

While a public version is still a little way off, Managing Editor Kate Kozuch got to try a demo of the technology in person at I/O. For the purpose of the demo, Google hooked up a stationary top-down camera to a machine running Gemini 1.5.

When presented with an array of dinosaur figurines, Gemini not only named each’s classification but came up with names and adventurous storylines that seemed surprisingly suitable.

♬ original sound – Tom’s Guide

The Next Google Glass?!

Google Project Astra Glasses

(Image credit: Google)

Was that, Google Glass? No, but Google did show off a pair of smart glasses as part of its demo of Project Astra. This new voice assistant, which is coming to Gemini Live on phones, can see the world around you and can do everything from identify landmarks to help solve equations. But a video demo during I/O showed a big surprise. This tech is coming to other form factors, including smart glasses. 

Google I/O 2024 wrap

(Image credit: Google)

And that’s a wrap folks! Unlike past Google I/O keynotes, there were no hardware announcements of any kind. Instead, the focus was strictly around Gemini and how this AI model is being integrated into more of Google’s services.

AI ethics

(Image credit: Google)

Google is ensuring that its AI services will follow its principles to protect and safeguard how they’re used. There will also be various watermarks stamped into images, audio, and video that are generated with the help of generative AI.

With learning and education, LearnLM is a new model that will make these AI experience more personal and engaging.

Pre-made Gems, like Learning Coach, will provide step-by-step guidance when it comes to teaching. Rather than giving you the direct answer to your question, Learning Coach will use guidance to help you understand how to get the answer.

Token pricing

(Image credit: Google)

Google’s Josh Woodward details Gemini 1.5 Pro and Flash’s pricing. Gemini 1.5 Flash is priced at 35 cents for 1 million tokens, which is very cheap compared to GPT-4o’s rate of  $5 per 1 million tokens.

Gemini gets new features on Android

Gemini integrated search in YouTube videos

(Image credit: Google)

Gemini on Android is more context aware to provide helpful actions. It’s also overlayed on top of whatever app you’re using, so you don’t have to switch back and forth. There’s also a neat feature that lets you drop and drag images from the Gemini app to another one.

Another demo shows how Gemini can analyze a long PDF document for even better responses, rather than the user searching through the entire document to find an answer to their question.

More AI coming to your Android phone

Making smartphones truly smart

(Image credit: Google)

Google’s Sameer Samat is on stage to talk more about Android with AI at the core with three breakthroughs coming this year: better searching on your Android, Gemini is becoming you AI assistant, and on-device AI will unlock new experiences.

Circle to Search will be broadly rolled out to more Android phones, no longer limited to Samsung’s Galaxy and Google’s Pixel phones.

Gemini is becoming a personal assistant

Planning trips in Gemini

(Image credit: Google)

Trip planning experience with Gemini advance can plan everything with search, shopping, and maps. It can generate an itinerary based on what you’re looking to do.

Gemini AI Teammate shows how AI can take on a job

Gemini AI Teammate detailed

(Image credit: Google)

Virtual Gemini teammate can be used for collaboration among different people in Google Chat, much like having a new member on your team to track expenses and schedules. You can even give this virtual Gemini teammate its own name.

Gemini integrates more with Gmail

Gemini summarizing emails close up

(Image credit: Google)

Gmail for mobile is getting a boost with Gemini that will overlay on top of emails when it’s initiated on your phone. Another example shows how Gemini can search your email and even find shopping comparison for a roof repair. Gmail’s smart replies will include responses that populate Gemini’s recommendations, complete with links and prices to services that offer roof repairs.

Gemini can do even more complex tasks, like taking receipts you might have in different emails and can create a spreadsheet that organizes them in one place. Rather than tracking all of those emails and receipts, Gemini is doing all the work for you.

In this example of a receipts tracker, Gemini can also analyze the data and uncover things like where’s the money being spent.

♬ original sound – Tom’s Guide

Here’s how Google Search is getting a big upgrade with Gemini.

Ask Google with video search

Ask in video live demo

(Image credit: Google)

Google’s Rose Yao demos how she can fix a broken record player with the help of searching with video. She shows an example of how she’s recording video and asks why the record player isn’t working correctly. Google Search is able to search frame by frame to answer questions.

Google Search in the Gemini era

Search in the Gemini Era

(Image credit: Google)

AI Overviews will make Google Search better with helpful search results. One way the company is going to do this is with Multi-step reasoning, which will let Google do all the research for you. For example, Gemini can research a yoga studio based on reviews.

Additionally, Gemini in Google Search can plan for meals by aggregating all your meals for a day with recipes for all the dishes. If making them yourself is too much, Google Search can also help you find where you can buy those meals with Gemini’s help.

Your search result page will also change with Gemini’s help, like finding restaurants that have live music and it can even pull recommendations based on the season — like restaurants with rooftops.

ML Compute Demand requires new hardware

Trillium TPU

(Image credit: Google)

Trilliium is the next-generation TPI, which will be available to Cloud customers in late 2024.

Axion Processor is a custom ARM-based CPU.

AI Hypercomputer is a groundbreaking supercomputer architecture, which is made possible with liquid cooling at Google’s data centers.

Google is aiming to invest even more into the future with these announcements.

Veo is Google’s generative video model

Veo AI video generator

(Image credit: Google)

Veo is Google’s generative video model, which creates 1080p video based on prompts. It can be used in a tool called VideoFX, giving video editors and creators a new way of making video. Filmmaker Donald Glover has used Veo to produce videos that will be coming soon. Generating video from scratch is helping to advance the frontier of AI.

Wycleaf Jean using Google AI for music

(Image credit: Google)

When it comes to music, Music AI Sandbox is a YouTube tool to help creators make more music by mixing different styles and creating something original.

Imagen 3

(Image credit: Google)

Bringing creative ideas to live with new generative media models for photos and video. Google’s Doug Eck explains the new image model that’s been built from the ground up, Imagen 3, the company’s most capable image model. It’ll use generative AI to produce even more realistic, detailed images. It can also be used to render text.

Project Astra uses video for Gemini analysis

Project Astra AI agent

(Image credit: Google)

Project Astra aims to use video to better deliver relevant answers. For example, you’ll be able to use your phone’s camera to ask what it is you’re looking at, including being able to analyze code and have it make it relevant changes. Additionally, it’s intelligent enough to use context clues visually in video to even help you locate your belongings if you forget where you left it.

Making AI helpful for everyone

Gemini, Chrome and Agents working together

(Image credit: Google)

Gemini can be used for searching your phone, helping you to find receipts, scheduling a pickup window, and much more. With travel, Gemini can be used to search useful and fun activities if you’re planning a trip. Pichai says that Google is “Making AI helpful for everyone.”

♬ original sound – Tom’s Guide

Wondering how Gemini will make Google Photos better? Check out how it can offer better search results with the help of Gemini.

Google Workspace gets help with Gemini

Gemini summarizing emails

(Image credit: Google)

Gemini’s presence is being integrated into many of Google’s Workspace, like Gmail, which you can use Gemini to search emails, get highlights from a Google Meeting, and much more.

Google Photos just got a big AI upgrade with Gemini

Using Gemini in Google Photos

(Image credit: Google)

Google Photos is getting help with the power of Gemini to get more relevant searches, including how it can search and select photos based on contextual responses with a new feature called Ask Photos with Gemini.

Gemini Era

(Image credit: Google)

Sundar Pichai is highlighting all of Google’s work around Gemini. “We want everyone to benefit from what Gemini can do,” said Pichai. He also teases how Gemini will be intertwined into many of Google’s services.

People are using Google Search in more ways than ever before, with even longer keyword search. Also,  AI Overviews is rolling out to everyone in the US and coming soon worldwide.

Google I/O 2024 keynote has begun

Sundar Pichai wide shot

(Image credit: Google)

Google I/O 2024 has officially started and the first order of business appears to center around artificial intelligence. A quick video right at the start highlights many of Google’s work around Gemini, with Sundar Pichai the first to come out on stage.

DJ making AI music at Google I/O 2024

(Image credit: Google)

For a bit more context on this rather bizarre warm-up act, the DJ Mode in MusicFX gives you really granular control over the AI generated sound, reacting in real time to changes rather than generating a new sound every time.

MusicFX can’t generate songs with vocals and refuses any request to include tracks mentioning an artist. But other than that it can generate backing tracks for a video, music for a game or a just something to fill the space before Sundar Pichai walks on…

The pre-show has started

Google I/O venue

(Image credit: Future)

Well, the pre-show has started courtesy of a guy in a multicoloured bathrobe and some “pulsating” music. I’m not really sure what to make of this but I hope the folks in the audience are entertained.

He’s currently DJing with Google’s MusicFX artificial intelligence experiment that lets you create Live Loops and experiment with sound. Perhaps Google is coming after Suno and Udio when it comes to creating the best AI music generator… 

Gemini’s memory feature

The Google I/O sign

(Image credit: Future)

One Gemini feature I think could sneak into the keynote today is “Memory”. This was actually first rumored last year (back when Gemini was still called Bard) but news has been quiet of late. In a nutshell, the feature lets you save facts (or other things) you want Gemini to remember which can then factor into future conversations.

According to a post on X by Dylan Roussel over the weekend, the feature could be launching “in the next few days” with a new UI to boot. This would probably be the likely time for Google to provide us with an update. 

Dino crisis

A screenshot of the Chrome Dinosaur game

(Image credit: Future)

Google’s also currently promoting a generative AI version of the Chrome dinosaur game where you can swap your character and the obstacles for different things.

You can try it out for yourself here, but it looks like it’ll only be available for another 30 minutes or so. If you hadn’t twigged yet, there’s going to be a lot of talk about AI today… 

Not long now!

Google I/O 2024

(Image credit: Future)

People are taking their seats as the Google I/O keynote is scheduled to kick off in less than an hour!

You may be interested to know that while Wi-Fi is spotty for those of us in attendance (always ironic at a tech event), we are being treated to a soundtrack that includes Taylor Swift. So it’s all good.

Gemini’s multimodal features

Google Pixel 8a in hand.

(Image credit: Future)

Okay, that’s enough deviation from what’s going to be the real highlight of today’s show: AI. Last year’s I/O had its fair share of AI announcements but 2024 is going to turn it up to 11 — probably through more explanation of multimodal features. That means exploring what Gemini can do when it’s shown video, music or pictures.

A demo posted by Google yesterday showed this in action when a Pixel phone correctly identified the I/O stage being set up. It’s going to be like Google Lens on steroids and will surely offer up all manner of assistance based on what’s going on in your camera’s view. Imagine being able to record live video and having Gemini provide a flawless transcript or pull out relevant information based on what it’s seeing. I expect we’ll hear a lot more about this kind of functionality as the event progresses.

Desktop mode on Android?

This isn’t something I’m really expecting to hear much about today, but it’s cool so I’m going to talk about it here.

One feature on the best Samsung phones that doesn’t often get talked about is DeX (pictured above), which turns the device into a pseudo desktop via a proprietary dock. It seems like Google is exploring this with the Pixel as, according to a report (h/t Android Authority), the company held a private event showing a Pixel 8 running Chromium OS, the open source version of ChromeOS, on an external monitor.

While Google isn’t going to merge Android and Chrome anytime soon, ChromiumOS can run alongside Android in a virtual machine, thanks to the Android Virtualization Framework (AVF), which debuted in Android 13.

This feature seems a little too far off to make headline news at today’s keynote but as someone always looking to streamline their kit, I’m intrigued to see if this gets a mention when Google unveils more about Android 15 today.    

Boots on the ground

Google I/O 2024 welcome sign in Mountain View

(Image credit: Future)

Google I/O 2024 is officially kicking off in a couple of hours and, as you can see from the image above, we’re ready and waiting in Mountain View to get started. Managing Editor Kate Kozuch will be live at the event to capture reaction to Google’s announcements as they unfold.

A quick reminder the keynote gets underway at 1 p.m. ET / 10 a.m. PT / 6 p.m. BST and the leading topic will almost certainly be Gemini and Google’s latest AI developments — aiming to steal the thunder from all the wonderful new features OpenAI announced at its own Spring Update yesterday.

Google Maps upgrades!

Google Maps on iPhone

(Image credit: Shutterstock)

Shifting gears to Google Maps for a second, ahead of Google I/O two new Maps upgrades are starting to go live. The first new feature lets you apply a filter to show the charging stations of electric cars instead of gas stations. And the other allows the map to show the entire length of a road and highlight it in blue when you search for it. This will give you a better sense of the layout for sure.

Maps is a big part of Google’s product portfolio and usually commands some good stage time. We’ll see if there are any other new features waiting in the wings. In the meantime, here’s 9 essential Maps tricks for your summer road trip.

Google Wear OS

(Image credit: Shutterstock)

Hi folks, it’s Jeff Parsons taking over this live blog for a spell as we look ahead to what Google may show us over in Mountain View later today. One area I’d be interested to hear about is Wear OS. Google announced Wear OS 4 at last year’s I/O and it brought with it a few nice-to-have upgrades like better battery life and some new watch face tools. But, overall, it feels Google’s been coasting a little bit here.

While it’s a fair bet we may hear about Wear OS 5 at today’s announcement, I’m keen to hear more about exactly how Google can differentiate from watchOS 10 — which had a redesigned home screen experience driven by widgets and Smart Stacks when it was announced in September. Will we get Gemini integration? Will we learn more about Android Health, the replacement for Google Fit? Will Fitbit continue to be subsumed into the Google brand? Bingo cards at the ready.

Google Messages logo on a smartphone screen in front of the Google branding

(Image credit: Getty Images)

A recent post on X has shared two screenshots that seemingly reveal that Google will allow users to edit messages on their app, even after they’ve been sent.

According to the post, the feature will function similarly to WhatsApp and iMessage, with a 15-minute grace period to edit a sent message. It appears that some users have been allowed to test this new feature, although there is no indication of when it will officially be announced. It could be that we will hear more at Google I/O as one of its new features.

Google Pixel Fold 2

(Image credit: SmartPrix)

Google I/O 2023 was when we first got to see the Google Pixel Fold, the company’s first ever foldable phone. So naturally we’re wondering if we’ll see the sequel announced, or at least teased, at this year’s I/O.

There are some odd rumors that the Pixel Fold 2, as we assumed it would be named, will actually be titled the Pixel 9 Pro Fold to tie it in with Google’s other flagship Pixels. Otherwise, we’re expecting a refined version of last year’s foldable Pixel, perhaps with some new unique design elements and likely some upgraded internals too.

Pixel 9 leaked image

(Image credit: @OnLeaks)

Google doesn’t do it consistently, but we may see a teaser for the Pixel 9 series today, as a token gesture toward the company’s hardware products among a keynote that’s likely to be all about software.

The Pixel 9 series is believed to consist of three models this year, featuring a new smaller Pixel 9 Pro model alongside a Pro XL model and the standard version. All the phones should be running a new Tensor G4 chip, and will likely be packed with lots of new AI features to build on those added to the Pixel 8 series last year.

Apple iPhone 15 review

(Image credit: Tom’s Guide)

Circle to Search is not just a Galaxy S24, Pixel 8 and Pixel 8a feature. You can also use it on your iPhone (kinda).

With a new shortcut, it’s possible to instantly take a screen shot and search it through the iOS Google app. It’s likely as close as iPhone users will ever get to sampling Circle to Search, at least for the time being while Google’s being very particular about who gets to use it.

and image of the Google Chrome logo on a laptop

(Image credit: Shutterstock)

Everyone’s favorite browser Google Chrome may get some attention today. And we may already know one upcoming feature for it.

Circle to Search, currently available on only a select few Android phones, could be appearing on Chrome via a Google Lens update. This would allow far more users to try out this handy search method, although given it’s seemingly not finished yet, maybe Google needs some more time before it starts rolling out.

Google Gemini logo

(Image credit: Google)

Our AI expert Ryan Morrison had some predictions about what AI announcements Google will have later today.

He believes we’ll see new features come to the Gemini large language model and chatbots, both the developer-only and public versions, Gemini-based features in familiar Google apps like YouTube Music and Google Docs, and perhaps a demo of some advanced tech to show it can keep up with OpenAI and GPT4o.

Google IO 2021

(Image credit: Google)

Google’s Project Starline is a very fancy-looking video conferencing tool that the company’s been demoing for some time. But it recently announced that it’ll finally launch in 2025, with HP getting first dibs on the tech, so we expect it to be at least mentioned briefly today at I/O.

The basics of Starline are that it works just like a normal video call, but with a full 3D model of the person you’re talking to that in theory allows for more natural conversation. The hardware required has gradually shrunk since Starline debuted, so hopefully it’ll soon be small enough to fit in a dedicated camera or within another device, rather than requiring specialized multi-camera monitors or even dedicated rooms to use.

GPT-40 closeup

(Image credit: OpenAI)

OpenAI clearly had its sights set on stealing some of the pre-Google I/O thunder by holding an event of its own Monday (May 13). And it’s easy to steer some of the attention away from Google when your announcements are as significant as what OpenAI had to say.

The highlights included:

The GPT-4o model is particularly significant in that it can analyze image, video and speech. It’s almost enough to make up for the fact that the event came and went without any sign of ChatGPT-5.

Maybe Google was always planning to tease some AI-powered feature ahead of Google I/O. Or maybe it was all those Open AI announcements 24 hours ahead of I/O that did the trick. But Google spent Monday afternoon showing off a new AI capability of the camera app on what appears to be a Pixel.

Using voice prompts, you would ask the camera what it sees in a conversational tone. And the AI can reply, describing what’s in its view finder with impressive accuracy. Watch the posted footage from Google to see for yourself.

Google Pixel Fold shown in hand

(Image credit: Tom’s Guide)

At last year’s Google I/O, the Pixel Fold made its debut giving Google a foldable device that could take on the leading foldable from Samsung. Indeed, in our Pixel Fold review, we praised its thin, durable design and wide cover display. (A feature Samsung may adopt for the upcoming Galaxy Z Fold 6, incidentally.)

We don’t expect to see the Pixel Fold 2 previewed at this year’s Google I/O, and not just because Google appears to want to keep the focus on AI. Rather, it’s because of a rumor that the next Pixel Fold might get folded into the Pixel 9 lineup and rebranded as the Pixel 9 Pro Fold. If so, that means a fall launch for the foldable follow-up.

android tracker deetction screenshots

(Image credit: Google)

Google had some news to announce in conjunction with Apple ahead of Google I/O. The two tech giants have teamed up to boost cross-platform protections against someone trying to monitor your location with a Bluetooth tracker. 

Should you be on the move with an unknown tracker along for the ride, you’re going to get an alert on your phone about that device, regardless of the platform its paired with. Previously, iPhone users would see alerts for Apple AirTags that were monitoring them without their knowledge, but not necessarily other trackers paired to Android devices.

The feature’s included on the just-released iOS 17.5 update for iPhones, while Google’s adding support for any Android device running Android 6 or later.

Google Pixel 8a.

(Image credit: Future)

We expected to see the Pixel 8a debut at Google I/O today, but Google had other plans in mind. Instead, it announced the Pixel 8a last week — no doubt to clear the stage for all the AI talk we’re going to hear later today.

We may not have a Pixel 8a demo, but we’ve got the next best thing — a Pixel 8a review. And it sounds like this phone really impresses, thanks to its Tensor-powered AI features in a sub-$500 device along with Google’s extended software and security support.


This website uses cookies. By continuing to use this site, you accept our use of cookies.