When you pass by a Japanese tea shop in New York’s Bowery Market, just point your iPhone at the storefront and press and hold the button on the side of your phone to see business hours, photos from customers, and to call the shop. options are displayed. Or order.
Apple’s new visual intelligence tool, which will be available on the iPhone 16 lineup, takes the intermediate step of unlocking your phone, opening Google or ChatGPT, and typing a query or uploading a photo to get an answer. It is intended to be omitted. An early version of the feature is available as part of Apple’s iOS 18.2 developer beta, which launched for program participants on Wednesday.
The version I tried was an early preview aimed at developers rather than consumers, but it gave me a good idea of how visual intelligence works and what it brings to the iPhone experience. We briefly tested this very early version and found it to be great for quickly retrieving information about points of interest on the fly. While it can be useful, I also think it will take some time for consumers to embrace this feature once it’s released, as it represents a new way of thinking about how to interact with a mobile phone.
Still, this hints at a future where you won’t need to open as many apps to do things on your mobile device, and that’s promising.
But I’d like to give it more time and talk more about it once the final version is released.
read more: “The Cambrian Explosion: Fundamentally Reimagining the Smartphone with AI” coming soon
How visual intelligence works
Visual intelligence relies on new camera control buttons on iPhone 16, 16 Plus, 16 Pro, and 16 Pro Max. If you hold down the button, you’ll see a prompt explaining what Visual Intelligence is and informing you that the images won’t be saved to your iPhone or shared with Apple.
When the Visual Intelligence interface is open, you can take a photo by simply tapping the camera shutter button. From there, you can tap the on-screen button to ask ChatGPT about the image, or hit the search button.Click the button to launch Google Search. You can choose to use ChatGPT with or without an account. If you don’t sign in, your requests will remain anonymous and will not be used to train your ChatGPT model.
The current version of Visual Intelligence also gives you the option to report a concern by pressing an icon that looks like three dots. If you want to delete the image and take another one instead, × The icon is located where the shutter button is normally located on the screen.
On iPhone, in addition to using Google or ChatGPT, you’ll also see certain options based on what you’re pointing the camera at. For example, business hours if you’re pointing the camera at a store or restaurant.
What does it feel like to use it?
In the short time I’ve spent with Visual Intelligence, I’ve used it to find out about restaurants and stores, and to ask questions about video games.
It’s a quick and convenient way to access ChatGPT and Google, but what’s most interesting to me is how you can identify restaurants and stores. So far, this has been most effective when pointing the camera at a storefront rather than a sign or banner.
For example, when you scan the exterior of Kettl, the Japanese tea shop we mentioned earlier, Visual Intelligence automatically extracts useful information, such as photos of the various drinks. I had a similar reaction when I took a photo of a vintage video game store near my office. When you pressed the shutter button, Apple displayed the store’s name along with a photo of the store, a link to its website, and an option to call the store.
Once inside the store, we used Visual Intelligence to ask ChatGPT for game recommendations based on the titles in the store and learn more about the consoles and games in the store. That answer was pretty accurate, but it’s always worth remembering that chatbots like ChatGPT aren’t always accurate.
After taking a photo of the game on the shelf, I asked ChatGPT for games similar to the Persona Dancing series, and they suggested other titles with a similar emphasis on music and story. This seems like a smart answer, since Persona Dancing Game is a rhythm-based spin-off of Persona, a popular Japanese role-playing game. To find out that the Game Boy Color was released in 1998, all you had to do was take a quick photo and ask when it was released. (By the way, I got similar results when I asked the same question on the ChatGPT app.)
I’ve enjoyed experimenting with Visual Intelligence so far, and I feel like it could be even more useful while traveling. Being able to tap your iPhone at a landmark, store, or restaurant and learn more about it would have come in handy during my travels to France and Scotland earlier this year. In cities that are already familiar, there is less of a desire to immediately know detailed information about nearby places.
read more: What I learned from replacing my Apple Watch with Samsung’s Galaxy Ring
Visual intelligence and the future of mobile phones
It’s hard not to compare Visual Intelligence to Google Lens. Google Lens lets you use your phone’s camera to learn about the world around you instead of typing search terms. In its current form (also an early preview for developers), Visual Intelligence almost feels like a dedicated Google Lens/ChatGPT button.
Considering Google Lens has been around for years, it might not feel new or different. But this kind of functionality is so important that the fact that modern iPhones have a dedicated button for it speaks volumes. This shows that Apple thinks there may be a better way to search and get things done on your phone.
Apple isn’t alone in this belief. Startups like Google, OpenAI, Qualcomm, and Rabbit all believe that AI can unlock new ways to use cameras on mobile devices by making them more than just a way to take photos, but a discovery tool. At the annual Snapdragon Summit this week, Qualcomm showed off a virtual assistant concept that uses a camera to do things like split a restaurant bill into three installments based on a photo of your receipt.
The trick is to get the public to adopt it. I’d wager that muscle memory might prevent many people from ditching the old method of tapping and swiping to take photos, even if it were faster and more efficient.
Developing new habits takes time. However, Visual Intelligence is still in early preview and more features will be added in the future.
Apple’s iPhone 16, 16 Plus show off bold colors and buttons
See all photos