Discover how to harness the potential of ChatGPT for advanced keyword research in SEO with our comprehensive guide.
Search On is constantly striving to improve search experiences by harnessing the power of machine learning advancements so that they better reflect how we all form a better understanding of our world. When we have a greater understanding of all forms of information, from images and languages to things within the world around us, we can unlock new ways of helping people to explore and gather information.
Search On is dedicated to improving visual search so that it is more natural and more intuitive, so people can navigate information in a more effective way. So, let’s take a more in-depth look at Google’s new and exciting upcoming ways of navigating the internet and discovering new information, new topics, and new ways of thinking in an effective and innovative way.
Lens is an effective tool that allows users to search the environment around them by using an image or their own camera. People are now using Lens to find answers to over 8 billion queries every month! Visual search was made more natural earlier in the year thanks to Multisearch – a significant milestone that was introduced to make searching for information easier.
Multisearch allows the user to take a screenshot or image and add text onto it in a similar way to naturally pointing at something then asking questions about it. Now available in the English language globally, Multisearch is set to be introduced in more than 70 languages over the months to come.
Google has been previewing how this search capability can be supercharged with a “Multisearch near me” capability which will allow the user to take either a screenshot or picture of an item or dish then instantly find it close by.
This exciting new way to search will enable users to find local businesses and connect instantly with them, whether they wish to support a shop in their neighbourhood or simply require something immediately. “Multisearch near me” is going to roll out in the English language in the United States soon.
A powerful element of understanding visually is being able to break the language barrier. Lens has gone far beyond being able to translate text and is now translating pictures. Every month, in fact, users are turning to Google for their translation needs, translating text within images more than a billion times in over a hundred different languages.
Thanks to the latest machine learning advancements, Google can now blend text that has been translated into complicated images, allowing it to feel and look more natural. Google has even optimised the models it uses for machine learning so it can achieve this in only 100 milliseconds, which is less time than it takes to blink your eye!
This is achieved by using GAN models (generative adversarial networks) which help to power the cutting-edge technology behind Pixel’s Magic Eraser.
We are constantly in a process of redefining the ways in which people can search for information and then interact with it. Google is working to allow people to ask the questions they want to know the answers to with fewer words, and even no words, and the search engine will still understand precisely what they mean and produce results that are helpful.
Users will also be able to explore information organised in ways that make more sense, whether going deeper into an unfolding topic or discovering a new viewpoint to expand perspectives.
A key element of this centres around being able quickly to uncover the results that users are searching for, so Google will soon be rolling out a quicker way of finding what users need. When users start typing a question into the search bar, Google will be able to supply relevant content immediately, before they even finish typing.
In some cases, users are unaware of which angle they need to explore before seeing it. Therefore, Google is introducing new searching experiences that help users to explore the topics that they are interested in more naturally when they visit Google.
Immediately a user begins to type their question, Google will provide topic or keywords options that allow the user to craft their question and that will allow the most relevant results to be returned.
As users learn more about a particular topic, they may find themselves wondering more about it, so Google is making it simpler to explore subjects by highlighting information that is especially helpful and relevant, including content that has been produced by open web creators.
So, for example, for a topic like a city, users may see short videos and visual stories from previous visitors, tips about exploring the city, fun things to see and do, the best ways of getting there, and all the other vital aspects they need to know when planning their travels.
Furthermore, thanks to Google’s deeper understanding of the ways in which people search, it will soon be able to show users topics that help them look into things more deeply, or find new directions on a particular subject.
Users will be able to remove or add topics when they wish to zoom out or in, and one of the best things about this is that it can help users to discover things they may never have thought of.
Google is also reimagining how it displays results so that they can reflect more effectively the ways in which people tend to explore topics. The content that is most relevant will be displayed from a range of different sources and in a range of formats from video to images and text. As users continue to scroll, they will see new ways of getting inspired by the topics that relate to their search.
Over the months to come, Google will be introducing these exciting new ways of exploring information, so that users can follow their curiosity and uncover new things.