Google Search Gets Multisearch, Lens AR Translate, and More

Google Search Gets Multisearch, Lens AR Translate, and More

At its ‘Search On’ event, Google introduced a plethora of new capabilities for its search engine, many of which would make results richer and more visually appealing. In order to create search experiences that function more like our minds—that is as multidimensional as people—we are moving well beyond the search box.

You’ll be able to locate exactly what you’re looking for as we move into this new era of search by fusing images, sounds, text, and speech. According to Google SVP of Search Prabhakar Raghavan, “We call this making Search more natural and intuitive.”

First, Google is bringing the multi-search function, which it first debuted in test in April of this year, to English globally. Over the course of the next several months, it will also be available in 70 additional languages. By fusing text and visuals, the multi-search capability allowed users to look up many items at once.

The function is also compatible with Google Lens. Google claims that customers use its Lens feature to search for what they see nearly eight billion times per month.

Google Search Gets Multisearch, Lens AR Translate, and More

Also: Google Maps Will Be Updated to Look More Like the Real World

However, users will be able to take a picture of an object and then use the term “near me” to find it close by thanks to the integration of Lens and Multisearch. This “new method of searching will help people find and engage with local companies,” according to Google. Later this fall, the “Multisearch near me” will begin to roll out in the US in English.

“This is made feasible by in-depth knowledge of the product inventories and geographic locations. In relation to multi-search and Lens, Raghavan stated that they were “influenced by the millions of photographs and reviews on the web.”

Google is enhancing the way translations appear over images. Over 1 billion times per month, consumers use Google to translate text on photos into more than 100 different languages, claims the business. Google will be able to “blend translated text into complicated pictures, so it looks and feels a lot more natural” thanks to the new technology.

As a result, rather than striking out from the original image, the translated text will appear more seamless and integrated. To guarantee this experience, Google says it is utilizing “generative adversarial networks (also known as GAN models), which helps fuel the technology powering Magic Eraser on Pixel.” Later this year, this feature will become available.

Additionally, it is updating its iOS app so that users can access shortcuts just beneath the search box. Users will be able to search for music, translate any text using their camera, and more thanks to this.

When people search for information about a location or subject, the search results on Google Search will also become more visually appealing. In the example, Google provided, the initial set of results for a search for a city in Mexico also included videos, photographs, and other content related to the location.

According to Google, this will prevent users from having to open numerous tabs in order to learn more about a location or a subject.

Even when a user starts to type a question, it will also offer more pertinent information about the upcoming month. In order to assist users in formulating their queries, Google will offer “keyword or topic possibilities.” Some of these themes, like cities, etc., will also feature content from creators on the open web, along with travel advice, etc.

According to the company’s blog post, the “most relevant content, from a number of sources, regardless of the format the information arrives in — whether that’s text, images, or video” will be displayed. The launch of the new functionality is scheduled for the upcoming months.

Google Search Gets Multisearch, Lens AR Translate, and More

Also: How to Make a Google Account!

When looking for food—whether it is a specific dish or an item at a restaurant—Google will display visually richer results, including images of the sought-after dish. Additionally, it is “covering additional ground for digital menus and enhancing their visual quality and dependability.”

The company claims that in order to produce these novel results, it combines “menu information provided by people and merchants, and found on restaurant websites that use open standards for data sharing.” It also relies on its “image and language understanding technologies, including the Multitask Unified Model.”

Google stated in a blog post that the menus “will highlight the most well-liked foods and conveniently flag out different dietary options, starting with vegetarian and vegan.”

Additionally, it will change how shopping results show up on Search, making them more visual with links and allowing users to shop for a “full look”. The user will be able to view these specific things in the 3D view from the search results, which will also offer 3D shopping for sneakers.

Google Search Gets Multisearch, Lens AR Translate, and More

Maps on Google

Though most of these will only be available in a few areas, Google Maps is also getting some new features that will provide more visual information. Users will be able to, among other things, determine where to eat, go sightseeing, etc. in a specific area by checking the “neighborhood vibe.”

Tourists will find this interesting because they can utilize the information to get to know a neighborhood better. Google claims that in order to provide this information, it combines “AI with local knowledge from Google Maps users.” In the upcoming months, Neighbourhood Vibe will launch on Android and iOS worldwide.

Additionally, the immersive view function is being expanded to provide users with 250 photorealistic aerial views of famous places throughout the world, ranging from the Acropolis to the Tokyo Tower. In Google’s blog post, the company said that “predictive modeling” is the method used for the immersive view to automatically understand previous trends for a location.

In the upcoming months, the immersive view will be made available for Android and iOS in Los Angeles, London, New York, San Francisco, and Tokyo.

The Live View functionality will also allow users to view useful data. Users who are out and about can utilize the search with the Live View tool to locate a market or store nearby. In the upcoming months, Search with Live View will be made available on Android and iOS in London, Los Angeles, New York, San Francisco, Paris, and Tokyo.

Through the Google Maps Platform, it is also making its eco-friendly routing tool, which debuted previously in the US, Canada, and Europe, available to outside developers. Google hopes that businesses in other sectors, like delivery or ride-hailing services, would provide customers the chance to enable eco-friendly routing in their apps and track fuel usage.


Kimberly is a freelance writer with a love of writing and traveling. She has been writing for most of her life and has been published in various magazines and online publications. She writes about entertainment, technology, and lifestyle-related topics at Kimberly is always looking for new writing opportunities and loves learning about new cultures and experiences.

Related Posts

Leave a Reply

Your email address will not be published. Required fields are marked *