Google is working on language-to-text AR glasses. It’s a complicated idea

​​Google appears to be working on AR glasses once again, but this time, it is showing a new feature that translates speech to readable text.  At last week’s Google I/O 2022, the company demonstrated an AR glasses prototype that can translate spoken language into readable display text. Google has not hinted whether they are developing these as a product, or when, but the fact that they showed them to developers is indicative that they are thinking of how to extend the model of AR glasses to utilize their gigantic datasets and existing technologies. If Google moves forward with the product, it is likely that it will frame it as a device that would  attempt to break down language barriers. Sounds great, right? No more trying to find Google Translate on the web and pecking phrases into our mobile phones to translate things. When (or if) these hit the market, we’ll finally be able to read foreign signs, order correctly in restaurants, and even make new friends more easily when we travel. More significantly, there would be a way to quickly translate communication in the event of an emergency, when people may not all speak the same language. On another level, these “translation glasses” could also open up communication channels for the deaf and hard of hearing community, giving them a new way to communicate with those around them. However, as with all new technology ideas, Google’s translation glasses could come with a huge social cost: to our privacy, our well-being, and our cooperation with each other in our communities. What does it mean when Google becomes the translator for our lives, and are we comfortable with that notion?  The problem with any type of technology translation device is that it has to “listen” to those around it, to procure the data to translate. And if the AR glasses are listening, we will need to know what, or whom, they are listening to—and when they are listening. At the moment, we don’t know if these glasses will be able to distinguish between more than one person at a time, either. Also, we will need to know if it is legal for these glasses to listen without consent—and if one needs the consent of someone to record them in order to translate them, will one need the glasses to translate the consent? We don’t know if in the future, these glasses will have the capacity to record what they translate, nor will we know if they could identify whom they are recording at any given time, or within what range they are capable of listening. And If they are recording glasses, or even with the transcribed text, we’ll need to know if  that is stored somewhere that can be erased, and if people could opt-out in a public space without being recorded while doing so. Let’s assume for the moment that these Google glasses won’t record us, and that Google manages to figure out consent and permission. Given that, in our crowded, noisy world, the usual problems with speech to text could still abound in the form of misunderstandings, misrepresentations, etc., in what Google ‘hears’ and what it writes as a result of that hearing. The tech might also have a lot of misspellings and confusion with mixing languages. As The Verge pointed out, many of us “code switch” using words from many different languages interspersed, with the added complexity of not all of them reading from left-to-right, which will need to be accommodated, too.  Now add to that an aggregate population using these while wandering around, which invokes much of what I wrote with Dr. Catherine Flick about Meta’s pre-Ray-Ban Stories Project Aria glasses. Many of the same issues persist, except for that with these new Google glasses, people may be walking around and reading transcripts, which again, is more like what was going on in the early days of cell phones and Divided Attention, creating potentially dangerous outcomes as distracted people walk into traffic or fall into fountains. One of the main concerns with the glasses is Google’s seeming assumption here that technology can solve cultural problems—and that if the technology isn’t working, the solution is to develop and apply more technology. In this case, solving cross-cultural communication problems cannot be fully solved with language translation. Tech can help, but these glasses won’t translate culture or cultural norms such as whether someone is comfortable being direct or indirect, or any one of multitudes of cultural nuances and cues found in the ways that different people in different groups communicate with each other. For that, we need other humans to guide us.

S.A. Applin, PhD, is an anthropologist whose research explores the domains of human agency, algorithms, AI, and automation in the context of social systems and sociability. You can find more at @anthropunk and PoSR.org.

https://www.fastcompany.com/90753311/google-is-working-on-language-to-text-ar-glasses-its-a-complicated-idea?partner=rss&utm_source=rss&utm_medium=feed&utm_campaign=rss+fastcompany&utm_content=rss

Établi 3y | 18 mai 2022 à 14:21:55


Connectez-vous pour ajouter un commentaire

Autres messages de ce groupe

What to know about the American Airlines jet and Army helicopter that collided over D.C.

The Army helicopter and regional American Airlines jet that collided over Washington are both workhorse aircraft that operate around the world on a daily basis.

There were 60 passengers

30 janv. 2025 à 22:10:09 | Fast company - tech
How Apple Vision Pro is finding a home in healthcare

In early 2024, Ryan Broderick, a surgeon at UC San Diego Health, was commiserating with some colleagues about the profusion of screens in today’s operating rooms. Though the displays provided essen

30 janv. 2025 à 22:10:08 | Fast company - tech
If you were bored watching videos at a plodding 2x speed, you’re in luck: They can go faster

MrBeast’s fast-cutting, non-stop dramatics are already quick enough. Now, some users can watch his videos (and more) at 4x speed.

YouTube is continuing their push for pace. In a

30 janv. 2025 à 22:10:07 | Fast company - tech
After a week of DeepSeek freakout, doubts and mysteries remain

Welcome to AI DecodedFast Company’s weekly newsletter that breaks down the most important news in the world of AI. You can sign up to receive this newsletter ever

30 janv. 2025 à 19:50:02 | Fast company - tech
Microsoft posts 10% growth for Q4 as it plans to spend $80 billion on AI

Microsoft said Wednesday that its profit for the October-December

30 janv. 2025 à 17:30:05 | Fast company - tech
Tesla’s Q4 results fell short of Wall Street’s forecasts. Here’s why

Tesla’s fourth-quarter adjusted profits rose slightly amid a big p

30 janv. 2025 à 15:10:09 | Fast company - tech