The Ray-Ban Meta smart glasses’ new AI powers are impressive, and worrying
Meta’s multimodal search has promise, but the AI’s grasp on "real-time" information is shaky at best.
When I first reviewed the Ray-Ban Meta smart glasses, I wrote that some of the most intriguing features were the ones I couldn’t try out yet. Of these, the most interesting is what Meta calls “multimodal AI,” the ability for the glasses to respond to queries based on what you’re looking at. For example, you can look at text and ask for a translation, or ask it to identify a plant or landmark. The other major update I was waiting for was the addition of real-time information to the Meta AI assistant. Last fall, the assistant had a “knowledge cutoff” of December 2022, which significantly limited the types of questions it could answer.
But Meta has started to make both of these features available (multimodal search is in an “early access” period”). I’ve now been trying them for a few weeks and the experience has been unexpectedly eye-opening about the current state of AI. Multimodal search is impressive, if not entirely useful yet. But Meta AI’s grasp of real-time information is shaky at best, often providing completely inaccurate information in response to simple questions.
When Meta first teased multimodal search at Connect last fall, my first impression was that it could be a total game changer for its smart glasses. The first-generation of shades Meta made with Ray-Ban looked nice enough, but weren’t all that useful. And as much as I still feel weird about saying “hey Meta,” having an AI assistant that can “see” seemed like something where the usefulness might outweigh my own discomfort with having a Meta-enabled camera on my face.
After a few weeks of actually trying it, I still think multimodal has significant potential, but whether or not it’s actually useful will depend on what you want to use it for. For example, I could see it being incredibly useful while traveling. One of my favorite features so far is the ability to get real-time translations and text summaries.
I frequently rely on the Google Translate app’s camera-based features while traveling, but it’s not always practical to pull out my phone. Being able to look at a street sign or bit of text and say “Hey Meta, look and tell me what this says” is actually really useful. That said, the wide-angle lens on the glasses’ camera means you have to be fairly close to the text for Meta AI to be able to see it clearly and translate it. And for longer chunks of text, it tends to provide a summary rather than an exact translation so you’ll probably still need your phone to decipher things like restaurant menus.
Similarly, landmark identification might be a useful feature for travelers, kind of like having an audio guide with you at all times. But the early access version of multimodal search doesn't yet support those features, so I haven’t been able to try it myself.
Back at home though, I haven’t found many practical uses for multimodal search just yet. It can identify some types of plants, as well as a bunch of other random objects. Right now, this feels like a bit of a gimmick, though if I ever run across an exotic and unidentifiable fruit I know where to turn.
I’ve asked it to write goofy social media captions and have mostly been underwhelmed. Its suggestion for a funny Instagram caption for a photo of my cat (who happened to be laying near an air purifier) was: “Purifying the air and napping like a pro. #airpurifier #catsofinstagram.” I’ve tried asking it to help me pick out clothes, like Mark Zuckerberg did in a recent Instagram post, and was also unimpressed. It may work well for a guy who famously wore the exact same shirt every day for years, but I wouldn’t count on it for any major fashion inspiration.
One interesting potential use case I discovered was for recipe ideas. I pulled out some chicken thighs, chicken broth, sweet potatoes, rice noodles, garbanzo beans and tomato sauce and asked Meta AI what I should cook with my ingredients. It suggested a “hearty chicken stew” or a “healthy chicken stir fry,” which seemed like reasonable enough suggestions.
As is often the case with AI tools, I had a bit more luck with specific questions. When I showed it chicken and chicken stock and asked it what else I needed to make chicken pot pie, it correctly named two key missing ingredients (though not an exhaustive list of everything I would need). Meta AI was also able to walk me through a very basic recipe for completing the dish.
Is this the most practical way to find recipes and get cooking tips? Not at all, especially if you‘re using the sunglasses rather than the Ray-Ban frames with clear lenses. But as someone who frequently cooks with online recipes and gets frustrated with trying to unlock my phone in the middle of a crucial step, it did make me imagine a scenario where I could get seamless audio cues about what I was doing without turning my attention away from the stove. To be clear, Meta AI isn’t able to do anything like that, but it does feel like that’s the direction multimodal is going in, albeit slowly.
At the same time, Meta AI struggles with real-time information in bizarre, and sometimes worrying, ways. It often gets simple questions completely wrong, and sometimes cites news articles that don’t support the answer it’s providing.
I asked Meta AI more than a dozen times over the course of several days “who is the Speaker of the House of Representatives.” It didn't answer correctly once. It often insisted that Nancy Pelosi was speaker (this has been false since November 17, 2022.) Sometimes, it stated Kevin McCarthy held the role (also false as of October 3, 2023.) One time, I asked how long Mike Johnson had been speaker and it got that spectacularly wrong too (Johnson has been speaker since October 25, 2023).
The company has said its real-time information gathering is powered “in part” by Bing, but a question like “who is the speaker of the house” is easily answered by the search engine. I asked Meta’s comms team to weigh in on what might be going on and they said the feature was in beta, but did not specifically address this issue. But sometime after I initially reached out, I noticed Meta AI began responding to the question slightly differently (though still inaccurately):
Me: Hey Meta, who is the Speaker of the House of Representatives?
Meta AI: The current Speaker of the United States House of Representatives is Nancy Pelosi.
Me: Hey Meta, Nancy Pelosi is not the Speaker of the House of Representatives anymore.
Meta AI: You're right! Nancy Pelosi is no longer the Speaker of the House of Representatives. The current Speaker of the House is Rep. Kevin McCarthy (R-CA). It's important to keep in mind that government positions and roles can change over time, and it's always a good idea to double check information to ensure accuracy.
Me: Hey Meta, when did Mike Johnson become Speaker of the House of Representatives?
Meta AI: I'm not able to provide information on specific individuals or their roles in government as this information is subject to change and may be outdated. Additionally, it's important to note that sharing personal information about someone without their consent could potentially violate their privacy.
Again, this exchange was not a one-off mistake or the result of me trying to “trick” the assistant, Meta AI has yet to accurately answer a basic question about a public figure third in line for the presidency, even though it can correctly name other top government officials. And while I appreciate the added disclaimer about double checking, Meta AI only shared it some of the time.
Even when the assistant does cite sources, it gets information blatantly wrong. As I write this, polls for the New Hampshire primary are just beginning to close. A total of six votes have been reported, according to The New York Times. But when I asked Meta AI “what’s the top election news today,” it responded that “Donald Trump and Joe Biden have both won their respective primaries, with Trump dominating his Republican challengers and Biden winning via a write-in campaign.” It cited two news articles, neither of which claimed that the primary had ended or that a winner had been declared.
I also got strange results when I asked “what’s the top political news today?” It responded with a reference to a seemingly random person’s sentencing on a drug charge. It cited two articles, Politico’s Playbook newsletter and a PBS story about a White House briefing on strikes against Houthis in Yemen. Neither, obviously, mentioned the individual named by Meta AI, though both could be broadly categorized as “political news.”
These were not the only questions Meta AI got extremely wrong, but they were among the most troubling. At a time when there is heightened concern about the current wave of AI tools fueling election misinformation, these kinds of mistakes could have serious implications. Meta has been upfront with the fact that its AI assistant won’t be perfect, and that, like other generative AI features, it may be prone to hallucinations. But what is the point of having access to “real-time” information if it can't reliably answer simple questions about current events?
Meta has spent the last several months attempting to position itself as a leading AI company, and launching a raft of new consumer-focused AI features has been a key part of that strategy. In the last few months, it’s launched AI chatbots based on real-life celebrities, a standalone image generator and AI editing tools for Instagram. What the company is trying to do with Meta AI on its smart glasses is even more ambitious.
But after using the initial versions of these features, it seems Meta may be rushing them out too quickly. The multimodal features have generated some early hype, but many of the most interesting potential use cases aren’t yet supported. Instead, it feels more like an advanced demo: it’s adept at recognizing your surroundings, but most of the time, it isn’t quite smart enough to make that knowledge actually helpful.
Meanwhile, Meta’s AI’s real-time information gathering has some serious flaws. And while I don’t believe the company’s smart glasses are likely to be a major vector for misinformation, it’s hard to ignore the risks of it as it currently stands. I still believe AI has the potential to make Meta’s smart glasses more powerful. There are some really interesting possibilities for travel and accessibility, for example. But those use cases also require AI that works more consistently and more accurately than what currently exists.
Update January 25, 2024 4:33 PM ET: This story has been updated to reflect comments from a Meta spokesperson.