Source link : https://tech-news.info/unlocking-the-future-ray-ban-smart-glasses-introduce-live-ai-and-translation-features/
Meta Enhances Ray-Ban Glasses with Cutting-Edge Features in 2023
After launching Ray-Ban glasses this year, Meta has been actively refining its functions through a series of feature updates. The newly released firmware version 11 brings exciting advancements in artificial intelligence, offering live AI-powered multimodal capabilities, instantaneous translation services, and Shazam integration.
Engage with Meta AI Effortlessly
This recent update significantly amplifies the multimodal features of the Ray-Ban ray-bans/” title=”Revolutionary Budget ChatGPT Smart Glasses Take on Meta’s Ray-Bans!”>smart glasses (review), enabling live video streaming that supports interactive AI functionalities.
The continuous live AI feature allows these smart glasses to capture real-time environmental data and adapt to user queries automatically—eliminating the need for wake commands or physical activation. Key highlights include the ability to identify ingredients for recipe suggestions and deliver contextual insights about visible objects and surroundings.
This functionality remains active even during phone calls through the Ray-Ban glasses, a fact that Meta showcased in its demonstration last September.
Real-Time Translation Functionality on Ray-Bans
The new assistant also supports natural conversation transitions by allowing users to change subjects or expand upon previous questions effortlessly. This enhancement builds upon Meta’s earlier “Look and Ask” feature that required users to activate the assistant for every inquiry—a limitation now resolved.
Similar to Google’s experimental smart glass prototype from two years ago that aimed at real-time translation capabilities, Meta is now incorporating this groundbreaking feature into its eyewear line.
The latest iteration of Ray-Bans can translate spoken languages directly into English audio played through the device’s speakers or displayed within the accompanying app. At launch, supported languages include French, Italian, and Spanish.
Song Identification with Shazam Integration
Additions don’t stop there; Meta has included Shazam for identifying music tracks too. However, unlike other advanced features such as real-time translation which operates continuously via live AI access without manual prompts—using Shazam requires activating Meta AI first. Users can request song identification or inquire about compositions playing nearby once powered on manually.
The latest updates are selectively available through a waitlist initially rolling out in North America before reaching additional locations worldwide; however, specific timelines for global expansion remain unspecified by Meta.
A Balancing Act: Advanced Features vs Privacy Concerns
While these innovative AI functionalities are strikingly impressive—and enhance user experience—they bring forth persistent concerns surrounding privacy and data handling—issues frequently associated with wearable technology like these smart glasses over recent years.
According to information from Meta itself,the images,text,and other data collected via these devices are processed remotely on cloud systems utilized to refine their artificial intelligence models—a consideration users should weigh carefully before activating such sophisticated features.
We Want Your Opinion!
What are your thoughts regarding evolving tools integrated into Ray–Ban Smart Glasses? We invite you share your views in comments below.
The post Unlocking the Future: Ray-Ban Smart Glasses Introduce Live AI and Translation Features! first appeared on Tech News.
Author : Tech-News Team
Publish date : 2024-12-22 05:13:27
Copyright for syndicated content belongs to the linked Source.