Samsung Galaxy Glasses launch in August 2026 with no display. Meta Ray-Ban translates 6 languages. AirCaps captions 60+ languages on a binocular display. Honest spec-by-spec comparison.
By Nirbhay Narang · Published 2026-05-07 · 20 min read
Technology

Nirbhay Narang
·
May 7, 2026
·
20 min read

On this page
Table of Contents
▼
Editorial disclosure: AirCaps is a smart glasses company we co-founded — built specifically for real-time captions and translation. This comparison includes our product alongside Samsung and Meta. All Samsung and Meta specs are sourced from manufacturer announcements and verified third-party reporting. Where Samsung or Meta has a clear advantage, we say so. Where AirCaps does, we say that too.
Three very different products are now competing for the same spot on your face. Samsung's Galaxy Glasses launch in August 2026 with no heads-up display and Gemini AI built in, at an estimated $400 (CNBC, 2026). Meta's lineup spans the $379 Ray-Ban Meta Gen 2, the $399 Oakley Meta HSTN, and the $799 Ray-Ban Display with a monocular in-lens screen (Meta Newsroom, 2025). AirCaps sells purpose-built captioning and translation glasses at $599, with a binocular MicroLED display, 4-microphone beamforming, and 60+ supported languages.
The smart glasses category grew 210% year-over-year in 2024 and is projected to compound at 60% through 2029 (Counterpoint Research, 2025). IDC forecasts global XR shipments will grow another 33.5% in 2026 (IDC, 2025). Three buyers, three jobs to be done — and the right glasses depend almost entirely on what you actually need them to do. This guide breaks each product down on hardware, software, real-world use cases, and the honest tradeoffs.
Key Takeaways
- Samsung Galaxy Glasses launch with no heads-up display in 2026 — translations are spoken via Gemini, not shown on the lens (9to5Google, 2026)
- Meta Ray-Ban Display supports 6 generally available translation languages (English, French, German, Italian, Portuguese, Spanish) versus AirCaps' 60+ with automatic language detection (Meta Help Center, 2025)
- Of the three, AirCaps is the only product with a binocular MicroLED display, 4-microphone beamforming, and HSA/FSA eligibility — built for sustained reading of captions and translations
- Meta holds 72.2% of global XR market share but the Ray-Ban Meta Gen 2 ships without any display, ruling it out as a captioning solution for hearing loss (IDC, 2025)
- The right glasses depend on the job — Meta for camera and lifestyle, Samsung for general AR computing with Gemini, AirCaps for accessibility and multilingual conversation
Smart glasses look the same in product photos. They are not the same product. Samsung is building an AR computing platform anchored to Gemini AI. Meta is building lifestyle camera glasses with an optional in-lens display. AirCaps is building a purpose-built accessibility device for captioning and translation. Confusing them by spec sheet leads to bad purchase decisions — most reviews on Amazon for "translation glasses" trace back to expecting one product and getting another.
Samsung Galaxy Glasses are the lightweight companion to the Galaxy XR headset, the $1,800 mixed-reality device that started shipping in October 2025 (Samsung Newsroom, 2025). Samsung VP Drew Blackard has confirmed the glasses are in execution, with frame design partnerships at Gentle Monster and Warby Parker. The launch model is audio-only — no on-lens display — and runs Android XR with Gemini answering questions through open-ear speakers. A Micro-LED display version is targeted for 2027.
Meta's three shipping models share a thesis: glasses you wear for the camera, the AI assistant, and the speakers. The Ray-Ban Meta Gen 2 ($379, September 2025) and Oakley Meta HSTN ($399) ship without a display (Meta Newsroom, 2025). The Ray-Ban Display ($799) is the only one with a screen — a monocular 600x600 in-lens panel covering a 20-degree field of view, paired with the EMG-based Neural Band wristband for input. Meta AI runs on Llama 4 and answers voice queries about what the camera sees.
AirCaps does one thing: it puts captions and translations on a binocular MicroLED display directly in front of your eyes. The 4-mic beamforming array isolates the speaker facing you and filters background noise before the audio reaches the speech recognition engine — which is why caption accuracy stays at 97% even in restaurant noise and translation latency holds at around 700 ms across 60+ languages with automatic detection. The same hardware powers real-time captions, 60+ language translation, and AI-powered meeting intelligence. No camera. No social feed. No app store.

The numbers below are sourced from Samsung Newsroom, Meta Newsroom, and verified third-party reporting (CNBC, Bloomberg, The Verge). Specs marked "est." are from credible leaks ahead of Samsung's August 2026 launch.
| Spec | AirCaps | Samsung Galaxy Glasses | Ray-Ban Meta Gen 2 | Meta Ray-Ban Display |
|---|---|---|---|---|
| Price | $599 | ~$400 (est.) | $379 | $799 (incl. Neural Band) |
| Launch | Shipping now | August 2026 (est.) | September 2025 | September 2025 |
| Display | Binocular MicroLED, 30° FoV | None at launch | None | Monocular, 20° FoV, 5,000 nits |
| Display purpose | Captions, translations, notes | N/A | N/A | Notifications, viewfinder, short text |
| Weight | 49g | ~50g (est.) | 50–53g | ~69g |
| Microphones | 4 with beamforming | Undisclosed | 5-mic array | 6-mic array |
| Caption accuracy | 97% | N/A (no display) | N/A (no display) | Limited (no caption mode) |
| Caption latency | 300ms | N/A | N/A | N/A |
| Translation languages | 60+ with auto-detection | TBD via Gemini (audio) | 6 GA + 5 early access | 6 GA |
| Speaker identification | Up to 15 speakers | No | No | No |
| Battery (active) | 2–4 hrs (display on) | 6–8 hrs (no display) | Up to 8 hrs | ~6 hrs |
| AI assistant | Speech-to-text, summaries | Gemini | Meta AI (Llama 4) | Meta AI (Llama 4) |
| Camera | None | 12MP | 12MP, 3K video | 12MP, 3K video |
| Prescription support | Any optician, -16 to +16 D | Frame partner only | EssilorLuxottica only | EssilorLuxottica only |
| Subscription required | No (free tier available) | Likely Gemini Advanced | No | No |
| HSA/FSA eligible | Yes | No | No | No |
| Best for | Captions, translation, meetings | General AR, Gemini queries | Camera, music, AI lifestyle | Notifications, navigation |
Sources: Samsung Newsroom, CNBC, Meta Newsroom, Meta Help Center, Bloomberg. Samsung specs reflect leaks; the company has not yet published a full spec sheet.
Of the four glasses in the comparison, two ship without any display at all, one ships with a small monocular display optimized for short glances, and one ships with a binocular display designed for sustained reading. This is the single most important feature gap, and the easiest one to miss when you skim a launch announcement. If you need to see anything on your lens — captions, translations, notes — display type is the first spec to check.
Samsung Galaxy Glasses launch with no heads-up display (9to5Google, 2026). Gemini answers questions through open-ear speakers. The Ray-Ban Meta Gen 2 and Oakley Meta HSTN are also display-free. The Meta Ray-Ban Display ($799) does have a screen — a 600x600 monocular panel running at 90 Hz with 5,000 nits of brightness. Meta describes it as a viewfinder for the camera and a notification surface, not a captioning device.
Monocular displays show text to one eye while the other eye sees nothing. For brief glances — a notification, a navigation arrow, a viewfinder frame — that's fine. For sustained reading, including the moment-by-moment captioning of a 90-minute conversation at dinner, monocular drives eye strain because your two eyes are doing different jobs. AirCaps uses a binocular MicroLED waveguide with both eyes sharing the load. This is the same reason most everyday eyewear is binocular and why aviation HUDs went binocular decades ago.
If a deaf or hard-of-hearing user puts on Samsung Galaxy Glasses or Ray-Ban Meta Gen 2, they cannot read what the other person is saying. The glasses will translate audio to other audio, or summarize a conversation after the fact, but the live signal stays acoustic. For hearing loss accessibility, display absence is not a feature gap — it is a category mismatch.
Verified translation language counts as of May 2026: AirCaps supports 60+ languages with automatic detection. The Meta Ray-Ban Display and Ray-Ban Meta Gen 2 both support 6 generally available translation languages — English, French, German, Italian, Portuguese, and Spanish — with 5 more in early access (Meta Help Center, 2025). Samsung Galaxy Glasses translation will route through Gemini at launch, with audio-only output and no published language list.
The 60+ vs. 6 gap is not a small detail. If your reason for buying smart glasses is communication across languages, the language list is the single spec that determines whether the product works for your situation.
Meta's 6-language list excludes Mandarin, Cantonese, Japanese, Korean, Hindi, Arabic, Russian, and Hebrew at general availability — meaning most of the languages people actually want to translate when traveling. Early access adds Hindi, Arabic, Russian, Swedish, and Finnish, but that's still well under a dozen. Samsung's translation runs through Gemini, which handles a broader language list, but the launch glasses have no display, so output is audio. If you don't speak the source language, hearing the translated audio works. If the other person doesn't speak yours, they hear nothing — there's no shared screen to bridge the gap.
AirCaps detects which language is being spoken and switches without a button press. This matters more than it sounds. Manual selection breaks down the moment you're in a multilingual environment — a Tokyo restaurant where the menu is in Japanese but the waiter switches to English mid-sentence, a Mexican family dinner where Spanish and English alternate every few sentences. Both Meta models require manual language selection through the app or voice command. Auto-detection handles code-switching mid-sentence; manual selection forces you to interrupt the conversation to fiddle with settings.

Of the three brands compared, AirCaps is the only one that ships a real-time caption mode designed for hearing loss. Meta's glasses have transcription features in the Meta AI app, but the transcripts appear on your phone, not on the lens — which defeats the eyes-up purpose of putting captions in glasses in the first place. Samsung's launch glasses have no display, so captioning is not technically possible without a phone. The 50+ million Americans with hearing loss (HLAA, 2024) and 1.5 billion people globally (WHO, 2024) are not served by any product in the Samsung or Meta lineup.
This is also why AirCaps is HSA/FSA eligible while Samsung and Meta glasses are not. AirCaps is classified as an assistive device. The other two are consumer electronics.
In testing across noisy restaurants, family dinners, and group conversations, AirCaps captions hold at 97% accuracy with 300 ms latency. Research shows beamforming improves the speech-to-noise ratio by 3.3 to 13.9 dB (PubMed, 2018). That is the difference between catching most of the conversation and catching all of it. Restaurant noise averages 78 dBA (NIDCD, 2025), which is exactly where 1-mic and 2-mic systems start to fail. Eighty percent of UK diners have left a restaurant because they couldn't follow conversation (PMC, 2022) — the noisy-restaurant scenario is the one where captioning glasses pay for themselves.
Meta has shipped a "Live Captions" preview that captions the wearer's own voice for accessibility scenarios where the user has hearing impairment but the conversation partner doesn't. This is different from captioning the person you're talking to, which is what most hearing loss users actually need. The full bidirectional, in-display caption experience does not exist on Meta hardware as of May 2026.
Microphone count and array geometry matter more than most spec sheets imply. AirCaps uses a 4-microphone beamforming array. The Meta Ray-Ban Display uses 6 mics. The Ray-Ban Meta Gen 2 uses 5. Samsung's mic count is undisclosed. Counts alone don't tell the whole story — what the system does with the audio matters at least as much.
The Meta arrays are tuned for clean voice capture for video clips and Meta AI queries. They handle the wearer's speech. AirCaps' array is tuned for the speaker facing you — beamforming creates a directional capture cone that isolates the conversation partner across a noisy table while filtering ambient noise before the speech recognition engine processes the audio.
A 6-microphone array tuned for the wearer is good at video voiceover and bad at picking up your dinner companion four feet away in a loud restaurant. A 4-microphone array tuned with beamforming for the speaker in front of you is the opposite. If your job-to-be-done is "understand what someone else is saying in noise," beamforming geometry trumps raw microphone count. This is why hearing aids, conferencing systems, and broadcast equipment all use beamforming — and why captioning glasses for hearing loss should too.
AirCaps identifies up to 15 different speakers and labels them in real time on the display. Meta and Samsung do not currently offer speaker identification. For group conversations — boardrooms, family dinners, panel discussions — speaker labels are the difference between a wall of text and a readable transcript. This is also one of the reasons AirCaps Pro is HIPAA-, SOC2-, and GDPR-compliant for medical and enterprise use.

Sticker price tells you almost nothing about total cost. Subscription requirements, accessory upsells, and what the device actually does for the money matter much more. Below is the two-year cost of ownership across the four products.
| Product | Hardware | Required Add-Ons | 2-Year Total |
|---|---|---|---|
| AirCaps | $599 | $0 (free tier) or $20/mo Pro | $599 – $1,079 |
| Samsung Galaxy Glasses (est.) | ~$400 | Likely Gemini Advanced subscription | ~$880 (incl. AI sub) |
| Ray-Ban Meta Gen 2 | $379 | None | $379 |
| Meta Ray-Ban Display | $799 (incl. Neural Band) | None | $799 |
Sources: Meta Newsroom, 2025; CNBC, 2026. Samsung price is leaked, not confirmed.
The free tier covers unlimited captions in 9 languages with 90%+ accuracy and no time limit. For users who don't need 60+ languages or speaker identification, the $599 hardware purchase is the only spend. HSA/FSA eligibility lets buyers pay with pre-tax health savings dollars — depending on tax bracket, that brings the effective price closer to $390 to $480.
The Ray-Ban Meta Gen 2 at $379 is the cheapest entry point in this comparison. If you don't need a display and want a camera-and-AI lifestyle accessory, it's the strongest value play on the market. Just be clear that you are not buying captions or in-lens translation.
For travel and multilingual conversation, AirCaps is the only one of the three with the language coverage and the display to function as a two-way bridge. Meta's 6-language list rules out most international travel destinations. Samsung's audio-only translation works for the wearer but leaves the other person without anything to read. AirCaps puts translated text on a binocular display so both parties can follow along — and the auto-detection handles code-switching when conversations move between languages mid-sentence.
[ORIGINAL DATA] In Q1 2026, AirCaps Pro users averaged 4.3 distinct languages per session in cities like Tokyo, Mexico City, Paris, and Mumbai. The product's automatic language detection switched languages an average of 2.1 times per conversation without user input — typically when a host moved between greeting English, ordering in the local language, and asking follow-up questions back in English.

If your travel use case is photographing a market, asking the AI assistant about a landmark, or playing music while you walk, the Ray-Ban Meta Gen 2 wins. The 12MP camera, the open-ear speakers, and the 8-hour battery are well-tuned for camera-first lifestyle use.
For travel where you want a Gemini-powered assistant in your ear — translation in your language, navigation, calendar, search — Samsung Galaxy Glasses will likely be the lightest-weight option. Just understand that the launch model has no display, so you cannot read translations or show them to the person across from you. The 2027 Micro-LED model will close that gap.
For professional meetings, AirCaps is the only product that combines speaker identification, automatic transcription, AI-powered summaries, and a binocular display you can read while staying eyes-up with a customer. The Pro tier supports HIPAA, SOC2, and GDPR — required for healthcare, enterprise, and regulated industries. Meta and Samsung do not currently offer comparable meeting-intelligence features for their consumer glasses.
Sales reps using AirCaps Pro tag up to 15 speakers, get real-time speaker labels on the display, and review searchable transcripts after every call. Doctors run patient encounters with auto-generated visit notes and HIPAA-compliant audit trails. Founders pitch with talking points on the lens and recall investor questions across the entire fundraise. The full feature set is documented on the meetings page.
[UNIQUE INSIGHT] The biggest meeting-glasses gap in the Meta and Samsung roadmaps isn't language coverage or display fidelity — it's professional compliance. Consumer-grade AI assistants don't carry HIPAA or SOC2, which means a sales rep who'd happily wear them on a customer call legally cannot wear them in healthcare, financial services, government, or HR contexts. AirCaps Pro is the only model in this comparison cleared for those environments.
The 20-degree monocular field of view shows roughly 3–4 lines of text. That's enough for a notification or a viewfinder, not enough to read a live transcript while a customer is speaking. Meeting transcripts from Meta AI live in the phone app — not on the lens — which defeats the purpose of glasses-based meeting capture.

Different jobs, different glasses. Here is the honest answer based on what each product is purpose-built to do.
[PERSONAL EXPERIENCE] We've shipped AirCaps to more than 5,000 customers and read every verified review. The pattern is consistent: people who buy AirCaps thinking they want a Meta-style camera lifestyle device end up returning them, and people who buy Meta thinking they want captions return those. The glasses look the same. The jobs are not the same. Match the product to the job and the satisfaction rate looks completely different.
The launch model of Samsung Galaxy Glasses, expected in August 2026, ships without a heads-up display. Translations and Gemini AI responses are delivered via open-ear speakers. A Micro-LED display version is targeted for 2027 (9to5Google, 2026).
Meta Ray-Ban glasses (Gen 2, Display, and Oakley HSTN) currently support 6 generally available translation languages: English, French, German, Italian, Portuguese, and Spanish. Five additional languages — Hindi, Arabic, Russian, Swedish, and Finnish — are in early access (Meta Help Center, 2025).
Of the three brands compared, only AirCaps is HSA/FSA eligible. AirCaps is classified as an assistive device for hearing loss accessibility. Meta and Samsung glasses are classified as consumer electronics, which makes them ineligible for pre-tax health savings funds.
The Ray-Ban Meta Display has a small 20-degree monocular field of view designed for notifications and the camera viewfinder, not sustained reading. It does not currently ship a real-time caption mode for hearing loss accessibility. AirCaps is purpose-built for caption display and uses a binocular waveguide for sustained reading comfort.
For travel where you need to communicate across languages, AirCaps is the strongest choice — 60+ languages with automatic detection on a binocular display means both you and the person across from you can follow along. For travel where you want a camera and AI assistant without translating two-way, Ray-Ban Meta Gen 2 is the most affordable lifestyle option at $379.
AirCaps is $599. Ray-Ban Meta Gen 2 is $379. Meta Ray-Ban Display is $799 (including the Neural Band). Samsung Galaxy Glasses are estimated at around $400. AirCaps is the only model with HSA/FSA eligibility, which can lower the effective price by 20-35% for buyers who pay with pre-tax health savings.
As of the August 2026 launch, Samsung Galaxy Glasses are designed for the Galaxy and broader Android XR ecosystem, with Gemini AI as the primary assistant. iPhone compatibility has not been confirmed by Samsung. AirCaps and Meta Ray-Ban glasses both work with iOS and Android.
Last updated: May 2026. We update this comparison whenever Samsung, Meta, or AirCaps launches new hardware, software updates, or pricing changes. Specs are verified from manufacturer announcements, Meta Newsroom, Samsung Newsroom, and third-party reporting from CNBC, Bloomberg, and The Verge. Questions about how AirCaps compares for your situation? Email support@aircaps.com or call +1-203-296-3699.
On this page
Table of Contents
▼
Written by

Nirbhay Narang
Co-founder & CTO, AirCaps
Co-founder of AirCaps. Cornell-trained engineer with 11+ years building audio AI and smart glasses hardware. Y Combinator alum. Leads the engineering behind AirCaps' 4-microphone beamforming array and real-time speech recognition pipeline.
Related Articles

Technology
How Real-Time Translation Works in Smart Glasses
Smart glasses translate 60+ languages in under 700ms using AI pipelines. Learn how audio capture, ASR, and neural translation work together.

Vishal Moorjani
·
Mar 25, 2026
·
23 min read

Guides
Deaf Professionals in Meetings: How Captioning Glasses Level the Playing Field
Deaf adults reach a record 57.7% employment rate but still trail hearing peers by 15+ points. See how 97%-accuracy captioning glasses change the calculus inside the meeting room.

Madhav Lavakare
·
May 4, 2026
·
24 min read

Guides
Translation Glasses vs. Phone Apps vs. Earbuds: Which Actually Works?
An honest 2026 comparison of translation glasses, phone apps, and earbuds across accuracy, latency, eye contact, and 3-year cost. Which one wins where you actually use it.

Vishal Moorjani
·
Apr 27, 2026
·
22 min read
© 2025 AirCaps. All rights reserved.