Throughout human evolution, vision and hearing have been the primary ways we perceive and interact with the world. Yet, for the past several decades of modern computing, technology has constrained us into unnatural forms of interaction: keyboards, mouse clicks, and touchscreens. These tools, while transformative, have required us to adapt to machines rather than the other way around.
Artificial Intelligence, as we know it today, followed this trajectory, with chat windows and text prompts dominating its early interactions. However, 2024 has marked a shift toward more intuitive engagements powered by advancements in Natural Language Processing (NLP) and Natural Language Understanding (NLU). Take Google’s recent commercials, for example, showcasing AI as a voice companion on their latest Pixel phones.
But last week, Google unveiled announcements that signal an even bigger leap forward—one that could redefine how we engage with AI in the future. Could AR glasses become the interface that unlocks AI’s true potential? Let’s explore.
Google made two major announcements last week in collaboration with Samsung and Qualcomm:
Together, these announcements provide a roadmap for AI’s future—one where devices understand and reason about the physical world in ways that align with our natural senses.
At the heart of this shift is Gemini 2.0, Google’s latest AI model. This iteration is designed not just for text-based interactions but for engaging with the physical world. Some key features include:
The implications: AI with vision and reasoning can transform how we understand and interact with the real world.
To complement Gemini’s capabilities, Google also announced Android XR, a new operating system for AR and XR devices. Developed in partnership with Samsung and Qualcomm, Android XR aims to:
The first hardware device running Android XR, codenamed Project Moohan, is being developed by Samsung and is expected to launch next year. Google also mentioned and showed prototype glasses that will soon be used for real-world testing.
The long-term goal is to create an ecosystem of devices and apps, making AR glasses a compelling platform.
These announcements hint at Google’s vision for AI’s “future form factor.” By embedding Gemini into AR glasses powered by Android XR, Google is positioning these devices as the ultimate interface for AI. Imagine:
Yet, while the potential is exciting, the path to mainstream adoption of AR glasses is far from guaranteed. Google isn’t alone in this pursuit—competitors like Apple and Meta are also exploring AR and XR ecosystems. But by combining AI, AR, and a robust OS, Google’s approach could set a new standard. The question is: will developers and consumers buy into this vision?
The announcements of Gemini 2.0 and Android XR mark a significant step forward in aligning technology with our natural modes of interaction. While the road to mainstream adoption of AR glasses is still long, these innovations provide a compelling glimpse into the future. One question remains: Will AR glasses truly become AI’s future form factor? Google might think so—and they’re making bold moves to turn that vision into reality.