Apple filed a patent two weeks ago for Apple Watch gesture tracking. The idea: offload hand recognition from Vision Pro's cameras to sensors on your wrist, cutting power consumption and extending battery life.
A few weeks before that, Mark Gurman and Ming-Chi Kuo both reported that camera-equipped AirPods are coming. Infrared sensors in each earbud. Gesture recognition. Spatial awareness. Kuo's supply chain analysis puts Foxconn preparing 18 to 20 million camera units.
You don't commission 18 million camera units for an accessory to a device that few people own.
Neither of these makes much sense as a Vision Pro enhancement. Vision Pro sells 45,000 units a quarter. You don't commission 18 million camera units for an accessory to a device that few people own. You build infrastructure for the product that comes next.
Apple Glasses need everything Vision Pro can't provide. They need to be light enough to wear all day. They need gesture control that doesn't require a headset. They need spatial sensing that works without a computer on your face. They need to offload processing to devices you're already wearing.
An Apple Watch that handles gesture tracking. AirPods that sense your environment and recognize your movements. Both always on, both already in hundreds of millions of pockets and ears. The distributed platform that makes lightweight glasses viable already exists across devices people bought for other reasons.
Vision Pro showed Apple what spatial computing could do. The Watch and AirPods patents show where Apple is building the foundation to do it without a headset.