A future Apple HMD (Vision Pro, Smartglasses, Contact Lenses) will allow a user to visualize non-visible phenomena like Gasses, Audible Fences+
Yesterday the U.S. Patent and Trademark Office officially granted Apple a patent that relates to extended reality (XR) environments and future devices that will provide visualization of non-visible phenomena. The granted patent describes a user visualizing "Augmented Reality" in eyewear and being able to see "non-visible phenomena" like gasses or audio fences based on environmental sensor feedback.
Apple clarifies that their future mixed reality devices could take on many different forms such as head mountable systems, projection-based systems, heads-up displays (HUDs), vehicle windshields having integrated display capability, windows having integrated display capability, displays formed as lenses designed to be placed on a person's eyes (e.g., similar to contact lenses) and more.
Apple further notes that an HMD may detect, and/or receive information regarding, one or more non-visible features within a direct or pass-through field of view of a physical environment, and display a visualization of those detected non-visible features at the correct location of those features in the physical environment.
For example, responsive to a detection of non-visible feature of a physical environment, the device may display a visualization of the non-visible feature overlaid on the view of the physical environment at a location that that corresponds to the detected non-visible features.
The non-visible features may correspond to, for example, electromagnetic signals such as Wi-Fi signals, airflow from an HVAC system, temperatures of physical objects, fluids or gasses, an audible fence created for a pet (e.g., using ultrasonic pitches), sounds generated by a musical instrument, and/or hidden physical objects such as objects with known locations that are obscured from view by other physical objects (as examples).
If the HMD, smartglasses or other eyewear "detects" a non-visible phenomenon (via specialized sensors), the device will then allow the user to actually see the range of ultrasonic pitch or see where a dangerous gas actually is within their view.
Apple's patent FIG. 6 below illustrates a flow chart of an example process for providing computer-generated visualizations of non-visible phenomena; In the example of FIG. 3, a visualization #302 of a non-visible ultrasonic fence generated by ultrasonic fencing device #210 is also displayed, to be perceived by a user of electronic device 105 at the three-dimensional location of the non-visible ultrasonic fence.
Apple's patent FIG. 4 below illustrates other visualizations of non-visible phenomena that can be provided by an eyewear device such as electronic device #105. In the example of FIG. 4, a visualization #400 of airflow from vent #206 is illustrated. In this example, the visualization is a representation, in the visible light spectrum, of the airflow existing outside the visible light spectrum. In this example, the visualization is based on a visual context that includes wavy lines with a wave frequency and a separation that increases with decreasing airflow.
The airflow may be detected using an airflow sensor in the eyewear device that detects airflow directly from the movement of air through the sensor and/or indirectly.
In Apple's patent FIG. 5 above, a musical instrument implemented as a guitar #500 is shown being tuned using a visualization #501 of the non-visible sound generated by a string of the guitar. In this example, the visualization is generated based on a visual context that includes two sine waves such as a first sine wave #502 and a second sine wave #504.
For more details, review Apple's granted patent 11715301