Gurman: AirPods to Gain Hearing Health Features Over the Next Two Years
Apple's AirPods could gain more prominent hearing health features in the next year or two, according to Bloomberg's Mark Gurman.
In the latest edition of his
"Power On" newsletter, Gurman said that he believes Apple will "upgrade the AirPods to become a health tool in the next year or two," adding "the ability to get hearing data of some sort."
Apple has already added several hearing-focused features to the AirPods in recent years such as Live Listen and Conversation Boost, but Gurman noted that such features are not yet FDA approved or designed to serve as a hearing aid replacement. Given Apple's health objectives, he expects AirPods to take on these sorts of functions "more officially" in the near future.
Apple analyst Ming-Chi Kuo has previously highlighted the company's intention to add biometric health monitoring capabilities to future AirPods models. Apple patent filings have described an earbud-based fitness monitoring system that integrates an advanced biometric sensor to detect physiological metrics including temperature, heart rate, perspiration levels and more, through skin contact and via built-in motion sensors. Apple's vice president of technology, Kevin Lynch, has also hinted that AirPods could be a source of additional health data in the future.
New hearing health capabilities would not necessarily be contingent on new AirPods being released, since current models already contain speakers and microphones to support such features, and they could be pushed to existing devices via a firmware update.
Popular Stories
Apple has announced it will be holding a special event on Tuesday, May 7 at 7 a.m. Pacific Time (10 a.m. Eastern Time), with a live stream to be available on Apple.com and on YouTube as usual. The event invitation has a tagline of "Let Loose" and shows an artistic render of an Apple Pencil, suggesting that iPads will be a focus of the event. Subscribe to the MacRumors YouTube channel for more ...
Apple today released several open source large language models (LLMs) that are designed to run on-device rather than through cloud servers. Called OpenELM (Open-source Efficient Language Models), the LLMs are available on the Hugging Face Hub, a community for sharing AI code. As outlined in a white paper [PDF], there are eight total OpenELM models, four of which were pre-trained using the...
Apple has dropped the number of Vision Pro units that it plans to ship in 2024, going from an expected 700 to 800k units to just 400k to 450k units, according to Apple analyst Ming-Chi Kuo. Orders have been scaled back before the Vision Pro has launched in markets outside of the United States, which Kuo says is a sign that demand in the U.S. has "fallen sharply beyond expectations." As a...
Apple is set to unveil iOS 18 during its WWDC keynote on June 10, so the software update is a little over six weeks away from being announced. Below, we recap rumored features and changes planned for the iPhone with iOS 18. iOS 18 will reportedly be the "biggest" update in the iPhone's history, with new ChatGPT-inspired generative AI features, a more customizable Home Screen, and much more....
Apple is finally planning a Calculator app for the iPad, over 14 years after launching the device, according to a source familiar with the matter. iPadOS 18 will include a built-in Calculator app for all iPad models that are compatible with the software update, which is expected to be unveiled during the opening keynote of Apple's annual developers conference WWDC on June 10. AppleInsider...
Top Rated Comments
The Mystic Meg of the Mac world.
What a fab job he has. Makes tons of money while "forecasting" the bleeding obvious…
*sigh* Wish my school careers advisors way back in the '70s saw this coming.
One thing they absolutely NEED is a mechanism to troubleshoot AirPods on the iPhone. My impairment is worst in my left ear, and my left AirPod microphone has—apparently—been malfunctioning the past few months. But I didn’t know it, I just assumed my hearing was getting worse (which happens, unfortunately). And thanks to how the automatic microphone switching feature works, seemingly randomly folks I spoke to on the telephone would say my voice was “quiet” or “breaking up”, but with cellular and Bluetooth and earpieces, that’s not an indicator. And, like I said, it seemed random. It was until I really started paying attention to which ear I was putting AirPods into first that I tracked it down to the left mic. And by then my AppleCare had expired, though it had been ongoing for over 3 months, probably longer. Why couldn’t Siri/iOS have alerted me when using that the input from one microphone was substantially degraded over the other?? The OS had all the data, could have. Should have. Except that would be inconvenient for Apple’s bottom line. When I called Apple Support, they had no ability to troubleshoot the microphones and speakers on the iPhone, I had to suggest using Voice Memos… which I’d thought of on-the-fly while on the phone with them. That’s pretty bad. The alternative was for me to drive an hour to the closest Apple Store, but with no guarantee they could troubleshoot them. (I figured that really meant it was a trip whereby they could try to convince me to just buy another pair.) (Also found out that the external mic on my iPhone 12 mini was bad. On which AppleCare had just expired also. Sigh. Explains why “Hey Siri” had been working like crap for the past several months.)