Apple is gearing up to provide developers with resources to create apps and experiences for the Apple Vision Pro headset, and yesterday, the company pushed out new backend assets for the Vision Pro.
Code for updating the battery pack that powers the Vision Pro has been introduced, along with other visionOS assets needed for test versions of the device. Apple appears to have three different model numbers for Vision Pro batteries, including A2781, A2988, and A2697. It is not clear why there are three separate numbers when Apple has only announced two-hour battery life for the device, but there could be multiple battery packs in development or there could be additional battery options that will be used solely in Apple Stores for testing.
The code additions come ahead of when Apple plans to introduce development kits to select developers. Apple on its developer website says that it plans to introduce Apple Vision Pro developer kits and other developer tools starting in July, and there are less than two weeks left in the month.
Apple has just pushed live the backend for the Vision Pro battery to receive firmware updates along with a few other visionOS assets. This comes as Apple is gearing up to send out Vision Pro developer kits sometime this month. pic.twitter.com/tNiyN0nB7E — Aaron (@aaronp613) July 21, 2023
Developer kits will help developers bring their creations to life directly on Vision Pro, according to Apple, providing the tools to quickly build, iterate, and test apps. Developers will be able to apply to get a kit.
Apple last offered a developer kit ahead of the launch of Apple silicon chips in 2020. Apple provided developers with a Mac mini that had an A-series chip so that developers could transition from Intel to Apple silicon. The Mac mini machines were priced at $500, and the Apple Developer Kits will also likely require a payment from developers.
Along with Apple Vision Pro developer kits, Apple says that it will offer Apple Vision Pro compatibility evaluations for existing apps and opportunities for developers to visit a Vision Pro developer lab that provides demonstrations of visionOS, iPadOS, and iOS apps running on the headset. Labs will be available in Cupertino, London, Munich, Shanghai, Singapore, and Tokyo.
Apple previously released the first visionOS beta in June, and Unity earlier this week released beta versions of Vision Pro game development tools. Apple is partnering with Unity on the "PolySpatial" development platform for visionOS that will allow developers to create and port 3D experiences to the Vision Pro headset.
Top Rated Comments
This is going to be a game changer. Instead of a virtual reality a la Meta that nobody wants this thing augments the real world. The possibilities are immense.
For example, I am a scientist and this instrument offers unprecedented research possibilities. So much so that I am writing grants to get funding to buy some and hire developers. From research perspective, this headset is a steal in terms of costs.
Imagine using something like in people that have cognitive deficits. An advanced lighter and more portable version of this could help people in various ways in their daily lives.
After a couple of years there will be consumer prices versions of this that are really portable. And even at a 1500 dollars price point, I believe it would be a good price. If you don’t want a computer monitor this might be the best screen you could buy.
My prediction is that in a decade it will be a commonplace gadget. And it will not be long that Samsung et al jump ok the bandwagon.
No fuss. No muss!
[MEDIA=twitter]1682081253544316928[/MEDIA]
And thanks to this feature, the system is capable of recording in 3D mind you what you are presently looking at. I don't think that anything like this is available in another product at this price point, or at all.
Focusing on what the system yet is not seems to ignore the framework it introduces. The first iPHone was very limited, but the framework was the grounbreaking aspect -- a device the size of a cell phone ONLY consisting of a touch screen and showing a clever interface to use fingers to interact with something like this in a completely novel way. I think this was one of the major conceptual aspects that made it back then clear to me this will be the future. I have the same feeling now about the AVP. The framework is exceptional -- control with you eyes and your hands, not huge controllers, and working with the real world at the same time. Size, weight, battery life, etc etc, these are all peripheral problems that will be solved over time. What counts is the concept, and that is just again very clever.