The company says the device has 23 sensors in total. To make this a standalone device (it can be used without being tethered to a phone or computer, although it does, awkwardly, have an external battery pack), the company packed an M2 processor into the headset, along with a brand-new R1 processor, which takes care of all the inbound data from the 12 (!) cameras, lidar, eye sensors, six microphones and more. How did Apple wind up there? It appears that its engineers were given carte blanche to make the best device they could, solving some crucial problems along the way. Yes, the Apple Vision Pro is weighed down by its $3,500 price tag, and judging from the videos and cut-away photos alone, I’d be surprised if the company will make any money at all on the devices. There’s a chasm of difference between building a single prototype of something and building a few million of something. Apple has a long history of launching first-generation products that are pretty decent before quickly following up with updates and amendments that make the original kit better.įrom hardware engineers, you occasionally hear an under-breath muttered comment along the lines of “the third generation of X was the one we wanted to ship in the first place.” That isn’t uncommon in hardware it’s pretty rare that a full product vision survives the constraints of supply chains and manufacturing. ![]() TechCrunch tried out the Apple Vision Pro headset, concluding that it is very good - but perhaps also too good to be true.
0 Comments
Leave a Reply. |