Apple has rolled up its sleeves for the Deep Fusion feature that adds a new meaning to the camera on the iPhone 11 and 11 Pro models. So, what does this feature offer?
Apple, in the event of the iPhone 11, while highlighting the camera features in the event mentioned a feature called Deep Fusion. Apple announced it rolled up its sleeves for this feature, which is not available with iOS 13.1. The new iOS 13 Beta tests will be released next week and will be officially announced. So, what features does this add to the camera?
This feature; 9 photos are analyzed by the neural motor in 1 second. All pixels are then scanned to select the best one. Detailed and at least 24 Megapixel photos can be obtained with this system.
- Survival-Themed Dune Awakening Announced
- Fitbit introduced three new products: Sense 2, Versa 4, and Insprie 3
- iPhone 14 promotional date has been announced
- US parents want crypto money training in schools
- How much money do technology companies earn per second? There is Apple at the summit
Deep Fusion takes advantage of artificial learning developed to optimize photos pixel-by-pixel processing, texture, detail and noise throughout the photo. This feature, which is not yet in use, will be available on the iPhone 11 and iPhone 11 Pro before the end of 2019.
There are currently two photos taken with this feature which will add a new meaning to the camera of the iPhone 11. It’s hard to believe that these pictures were taken with the iPhone camera, but it was indeed taken with the iPhone 11.