As part of Apple’s Advanced Manufacturing Fund, Apple is investing $250 million in Corning, a supplier that has been working on glass for the iPhone, Apple Watch and iPad. Apple had previously invested $200 million in May 2017.
The company says that the new investment will support research and development for precision glass processes. While Corning has supplied glass to Apple for every generation of iPhone and iPad, Apple says that glass in the iPhone 11 and 11 Pro is even tougher than before. Apple also uses glass for the back of the device in order to enable wireless charging.
As Apple mentioned before, the company has spent $60 billion with 9,000 American suppliers in 2018. It represents 450,000 jobs.
Today’s investment is part of a commitment to spend billions of dollars in U.S.-based companies with its Advanced Manufacturing Fund in order to build new facilities and help manufacturers. Apple originally planned to invest $1 billion, but it has deployed the entire initial fund.
Apple has now spent $1 billion out of its $5 billion subsequent fund. For instance, Apple has invested $390 million in Finsar, the maker of the TrueDepth camera and $10 million in Elysis, an aluminum maker.
The newest version of Snap’s Spectacles already has a 3D feature that lets you see the world with immersive filter effects, and now the company’s flagship app Snapchat is levelling up. Today the company announced a new 3D Camera Mode that will let users make and share images with diorama-like depth effects that move when you tilt your phone.
The 3D Camera Mode is available as of today for those using the iPhone X and above with an update of the Snapchat app, where the feature can be accessed via the camera mode, using the drop-down menu on the right. The pictures can also be viewed (but not created) on older and other phone models (including Android) as well.
Alongside the 3D Camera Mode, those creating pictures will also, naturally, be supplied with a new library of 3D effects, lenses and filters; and after you are finished making the images, you can also save them to your camera roll to use elsewhere as well.
The move into 3D is the latest salvo for Snapchat in what has been a long-term feature battle with Facebook, and specifically Instagram . We have long documented the history of how Snapchat has led the charge with new concepts in photo art on its app — from the very basic aspect of ephemeral images, through to the emergence of lenses and filters, and stories to build narratives of Snaps and videos — only to see Instagram (and to a lesser extent, Facebook itself) follow suit with the similar features.
The feature-copy situation is particularly a tough one for Snapchat, which still trails Instagram in overall users and once saw its growth slow massively after its rival implemented Stories. Snapchat most recently reported 203 million daily active users, while Instagram currently says its DAUs are over 500 million.
Lenses and photo effects overall still have a long way to go, though — not least because currently some 70% the company’s daily users turn to lenses to spice up their pictures, pointing to a very sticky feature that helps keep them on Snapchat overall. So Snapchat’s push to keep innovating (even if it gets copied) is commendable.
And in the moving target that is consumer taste, that model is likely to also get changed up with more recent competitive developments: specifically it will be worth watching how and if the rise of the popular music-based TikTok app will impact what features we see on these two older rivals.
Despite all that, ironically, with 3D, Facebook was actually ahead of the game, launching AI-based 3D images back in October 2018. Up to now, it’s never extended that feature to Instagram. However, with Snapchat getting in on the action, I wouldn’t be surprised to see 3D show up on Insta, too.
To be clear, the 3D feature’s reliance on models of the iPhone X and newer cuts more legacy models of the iPhone out as a matter of necessity, since they are made using image and depth data that can be collected on the iPhone X’s front-facing lens.
On the other hand, you mind find it a strange oversight that the same feature is not showing up on its Android app — not least given that there are a fair number of high-end Android devices that can capture the same types of depth and other image data as the higher models of the iPhone.
Snapchat has had a history with Android. While it is a popular platform for mobile apps overall, at one point Snap had to redesign its Snapchat Android app because it was so slow and buggy, leading to plummeting users. Eventually it clawed some of that back, but it seems that for now, enough of its biggest users are on iOS that Snap continues to prioritise it when it comes to new features. It will be worth watching to see how long it takes Snap to extend this feature to Android. (We are asking.)
Cloudian, a company that enables businesses to store and manage massive amounts of data, announced today the launch of Edgematrix, a new unit focused on edge analytics for large data sets. Edgematrix, a majority-owned subsidiary of Cloudian, will first be available in Japan, where both companies are based. It has raised a $9 million Series A from strategic investors NTT Docomo, Shimizu Corporation and Japan Post Capital, as well as Cloudian co-founder and CEO Michael Tso and board director Jonathan Epstein. The funding will be used on product development, deployment and sales and marketing.
Cloudian itself has raised a total of $174 million, including a $94 million Series E round announced last year. Its products include the Hyperstore platform, which allows businesses to store hundreds of petrabytes of data on premise, and software for data analytics and machine learning. Edgematrix uses Hyperstore for storing large-scale data sets and its own AI software and hardware for data processing at the “edge” of networks, closer to where data is collected from IoT devices like sensors.
The company’s solutions were created for situations where real-time analytics is necessary. For example, it can be used to detect the make, model and year of cars on highways so targeted billboard ads can be displayed to their drivers.
Tso told TechCrunch in an email that Edgematrix was launched after Cloudian co-founder and president Hiroshi Ohta and a team spent two years working on technology to help Cloudian customers process and analyze their data more efficiently.
“With more and more data being created at the edge, including IoT data, there’s a growing need for being able to apply real-time data analysis and decision-making at or near the edge, minimizing the transmission costs and latencies involved in moving the data elsewhere,” said Tso. “Based on the initial success of a small Cloudian team developing AI software solutions and attracting a number of top-tier customers, we decided that the best way to build on this success was establishing a subsidiary with strategic investors.”
Edgematrix is launching in Japan first because spending on AI systems there is expected to grow faster than in any other market, at a compound annual growth rate of 45.3% from 2018 to 2023, according to IDC.
“Japan has been ahead of the curve as an early adopter of AI technology, with both the governmetn and private sector viewing it as essential to boosting productivity,” said Tso. “Edgematrix will focus on the Japanese market for at least the next year, and assuming that all goes well, it would then expand to North America and Europe.”
Fintech startup FairMoney is building a challenger bank in Nigeria. The company first started offering microcredit and now plans to expand to current accounts and savings. FairMoney just raised an $11 million Series A round (€10 million) led by Flourish, DST Global partners and existing partners Newfund, Speedinvest and Le Studio VC.
FairMoney lets you get a loan from its mobile app. After answering a few questions and sharing financial information, the startup analyzes this data set as well as your geolocation, other apps installed on your phone and other factors to give you an answer in a few minutes.
On average, people borrow the equivalent of $33. Eventually, if you always repay on time, you are able to borrow as much as $415. Interests vary depending on repayment periods and other factors, but the maximum annual percentage rate is 13%.
When you apply for a loan, FairMoney then uses traditional bank transfers to credit the money — bank transfers occur within a few minutes in Nigeria. You can then repay using cash with partner bank tellers, bank transfers or SMS.
FairMoney has a lending license in Nigeria to operate. The company will partner with microfinance institution to launch current accounts, savings and facilitate payments. Eventually, FairMoney hopes that it’ll get its own microfinance license from the central bank.
Like many challenger banks, FairMoney wants to become your financial hub for all your banking needs — one app to rule them all. That’s why the ability to hold money in your FairMoney wallet will be key. For users without smartphones, the startup is also working on an SMS interface to transfer money.
Let’s get this out of the way right up front: iPhone 11’s Night Mode is great. It works, it compares extremely well to other low-light cameras and the exposure and color rendition is best in class, period.
If that does it for you, you can stop right here. If you want to know more about the iPhone 11, augmented photography and how they performed on a trip to the edge of a galaxy far, far away, read on.
As you’re probably now gathering, yes, I took the new iPhones to Disneyland again. If you’ve read my otherreviews from the parks,you’ll know that I do this because they’re they ideal real-world test bed for a variety of capabilities. Lots of people vacation with iPhones.
The parks are hot and the network is crushed. Your phone has to act as your ticket, your food ordering tool, your camera and your map. Not to mention your communication device with friends and family. It’s a demanding environment, plain and simple. And, I feel, a better organic test of how these devices fare than sitting them on a desk in an office and running benchmark tools until they go dead.
I typically test iPhones by using one or two main devices and comparing them with the ones they’re replacing. I’m not all that interested in having the Android vs. iPhone debate because I feel that it’s a bit of a straw man given that platform lock-in means that fewer and fewer people over time are making a truly agnostic platform choice. They’re deciding based on heredity or services (or price). I know this riles the zealots in both camps, but most people just don’t have the luxury of being religious about these kinds of things.
Given the similarities in models, (more on that later) I mainly used the iPhone 11 Pro for my testing, with tests of the iPhone 11 where appropriate. I used the iPhone XS as a reference device. Despite my lack of a desire to do a platform comparison, for this year’s test, given that much discussion has been had about how Google pulled off a coup with the Pixel 3’s Night Sight mode, I brought along one of those as well.
I tried to use the iPhone XS only to compare when comparisons were helpful and to otherwise push the iPhone 11 Pro to handle the full load each day. But, before I could hit the parks, I had to set up my new devices.
Setup
Much of the iPhone 11 Pro’s setup process has remained the same over the years, but Apple has added one new feature worth mentioning: Direct Transfer. This option during setup sits, philosophically, between restoring from a backup made on a local Mac and restoring from an iCloud backup.
Direct Transfer is designed to help users transfer their information directly from one device to another using a direct peer-to-peer connection between the two devices. Specifically, it uses Apple Wireless Direct Link (AWDL), which also powers AirDrop and AirPlay. The transfer is initiated using a particle cloud link similar to the one you see setting up Apple Watch. Once it’s initiated, your old iPhone and new iPhone will be out of commission for up to 2-3 hours depending on how much information you’re transferring.
The data is encrypted in transit. Information directly transferred includes Messages history, full resolution photos that are already stored on your phone and any app data attached to installed apps. The apps themselves are not transferred because Apple’s app signing procedure locks apps to a device, so they must be (automatically) re-dowloaded from the App Store, a process that begins once the Direct Transfer is complete. This also ensures that you’re getting the appropriate version of the app.
Once you’ve done the transfer, the data on your phone is then “rationalized” with iCloud. This helps in cases where you have multiple devices and one of those other devices could have been making changes in the cloud that now need to be updated on the device.
Apple noted that Direct Transfer is good for a few kinds of people:
People without an iCloud backup
People who have not backed up in a while
People in countries where internet speeds are not broadly strong, like China
People who don’t mind waiting longer initially for a ‘more complete’ restore
Basically what you’ve got here is a choice between having your iPhone ‘ready’ immediately for basic functionality (iCloud backup restore) and waiting a bit longer to have far more of your personal data accessible from the start, without waiting for iCloud downloads of original photos, Messages history etc.
Direct Transfer also does not transfer Face ID or Touch ID settings, Apple Pay information or Mail Data aside from usernames and passwords.
After iPhone Migration is complete the Messages content from the device will be reconciled the Messages content in iCloud to ensure they are in sync. The same is true for Photos stored in iCloud.
Anecdotally, I experienced a couple of interesting things during my use of Direct Transfer. My first phone took around 2.5 hours to complete, but I still found that the messages archive alerted me that it needed to continue downloading archived messages in the background. Apple suggested that this may be due to this rationalizing process.
I also noticed that when simultaneous Direct Transfer operations were active, side-by-side devices took much longer to complete. This is very likely due to local radio interference. Apple has a solution to that. There is a wired version of the Direct Transfer option using the Camera Connection Kit with a source device and connecting them via USB. Ideally, Apple says, the transfer speeds are identical, but of course the wired option side-steps the wireless interference problem entirely — which is why Apple will be using it for in-store device restores for new iPhones using the Direct Transfer option.
My experience with Direct Transfer wasn’t glitch free, but it was nice having what felt like a ‘more complete’ device from the get go. Of note, Direct Transfer does not appear to transfer all keychain data intact, so you will have to re-enter some passwords.
Design and Display
I’ve been naked for years. That is, team no case. Cases are annoying to me because of the added bulk. They’re also usually too slippery or too sticky. I often wear technical clothing too and the phones go into slots designed for quick in out or fun party trick things like dropping into your hand with the pull of a clasp. This becomes impossible with most cases.
Apple provided the clear cases for all three iPhones, and I used them to keep them looking decent while I reviewed them, but I guarantee you my personal will never see a case.
I’m happy to report that the iPhone 11 Pro’s matte finish back increases the grippyness of the phone on its own. The smooth back of the iPhone 11 and the iPhone XS always required a bit of finger oil to get into a condition where you could reliably pivot them with one hand going in and out of a pocket.
Traveling through the parks you get sweaty (in the summer), greasy with that Plaza fried chicken and turkey legs and all kinds of kid-related spills. Having the confidence of a case while you’re in these fraught conditions is something I can totally understand. But day-to-day it’s not my favorite.
I do like the unified design identity across the line of making the bump surface blasted glass on the iPhone 11 with a glossy back and then flipping those on the iPhone 11 Pro. It provides a design language link even though the color schemes are different.
At this point either you’ve bought into the camera bump being functional cool or you hate its guts. Adding another camera is not going to do much to change the opinion of either camp. The iPhone 11 Pro and Pro Max have a distinctly Splinter Cell vibe about them now. I’m sure you’ve seen the jokes about iPhones getting more and more cameras, well, yeah, that’s not a joke.
I think that Apple’s implementation feels about the best it could be here. The iPhone 11 Pro is already thicker than the previous generation, but there’s no way it’s going to get thick enough to swallow a bump this high. I know you might think you want that too, but you don’t.
Apple gave most reviewers the Deep Green iPhone 11 Pro/Max and the minty Green iPhone 11. If I’m being honest, I prefer the mint. Lighter and brighter is just my style. In a perfect world, I’d be able to rock a lavender iPhone 11 Pro. Alas, this is not the way Apple went.
The story behind the Deep Green, as it was explained to me, begins with Apple’s colorists calling this as a color set to break out over the next year. The fashion industry concurs, to a degree. Mint, seafoam and ‘neon’ greens which were hot early in the year have given way to sage, crocodile and moss. Apple’s Deep Green is also a dark, muted color that Apple says is ideal to give off that Pro vibe.
The green looks nearly nothing like any of the photographs I’ve seen of it on Apple’s site.
Inperson, the Deep Green is reads as dark grey in anything but the most direct indoor light. Outdoors, the treated stainless band has an “80’s Mall Green” hue that I actually really like. The back also opens up quite a bit, presenting as far more forest green than it does inside. Overall, though, this is a very muted color that is pretty buttoned up. It sits comfortably alongside neutral-to-staid colors like the Space Gray, Silver and Gold.
The Silver option is likely to be my personal pick this time around just because the frosted white back looks so hot. The first time I won’t have gone gray or black in a while.
Apple’s new super retina display has a 2M:1 contrast ratio and displays up to 1200 nits in HDR content and 800 in non-HDR. What does this mean out in the sun at the park? Not a whole lot, but the screen is slightly easier to read and see detail on while in sunny conditions. The “extended” portion of Apple’s new XDR screen terminology on the iPhone 11 Pro is due to lux, a luminance metric, not a color metric, so the color gamut remains the same. However, I have noticed that HDR images look slightly flatter on the iPhone XS than they do on the iPhone 11 Pro. The iPhone 11’s screen, while decent, does not compare to the rich blacks and great contrast range of the iPhone 11 Pro. It’s one of two major reasons to upgrade.
Apple’s proprietary 3D touch system has gone the way of the dodo with the iPhone 11. The reasoning behind this was that they realized that they would never be able to ship the system economically or viably on the iPad models. So they canned it in favor of haptic touch, bringing more consistency across the lineup.
By and large it works fine. It’s a little odd for 3D touch users at first. You retain peek and quick actions but lose pop, for instance, because there’s no additional level of pressure threshold. Most of the actions that you probably commonly use 3D touch for, like the camera or flashlight or home screen app shortcuts work just fine.
I was bullish on 3D touch because I felt there was an opportunity to add an additional layer of context for power users — a hover layer for touch. Unfortunately I believe that there were people at Apple (and outside of it) that were never convinced that the feature was going to be discoverable or useful enough so it never got the investment that it needed to succeed. Or, and I will concede this is a strong possibility, they were right and I was wrong and this just was never going to work.
Performance and Battery
Apple’s A13 Bionic features efficiency cores that are 20% faster and use 40% less power than the A12 bionic — part of where some impressive battery life improvements come from. Its overall clock speed and benchmarks are up around 20% overall. The performance cores also use 30% less power and the GPU uses 40% less power. The Neural Engine doesn’t escape either and uses 15% lower power. All compared to the iPhone XS.
My focus there on the cores power usage is not to say this feels any less juicy, but all new iPhones feel great out of the box because Apple (usually) works to neatly match power requirements with its hardware. And any previous generation software is going to have plenty of overhead out of the box. No change here.
The biggest direct effect that this silicon work will have on most people’s lives will likely be battery life.
The iPhone 11 Pro has a larger battery than the iPhone XS, with a different higher voltage chemistry. That, coupled with power savings improvements mentioned above, along with more in the screen and other components means better battery life.
My battery tests over several days at the parks point to Apple’s claims about improvements over the iPhone XS being nearly dead on. Apple claims that the iPhone 11 Pro lasts 4 hours longer then the iPhone XS. The iPhone XS came in at roughly 9.5 hours in tests last year and the iPhone 11 Pro came in nearly bang on at 12 hours — in extremely challenging conditions.
It was hot, the networks were congested and I was testing every feature of the camera and phone I could get my hands on. Disneyland has some WiFi in areas of the park, but the coverage is not total, so I relied on LTE for the majority of my usage. This included on-device processing of photos and video (of which I shot around 40 minutes or so each day). It also included using Disney’s frustrating park app, about which I could write a lot of complaints.
I ordered food, browsed twitter while in lines, let the kids watch videos while the wife and I had a necessary glass of wine or six and messaged continuously with family and team members. The battery lasted significantly longer on the iPhone 11 Pro with intense usage than the iPhone XS, which juuuust edged out my iPhone X in last year’s tests.
One of the reasons that I clone my current device and run it that way instead of creating some sort of artificially empty test device is that I believe that is the way that most people will be experiencing the phone. Only weirdos like device testers and Marie Kondo acolytes are likely to completely wipe their devices and start fresh on purchase of a new iPhone.
I’m pretty confident you’ll see an improvement in the battery as well. I’ve done this a lot and these kinds of real world tests at theme parks tend to put more of the kind of strains you’ll see in daily life on the phone than a bench test running an artificial web browsing routine is. On the other hand, maybe you’re a worker at a bot farm and I’ve just touched a nerve. If so, I am sorry.
Also, an 18W power adapter, the same one that ships with iPad Pro, comes in the box. Finally, etc. It is very nice having the majority of my cables have at least one end that is USB-C now because I can use multi-port GaN chargers from Anker and power bricks that have USB-C. Apple’s USB-C lightning cables are slightly thicker gauge now, and support data transfer as well as the 18W brick. The bigger charger means faster charging, Apple claims up to 50% charge in 30 minutes with the new charger, which feels about like what I experienced.
It’s quicker, much nicer to top off while nursing a drink and a meatball at the relatively secret upstairs bar at the Wine Country Trattoria in DCA. There’s an outlet behind the counter just ask to use it.
Oh, about that improved FaceID angle — I saw, maybe, a sliiiiiiight improvement, if any. But not that much. A few degrees? Sometimes? Hard to say. I will be interested to see what other reviewers found. Maybe my face sucks.
Camera and Photography
Once upon a time you could relatively easily chart the path of a photograph’s creation. Light passed through the lens of your camera onto a medium like film or chemically treated paper. A development process was applied, a print was made and you had a photograph.
When the iPhone 8 was released I made a lot of noise about how it was the first of a new wave of augmented photography. That journey continues with the iPhone 11. The ISP that normally takes on the computational tasks associated with color correction and making images look presentable from the raw material the sensor produces. Apple has added the Neural Engines’s machine learning expertise to the pipeline and it’s doing a bunch of things in various modes.
Deep Fusion shoots 9 images, it pre shoots 4 long and 4 short exposure images into a buffer. Then when you press the shutter button it takes a longer exposure. Then the neural engine and ISP combine these on a pixel by pixel basis into your image.
This is what makes the camera augmented on the iPhone 11, and what delivers the most impressive gains of this generation, not new glass, not the new sensors — a processor specially made to perform machine learning tasks.
What we’re seeing in the iPhone 11 is a blended apparatus that happens to include 3 imaging sensors, 3 lenses, a scattering of motion sensors, an ISP, a machine learning tuned chip and a CPU all working in concert to produce 1 image. This is a machine learning camera. But as far as the software that runs iPhone is concerned, It has one camera. In fact, it’s not really a camera at all, it’s a collection of devices and bits of software that work together towards a singular goal: producing an image.
This way of thinking about imaging affects a bunch of features from night mode to HDR and beyond, and the result is the best camera I’ve ever used on a phone.
But first, let’s talk new lenses.
Ultra Wide
Both the iPhone 11 and the iPhone 11 Pro get a new “ultra wide angle” lens that Apple is calling a 13mm. In practice it delivers about what you’d expect from a roughly 13mm lens on a full-frame SLR — very wide. Even with edge correction it has the natural and expected effect of elongating subjects up close and producing some dynamic images. At a distance, it provides options for vistas and panoramic images that weren’t available before. Up close, it does wonders for group shots and family photos, especially in tight quarters where you’re backed up against a wall.
In my testing of the wide angle, it showed off extremely well especially in bright conditions. It allowed for great close up family shots, wide angle portraits that emphasized dynamism and vistas that really opened up possibilities for shooting that haven’t been on iPhone before.
One clever detail here is that when you shoot at 1x or 2x, Apple blends the live view of the wider angle lenses directly into the viewfinder. They don’t just show you the wide with crop marks over it, they are piping in actual feeds from the sensor so that you get a precise idea of how the image might look, while still letting you see that you have other options outside of the frame. It’s the camera viewfinder engineer version of stunting.
I loved shooting people with it up close, but that won’t be for everyone. I’d guess most people will like it for groups and for landscapes. But I found it great to grab fun tight shots of people or really intimate moments that feel so much more personal when you’re in close.
Of note, the ultra wide lens does not have optical image stabilization on either the iPhone 11 or iPhone 11 Pro. This makes it a much trickier proposition to use in low light or at night.
The ultra wide camera cannot be used with night mode because its sensor does not have 100% focus pixels and, of course, no OIS. The result is that wide angle night shots must be held very steady or soft images will result.
The ultra wide coming to both phones is great. It’s a wonderful addition and I think people will get a ton of use out of it on the iPhone 11. If they had to add one, I think adding the UW was the better option because of group shots of people are likely far more common than landscape photographers.
The ultra wide is also fantastic for video. Because of the natural inward crop of video (it uses less of the sensor, so it feels more cramped), the standard wide lens has always felt a little claustrophobic. Taking videos on the carousel riding along with Mary Poppins, for instance, I was unable to get her and Burt in frame at once with the iPhone XS, but was able to with the iPhone 11 Pro. Riding Matterhorn you get much more of the experience and less ‘person’s head in front of you’. Same goes with Cars where the ride is so dominated by the wind screen. I know these are very specific examples, but you can imagine how similar scenarios could play out at family gatherings in small yards, indoors or in other cramped locations.
One additional tidbit about the ultra wide: you may very well have to find a new grip for your phone. The lens is so wide that your finger may show up in some of your shots because your knuckle is in frame. It happened to me a bunch over the course of a few days until I found a grip lower on the phone. iPhone 11 Pro Max users will probably not have to worry.
HDR and Portrait Improvements
Because of those changes to the image pathway I talked about earlier, the already solid HDR images get a solid improvement in portrait mode. The Neural Engine works on all HDR images coming out of the cameras in iPhone to tone map and fuse image data from various physical sensors together to make a photo. It could use pixels from one camera for highlight detail and pixels from another for the edges of a frame. I went over this system extensively back in 2016 and its only gotten more sophisticated since with the addition of the Neural Engine.
It seems to be getting another big leap forward when Deep Fusion launches, but I was unable to test that yet.
For now, we can see additional work that the Neural Engine puts in with Semantic Rendering. This process involves your iPhone doing facial detection on the subject of a portrait, isolating the face and skin from the rest of the scene and applying a different path of HDR processing on it than on the rest of the image. The rest of the image gets its own HDR treatment and then the two images are fused back together.
This is not unheard of in image processing. Most photographers worth their salt will give faces a different pass of adjustments from the rest of an image, masking off the face so that it doesn’t turn out too flat or too contrasty or come out with the wrong skin tones.
The difference here, of course, is that it happens automatically, on every portrait, in fractions of a second.
The results are portraits that look even better on iPhone 11 and iPhone 11 Pro. Faces don’t have the artificially flat look they could sometimes get with the iPhone XS — a result of the HDR process that is used to open up shadows and normalize the contrast of an image.
Look at these two portraits, shot at the same time in the same conditions. The iPhone 11 Pro is far more successful at identifying backlight and correcting for it across the face and head. The result is better contrast ant color, hands down. And this was not an isolated experience, I shot many portrait shots side by side and the iPhone 11 Pro was the pick every time. With especially wide margins if the subject was back lit, which is very common with portraiture.
Here’s another pair, the differences are more subtle here but look at the color balance between the two. The skin tones are warmer, more olive and (you’ll have to trust me on this one) truer to life on the iPhone 11 Pro.
And yes, the High Key Mono works, but is still not perfect.
Night Mode
Now for the big one. The iPhone 11 finally has a Night Mode. Though I wouldn’t really call it a mode because it doesn’t actually require that you enable it, it just kicks in automatically when it thinks it can help.
On a technical level, Night Mode is a function of the camera system that strongly resembles HDR. It does several things when it senses that the light levels have fallen below a certain threshold.
It decides on a variable number of frames to capture based on the light level, the steadiness of the camera according to the accelerometer and other signals.
The ISP then grabs these bracketed shots, some longer, some shorter exposure.
The Neural Engine is relatively orthogonal to Night Mode working, but it’s still involved because it is used for semantic rendering across all HDR imaging in iPhone 11.
The ISP then works to fuse those shots based on foreground and background exposure and whatever masking the Neural Engine delivers.
The result is a shot that brightens dark-to-very-dark scenes well enough to change them from throw away images to something well worth keeping. In my experience, it was actually difficult to find scenes dark enough to make the effect intense enough. The new 33% improvement in ISO in the wide camera and 42% improvement on telephoto on iPhone XS already help a lot.
But once you do find the right scene, you see detail and shadow pop and it becomes immediately evident even before you press the shutter that it is making it dramatically brighter. Night Mode works only in 1x and 2x shooting modes because only those cameras have the 100% focus pixels needed to do the detection and mapping that the iPhone 11 needs to make the effect viable.
I have this weird litmus test I put every new phone camera through where I take it on a dark ride, like Winnie the Pooh, to see if I can get any truly sharp usable image. It’s a great test because the black light is usually on, the car is moving and the subject is moving. Up until this point I have succeeded exactly zero times. But the iPhone 11 Pro pulled it off. Not perfect, but pretty incredible all things considered.
A few observations about Night Mode:
The night images still feel like night time. This is the direct result of Apple making a decision not to open every shadow and brighten every corner of an image, flaring saturation and flattening contrast.
The images feel like they have the same genetic makeup as an identical photo taken without night mode. They’re just clearer and the subject is brighter.
Because of the semantic mapping working on the image, along with other subject detection work, the focal point of the image should be clearer/brighter, but the setting and scene does not all come up at once like a broad gain adjustment.
iPhone 11, like many other ‘night modes’ across phones, has issues with moving subjects. It’s best if no one is moving or they are moving only very slightly. This can vary depending on the length of exposure from 1-3 seconds.
On a tripod or another stationary object, Night Mode will automatically extend up to a 10 second exposure. This allows for some great night photography effects like light painting or trailing.
The result is naturally bright images that retain a fantastic level of detail while still feeling like they have natural color that is connected to the original subject matter.
Back when the Pixel 3 shipped Night Sight I noted that choosing a gain-based night mode had consequences, and that Apple likely could ship something based on pure amperage but that it had consistently made choices to do otherwise and would likely do so for whatever it shipped. People really hated this idea, but it holds up exactly.
iPhone XS Max. “Night Mode” is a healthy analog gain boost in the mid tone and shadow portions of the curve. Compromise is grain, grey shadows etc. There’s a lot of info in the sensor that isn’t being used in order to present something ‘realistic.’. Choices.
Though the Galaxy 10+ has a great night mode as well, the Pixel 3 was the pioneer here and still jumps to mind when judging night shots. The choices Google has made here are much more in the realm of ‘everything brighter’. If you love it, you love it, and that’s fine. But it is absolutely not approaching this from a place of restraint.
Here are some examples of the iPhone 11 Pro up against images from the Pixel 3. As you can see, both do solid work brightening the image, but the Pixel 3 is colder, flatter and more evenly brightened. The colors are not representative at all.
In addition, whatever juice Google is using to get these images out of a single camera and sensor, it suffers enormously on a detail level. You can see the differences here in the rock work and towers. It’s definitely better than having a dark image, but it’s clear that the iPhone 11 Pro is a jump forward.
The Pixel 4 is around the corner, of course, and I can’t wait to see what improvements Google comes up with. We are truly in a golden age for taking pictures of dark shit with phone cameras.
Of note, the flash is now 36% brighter than the iPhone XS, which is a nice fallback for moving subjects.
Tidbits
Auto crop
The iPhone 11 will, by default, auto crop subjects back into your videos shot at 1x or 2x. If you’re chasing your kid and his head goes out of frame, you could see an auto button on the 1 up review screen after a bit of processing. Tapping this will re-frame your video automatically. Currently this only works with the QuickTake function directly from the iPhone’s lock screen. It can be toggled off.
You can toggle on auto cropping for photos in the Camera settings menu if you wish, it is off by default. This has a very similar effect. It’s using image analysis to see if it has image data that it can use to re-center your subject.
Slofies
Yeah, they’re fun, yeah, they work. They’re going to be popular for folks with long hair.
U1
Apple has included a U1 chip in the iPhone 11 – can’t test it but it’s interesting as hell. Probably best to reserve talking about this extensively for a bit as Apple will ship the U1’s first iPhone functionality with a directional…AirDrop feature? This is definitely one of those things where future purposes, tile-like locator perhaps, were delayed for some reason and a side project of the AirDrop team got elevated to first ship. Interestingly, Apple mentioned, purely as an example, that this feature could be used to start car ignitions given the appropriate manufacturer support.
If this sounds familiar, then you’ve probably read anything I’ve written over the last several years. It’s inevitable that iPhones and Apple Watches begin to take on functionality like this, it’s just a matter of how to do it precisely and safely. The U1 has a lot to do with location on a micro-level. It’s not broad, network based or GPS based location, it’s precise location and orientation. That opens up a bunch of interesting possibilities.
No Night Mode
Night Mode
About that Pro
And then there was the name. iPhone 11 Pro. When I worked at a camera shop, you learned the power of the word “pro”. For some people it was an aphrodisiac, for others, a turn off. And for others, it was simply a necessity.
Is this the pro model? Oh I’m not a pro. Oooh, this is the pro!
We used it as a sales tool, for sure. But every so often it was also necessary to use it to help prevent people from over-buying or under-buying for their needs.
In the film days one of the worst things you could ever shoot as a pro-am photographer was gym sports. It was fast action, inside where it’s comparatively dim, and at a distance from court-side. There was no cheap way to do it. No cranking the ISO to 64,000 and letting your camera’s computer clean it up. You had to get expensive glass, an expensive camera body to operate that glass and an expensive support like a monopod. You also had to not be a dumbass (this was the most expensive part).
Amateurs always balked at the barrier of entry to shooting in these kinds of scenarios. But the real pros knew that for every extra dollar they spent on the good stuff, they’d make it up ten fold in profits because they could deliver product no parent with a point and shoot could hope to replicate.
However, the vast majority of people that walked into the shop weren’t shooting hockey or wrestling. They were taking family photos, outdoor pics and a few sunsets.
Which brings us to what the term Pro means now: Pro is about edge cases.
It’s not about the 80% case, it’s about the 20% of people that need or want something more out of their equipment.
For this reason, the iPhone 11 is going to sell really well. And it should because it’s great. It has the best new lens, an ultra wide that takes great family photos and landscape shots. It has nearly every software feature of iPhone 11 Pro. But it doesn’t have the best screen and it doesn’t have telephoto. For people that want to address edge cases – the best video and photo options, a better dark mode experience, a brighter screen — the iPhone 11 Pro is there — for everyone else, there’s still fiscal 2020’s best selling iPhone.
Self-driving truck startup TuSimple has added another $120 million to a Series D funding round led by Sina, operator of China’s biggest microblogging site Weibo, bringing the total haul to $215 million as it seeks to expand.
The company, which launched in 2015 and has operations in China, San Diego and Tucson, Arizona, hit unicorn status in February when it raised $95 million in the Series D round with a post-money valuation of $1.095 billion. This additional funding includes investment from UPS, which announced in August that it had taken a minority stake in TuSimple just months after the two companies began testing the use of autonomous trucks in Arizona.
TuSimple’s total funding is $298 million. New participants in the round include CDH Investments, Lavender Capital, and Tier 1 supplier Mando Corporation.
The company plans to use the funds to continue developing its autonomous vehicle technology and expand its long-haul routes in Arizona and Texas.
TuSimple is working on a “full-stack solution,” an industry term that means developing and bringing together all of the technological pieces required for autonomous driving. TuSimple is developing a Level 4 system, a designation by the SAE that means the vehicle takes over all of the driving in certain conditions.
In late 2017, TuSimple raised $55 million with plans to use those funds to scale up testing to two full truck fleets in China and the U.S. By 2018, TuSimple started testing on public roads, beginning with a 120-mile highway stretch between Tucson and Phoenix in Arizona and another segment in Shanghai.
The company has more than 50 trucks and 18 contracted customers, according to TuSimple CFO Cheng Lu.
One of those customers is UPS, which initially tapped TuSimple to help it better understand how Level 4 autonomous trucking might function within its network. That relationship expanded in May when the companies began using self-driving tractor trailers to carry freight on a freight route between Tucson and Phoenix to test if service and efficiency in the UPS network can be improved. UPS and TuSimple conduct daily testing between Phoenix and Tucson.
Africa focused tech talent accelerator Andela will cut 400 junior engineers across Kenya, Uganda, and Nigeria, CEO Jeremy Johnson told TechCrunch.
The layoffs come as the startup released first time income figures indicating it will surpass $50 million in annual revenues for 2019.
Yes, the news seems a bit disjointed. Not everything moves in the same direction in the business of startups.
On the staff cuts, “they are due to market demand for more senior engineering talent,” Andela said in a company release.
“We’ve seen shifts in the market and what our customers are looking for…toward more experienced engineers,” Johnson said on a call.
For those who may not know Andela’s business, the startup’s client-base is over 200 companies around the world that pay for the African developers Andela selects and trains to work on projects.
Founded in 2014, Andela has offices in New York and five African countries: Nigeria, Kenya, Rwanda, Uganda, and Egypt. The Series D tech-venture is one of Africa’s most visible (by press volume) and best funded ― backed by $181 million in VC from investors that include the Chan Zuckerberg initiative.
Andela selects a roster of developers each year who come on staff for a salary (similar to a management consulting firm) and are encouraged to continue working and living in their home markets in Africa.
By pre-layoff numbers, Andela had 1575 engineers on board. Big job cuts usually point to financial distress and decreasing demand for a company’s goods or services. That’s not the case with Andela’s personnel move, according to Johnson, who describes the layoffs more as a result of misreading the market.
“We’re actually actively and intensely growing, the mid and senior developer populations and next year we’re going to bring in 500 more developers,” he said.
“We’ve hired more junior developers than we are able to place in remote roles.”
The departing Andela software-engineers will gain severance packages and placement assistance, according to Johnson. The company is working with partners such as CcHub and iHub to connect the developers to new opportunities.
“Many of these people will rapidly get jobs in the local ecosystem and some day may come back and work at Andela again,” he said.
On Andela’s $50 million in 2019 projected income, “It’s the first time we’ve ever confirmed anything on revenue,” said Johnson ― who acknowledged the venture is still not profitable.
He wouldn’t say why the company released those figures now, but one can speculate it is to soften concerns about Andela’s financial performance in light of major staff cuts.
Johnson flagged the revenue significance in a global startup context. “What it means is the world needs what we do. Very few companies have gotten to a $50 million run rate in under five years.”
If that’s rare in developed markets, its even more scarce in Africa’s tech scene — where startups releasing any financial stats is scarce overall. Only one VC backed digital company has revealed revenues between $50 and $100 million. That’s e-commerce startup Jumia, that listed in an NYSE IPO earlier this year.
The release of 400 developers may be welcome in Africa’s most active tech hubs, such as Nigeria and Kenya, where rapid startup formation and funding is starting to outpace software engineering talent — according to a number of founders.
Job-placement will partially depend on whether local tech companies can offer competitive packages to incentivize the Andela alums.
If they do, the net effect of Andela’s layoffs could be more software-engineering capacity for Africa’s tech ecosystem ― so long as most of the developers remain in Africa.
Holidu, the Munich-headquartered holiday rentals search engine that is now active in 21 country markets, has raised €40 million in Series C funding.
The round was led by Prime Ventures, with participation from coparion and MairDuMont Ventures. Existing investors, including EQT Ventures, Venture Stars, Senovo and business angel Chris Hitchen, also followed on.
Founded in 2014 by siblings Johannes and Michael Siebers after they say they had a frustrating experience trying to book a vacation rental for a surfing trip in Portugal, Holidu’s search engine lets you easily search for and book holiday accommodation.
Claiming to use proprietary image recognition technology, Holidu compares the prices of more than 15 million rental properties across 600 different websites including Airbnb, Booking.com and Homeaway. This enables users to save “up to 55%” on their booking by automatically spotting price differences for the same property across various listings.
“Many of the sites offer the same rentals but at different prices,” Holidu co-founder and CEO Johannes Siebers told TechCrunch back in 2016. “Also, there is a large rate of rejected bookings as the different sites don’t synchronize calendars with each other and properties get double-booked. For consumers it is impossible to gain a transparent overview”.
To help solve this, Holidu has also developed a service for holiday property owners. Dubbed “Bookiply,” it offers a single interface to list properties on the largest travel websites, including synchronizing calendars, creating multilingual descriptions and sourcing professional photography. In addition, Bookiply’s team handles traveler communication.
Holidu says that Bookiply already manages 5,000 properties and claims it is the market leader in several European leisure destinations. “The focus is on property owners who are not yet online or whose digital presence can be optimised,” says the company.
Meanwhile, to boost growth, last year Holidu acquired its Spanish competitor Hundredrooms. The startup now claims 10 million visitors per month and says it will use the Series C funding for product development (both the Holidu website and the Bookiply software). It will also grow its Holidu partners and sign up more property owners to Bookiply. To achieve this, the company says it plans to open multiple regional offices.
Google said on Tuesday it is bringing a set of new features to Android TVs to improve the experience of users who rely on mobile hotspots to connect their giant devices to the internet. The features, developed by Google’s Next Billion Users team, will be first rolled out to users in India and then in other countries, the company said.
Ahead of its yearly event in New Delhi on Thursday, where the company is expected to make a number of announcements, Google said it has identified and addressed a problem faced by millions of users: Their TVs are not connected to the internet through Wi-Fi or wired/ethernet line.
Instead these users rely on hotspots (local network) created through their smartphones or tablets. “But that presents problems,” wrote Joris van Mens, Product Manager at Google’s Next Billion Users team, in a blog post. “Watching HD TV on a mobile data connection can quickly drain your daily data plan.”
To address this, Google says it is introducing a feature called ‘data saver’ to Android TVs that would reduce the data usage on mobile connections by up to three times, thereby allowing users to consume more content on their TVs. It is also introducing a ‘data alerts’ feature to help users better monitor how much data they have consumed watching TV.
The data saver feature will be optional to users
Another feature dubbed ‘hotspot guide’ will allow users to set up their TV with their mobile hotspot. And last, Google is introducing the ability in its Files app to allow users to cast video files locally stored on their phones to the TV without using internet data. Files app, which Google launched two years ago, allows users to easily free up content on their phones. The company said last month that Files app had amassed over 100 million users.
These four features will roll-out to Android TV devices starting with those manufactured by Xiaomi, TCL, and Marq by Flipkart, Google said. The company expects to rollout the features globally soon.
At an event in Bangalore on Tuesday, Xiaomi unveiled a new lineup of TVs that will support Netflix and Prime Video. The Chinese electronics giant, which is the top smartphone vendor in India, confirmed that its new TV models will support Google’s ‘data saver’ feature.
Later this week, Google is expected to make a number of announcements around its payments app and other services in its yearly Google for India event. Indian newspaper Economic Timesreported this week that one of those announcements could be the launch of Kormo, a job discovery app that is currently available in select developing markets, in India.
Missed the deadline to apply for Startup Battlefield at Disrupt Berlin 2019? We get it. Founder life is tough. Well guess what, we’re extending applications by one week. The extension ends September 27th at 11:59pm (PT). It’s time to buckle down and apply to the Startup Battlefield right now!
It’s easy and free to apply. Add your startup name to the mix to see if you are one of the chosen few to launch on the prestigious TechCrunch Disrupt Stage – equity free price money, global exposure and the best place to launch your startup this December. Selected teams will receive intensive pitch coaching from TechCrunch editors and the Startup Battlefield team. They’ll train you, grill you and get you ready for the big competition. All selected teams will gain access to private VIP events, participation in CrunchMatch: TechCrunch’s investor startup matching program, and complimentary exhibition space with event passes. Companies will pitch on stage for six minutes followed by a six minute Q&A in front of a panel of elite judges.
The judges then select the top few companies to compete in a final round where companies will pitch to a new set of judges, followed by an even more intense Q&A. One team will win the coveted Disrupt Cup trophy, the attention of international press and investors and, of course, a $50,000 in equity-free cash prize money..
The Startup Battlefield is one of the best platforms for launching your early-stage startup to the world’s “technorati.” We live-stream the entire event on TechCrunch.com, YouTube, Facebook and Twitter. Plus, it’s available later on-demand.
Need more convincing? Consider these stats and you’ll see it’s not hyperbole, people. In 12 years of Battlefields, 857 companies have competed and form the Startup Battlefield alumni community. Those startups, including the likes of JukeDeck, N26, Vurb, Dropbox, Mint, Yammer and more, have collectively raised more than $8.9 billion in funding and generated 112 exits.