ADVERTISEMENT
ADVERTISEMENT

Google and MIT built software that dramatically improves smartphone photos before you take them

The researchers trained their system with machine learning algorithms by letting it compare raw photos to images retouched by professional photographers.

Google CEO Sundar Pichai recently remarked that google is a AI first company.

Google has collaborated with researchers at the MIT on a system that can retouch your smartphone photos before you've taken them, The Verge first reported.

Their system uses machine learning algorithms to make photos look like they've been taken by a professional photographer.

The idea is to transform a picture by tweaking specific aspects of it, like brightness or saturation, using settings that are specific to the image in frame.

That is nothing particularly new by itself; systems like this already exist, but they work on generally heavy high-resolution images.

ADVERTISEMENT

This kind of data processing is huge, and not practical for mobile devices. This is why researchers tried to use a smaller image as the source instead, but still "this doesn’t work well in practice; the low-res image just leaves out too much data," the MIT researchers say.

The new system learned how to tweak images by analysing a set of 5,000 photos, all of which had been retouched by five photographers. The system compared the raw, original shot to the retouched version, and understood how to mimick the readjustments.

The new method consists of two new clever tricks. The first is to render changes to the original image as so-called "formulae" (think of them as a layer) rather than masks applied to a high-res image. The formulae simply contain information on how to modify the source image, without having to take and process a picture. It all happens in real time.

As the user is in their device's viewfinder, the system then works out how to apply those formulae to get to the final image. In short, the system never applies the filters on a compressed image, but rather imposes the filtering layers on top of a shot that hasn't been taken yet — and users can see it while framing the shot inside the camera app.

The whole system weighs around 100 megabytes, which means it's light and fast enough to operate on the fly on a mobile device. When the researchers compared their new method to another machine learning algorithm that only uses high-res shots (both input and output), they found their system took one-hundredth of the time to process.

ADVERTISEMENT

"This technology has the potential to be very useful for real-time image enhancement on mobile platforms," said Jon Barron, one of Google's AI scientists. "This paper may provide us with a way to sidestep these issues and produce new, compelling, real-time photographic experiences without draining your battery or giving you a laggy viewfinder experience."

Enhance Your Pulse News Experience!

Get rewards worth up to $20 when selected to participate in our exclusive focus group. Your input will help us to make informed decisions that align with your needs and preferences.

I've got feedback!

JOIN OUR PULSE COMMUNITY!

Unblock notifications in browser settings.
ADVERTISEMENT

Eyewitness? Submit your stories now via social or:

Email: eyewitness@pulse.ng

ADVERTISEMENT
ADVERTISEMENT