Fixing Our Trust in Content

According to a recent study, over 63% of Americans say they often come across fake images online. Whether those images are truly faked or perceived as such, the damage is done. We are rapidly losing trust in what we see, our evolutionary number source of information.

If this trend continues, photography and video will soon be seen as purely artistic mediums, alongside painting or sketching. Adobe, with its year-old initiative, the Content Authenticity Initiative, plans to combat this trend and recently displayed how.

Interestingly enough, Adobe did not rely on the industry’s historical mammoth ( Canon, Nikon, Leica,…) to showcase the practical application of its initiative but rather a partnership made of a start-up, TruePic, and a microchip manufacturer, Qualcomm. It’s not a mishap but rather a reflection of the current photography landscape: 85% of the trillion images uploaded online today are taken with a cellphone. Qualcomm provides chips to some of the most popular cellphones in the world, and the Truepic solution is mobile native.

Armed with an identified mobile camera, award-winning photographer Sara Naomi Lewkowicz took a series of pictures embedded with the new feature. The result is images that contain information on who produced the content and what alterations were done to it since its creation.

The first CAI-enabled images were taken by Sara Naomi Lewkowicz of Maresa D’Amore-Morrison in New York City on November 15.

For those who would like to explore deeper, the “view more” opens up a verify page online displaying the previous version’s thumbnails. It also contains the signature of the capture device, which will need to be vetted by a yet established consortium to be accepted as trusted.

The verify page of the Adobe Content Authenticity Initiative.

Also, participating in this soft-launched were Twitter and the New York Times, both co-founders of this initiative but still inactive on its implementation. For good reasons. While Adobe has been generously sharing demonstration of the Initiative capabilities, there is still a lot to happen before the general public consumes this. First and foremost is a wide enough industry acceptance and implementation.

Key to this initiative is the realization that technology will not solve the issue of trust in images. As we have seen, tools used to deceive evolve much faster than those built to catch them. Adobe is well aware of this issue, thanks to Photoshop, probably the number one image falsifying tool. Rather, knowing who created the image and what alterations have been done is enough to provide decision-making information to the viewer. In time, that decision could be made by an AI.

The CAI is not for everyone. The majority of photos will never need it. Your aunt Annie will not need it when you send her pictures of her nephew playing soccer. Nor would your friends on Facebook really doubt that those images you took of your latest vacation in Greece are fake. If they do, there is a deeper relationship issue to be solved here.

It will not be a requirement for images to be published. The CAI will find its home in professional photography like photojournalism and documentary as well as a myriad of business applications like insurance or e-commerce. In time, it will make all photography more trustworthy as we learn to better understand the relationship between content, context, and motivation.


About the author: Paul Melcher is a photography and technology entrepreneur based in New York, and the founder of Kaptur, a news magazine about the visual tech space. The opinions expressed in this article are solely those of the author. You can find more of his writings on his blog, Thoughts of a Bohemian. Melcher offers his services as a consultant as well. This article was also published here.


Image credits: Header photo by The Creative Exchange