Xe8472 is a first-of-its-kind technology that simulates the human visual system
The Human Visual System
The visual perception of humans is a highly complex process. What we ’see‘ is not a representation of technical optical data, it’s an interpretation of this data by the human brain.
Any kind of optical illusions exploit this.
Direct warm sunlight. The human skin and all objects also reflect the ambient light from the blue sky. The camera sees and records this. Even the latest generation of professional Cameras.
The human brain however will compensate for this. What we (humans) actually experience is the same skin color that we would under varying lighting conditions. (within some limits)
simulates these physiological and psychological processes inside the human brain.
Xe8472 deals with a topic called „Color Constancy“ which both NASA and MIT researched for decades with no practical outcome to date. NASA Langley Research Center’s comments on the topic:
„The goal is to make the recorded image look like what you would have seen if you had been observing the scene in person. This means full visual realism analogous to the „concert hall“ sound of audio fidelity.“
Color Constancy is something completely diﬀerent than „automatic color correction“ or „white balance“.
Technology in general is supposed to be smart. In this case, smart enough to acknowledge when the original data should not be altered.
It is no challenge to invent algorithms that correct color shifts, the real challenge is to know if there is a visible color shift at all and to simulate how the human eye-brain would actually perceive it.
And this is precisely what Xe8472 does.
The way bigger challenge than applying such an autonomous algorithm to a single image without getting into problems is to apply the algorithm to moving pictures.
The algorithm was first introduced as Xe847 in 2002 for Photos.
It took more than a decade of R&D to get it right for professional Film.
Therefore it is now called Xe847-2