If you have great lighting, a good photographer can take decent photos even with the crappiest camera imaginable. In low light, though, all bets are off. Sure, some cameras can shoot haunting video lit only by the light of the moon, but for stills — and especially stills shot on a smartphone — digital noise continues to be a scourge. We may be getting close to what is possible to achieve with hardware; heat and physics are working against us making even better camera sensors. But then Google Research came along, releasing an open source project it calls MultiNerf, and I get the sense that we’re at the precipice of everything changing.
I can write a million words about how awesome this is, but I can do better; here’s a 1-minute-51-second video, which, at 30 frames per second and “a picture tells a thousand words,” is at least 1.5 million words worth of magic:
Video Credits: DIYPhotography
The algorithms run on raw image data and adds AI magic to figure out what footage “should have” looked like without the distinct video noise generated by imaging sensors.