Skip to main content

Novel de-noising method generates sharper photorealistic images faster

Monte Carlo computational methods are behind many of the realistic images in games and movies. They automate the complexities in simulating the physics of lights and cameras to generate high-quality renderings from samples of diverse image features and scenes. But the process of Monte Carlo rendering is slow and can take hours -- or even days -- to produce a single image, and oftentimes the results are still pixelated, or "noisy."
A global team of computer scientists from MIT, Adobe, and Aalto University has developed an innovative method for producing higher-quality images and scene designs in much less time by using a deep-learning-based approach that considerably cuts the noise in images. Their method results in sharper images that effectively capture intricate details from sample features, including complex lighting components like shadowing, indirect lighting, motion blur, and depth of field.
The researchers are set to present their work at SIGGRAPH 2019, held 28 July-1 August in Los Angeles. This annual gathering showcases the world's leading professionals, academics, and creative minds at the forefront of computer graphics and interactive techniques.
"Our algorithm can produce clean images from noisy input images with very few samples, and could be useful for producing quick rendered previews while iterating on scene design," says study lead author Michaël Gharbi, research scientist at Adobe. Gharbi began the research as a PhD student at MIT in the lab of Frédo Durand, who also is a coauthor.
The team's work focuses on so-called "denoising," a post-processing technique to reduce image noise in Monte Carlo rendering. It essentially retains the details of an image and removes anything that detracts from its sharpness. In previous work, computer scientists have developed methods that smooth the noise out by taking the average from the pixels in a sample image and neighboring pixels.
"This works reasonably well, and several movies have actually used this in production," notes coauthor Tzu-Mao Li, a recent PhD graduate from MIT who also studied under Durand. "However, if the images are too noisy, oftentimes the post-processing methods are not able to recover clean and sharp images. Usually users still need hundreds of samples per pixel on average for an image with reasonable quality -- a tedious, time-consuming process."
Somewhat comparable is the process of editing a photo in a graphics software program. If a user is not working from the original, raw file, altered versions of the photo will likely not result in a clear, sharp, high-res final image. A similar yet more complex problem is image denoising.
To this end, the researchers' new computational method involves working with the Monte Carlo samples directly, instead of average, noisy images where most information has already been lost. Unlike typical deep learning methods that deal with images or videos, the researchers demonstrate a new type of convolutional network that can learn to denoise renderings directly from the raw set of Monte Carlo samples rather than from the reduced, pixel-based representations.
A key part of their work is a novel kernel-predicting computational framework that "splats" individual samples -- colors and textures -- onto nearby pixels to sharpen the overall composition of the image. In traditional image processing, a kernel is used for blurring or sharpening. Splatting is a technique that addresses motion blur or depth-of-field issues and makes it easier to even out a pixelated area of a sample.
In this work, the team's splatting algorithm generates a 2D kernel for each sample, and "splats" the sample onto the image. "We argue that this is a more natural way to do the post-processing," says Li. The team trained their network using a random scene generator and extensively tested their method on a variety of realistic scenes, including various lighting scenarios such as indirect and direct illumination.
"Our method gives cleaner outputs at very low sample counts, where previous methods typically struggle," adds Gharbi.
In future work, the researchers intend to address scalability with their method to extend to more sample features and explore techniques to enforce frame-to-frame smoothness of the denoised images.
The paper, "Sample-based Monte Carlo Denoising Using a Kernel-Splatting Network," is also coauthored by Miika Aittala at MIT and Jaakko Lehtinen at Aalto University and Nvidia. 
Story Source:
Materials provided by Association for Computing Machinery
Note: Content may be edited.

Comments

Popular posts from this blog

Dark matter may be older than the Big Bang

Dark matter, which researchers believe make up about 80% of the universe's mass, is one of the most elusive mysteries in modern physics. What exactly it is and how it came to be is a mystery, but a new Johns Hopkins University study now suggests that dark matter may have existed before the Big Bang. The study, published August 7 in  Physical Review Letters , presents a new idea of how dark matter was born and how to identify it with astronomical observations. "The study revealed a new connection between particle physics and astronomy. If dark matter consists of new particles that were born before the Big Bang, they affect the way galaxies are distributed in the sky in a unique way. This connection may be used to reveal their identity and make conclusions about the times before the Big Bang too," says Tommi Tenkanen, a postdoctoral fellow in Physics and Astronomy at the Johns Hopkins University and the study's author. While not much is known about its origins,...

Home births as safe as hospital births: International study suggests

A large international study led by McMaster University shows that low risk pregnant women who intend to give birth at home have no increased chance of the baby's perinatal or neonatal death compared to other low risk women who intend to give birth in a hospital. The results have been published by  The Lancet 's  EClinicalMedicine  journal. "More women in well-resourced countries are choosing birth at home, but concerns have persisted about their safety," said Eileen Hutton, professor emeritus of obstetrics and gynecology at McMaster, founding director of the McMaster Midwifery Research Centre and first author of the paper. "This research clearly demonstrates the risk is no different when the birth is intended to be at home or in hospital." The study examined the safety of place of birth by reporting on the risk of death at the time of birth or within the first four weeks, and found no clinically important or statistically different risk between home...

GSAT-11 satellite to be launched from French Guiana on Dec 5th

GSAT-11 satellite to be launched from French Guiana on Dec 5th GSAT-11 would be located at 74 East and is the fore-runner in a series of advanced communications satellite with multi-spot beam antenna coverage over Indian mainland and Islands, ISRO said. GSAT-11 is the next generation “high throughput” communication satellite configured around ISRO’s I-6K Bus. (PTI/Representational). Indian space agency ISRO is scheduled to launch GSAT-11, the “heaviest” satellite built by it, on-board Ariane-5 rocket of Arianespace from French Guiana on December 5. Weighing about 5,854 kg, GSAT-11 would play a vital role in providing broadband services across the country, and also provide a platform to demonstrate new generation applications, the Indian Space Research Organisation (ISRO) said. It is the “heaviest” satellite built by ISRO, the space agency said. GSAT-11 is the next generation “high throughput” communication satellite configured around ISRO’s  I-6K Bus, and it...