We are going to talk about techniques that create physically based material models from photographs that we can use in our light simulation programs. In an earlier work, two photographs are required for high-quality reconstruction. It seems that working from only one photograph doesn't seem possible at all. However, with the power of deep learning...
___________________________________
The paper "Two-Shot SVBRDF Capture for Stationary Materials" is available here:
https://mediatech.aalto.fi/publications/graphics/TwoShotSVBRDF/
The paper "Reflectance Modeling by Neural Texture Synthesis" is available here:
https://mediatech.aalto.fi/publications/graphics/NeuralSVBRDF/
NVIDIA has implemented the two-shot model! Have a look:
https://twitter.com/karoly_zsolnai/status/839570124017438726
Our earlier episode on Gradient Domain Light Transport is available here:
https://www.youtube.com/watch?v=sSnDTPjfBYU
The light transport course at the Technical University of Vienna is available here:
https://www.youtube.com/playlist?list=PLujxSBD-JXgnGmsn7gEyN28P1DnRZG7qi
WE WOULD LIKE TO THANK OUR GENEROUS PATREON SUPPORTERS WHO MAKE TWO MINUTE PAPERS POSSIBLE:
David Jaenisch, Sunil Kim, Julian Josephs, Daniel John Benton, Dave Rushton-Smith, Benjamin Kang.
https://www.patreon.com/TwoMinutePapers
We also thank Experiment for sponsoring our series. - https://experiment.com/
Image credits:
Sherrie Thai - https://flic.kr/p/7pNLqB
Giulio Bernardi - https://flic.kr/p/b2GMJ
Andy Beatty - https://flic.kr/p/7j3dUX
Open Grid Scheduler - https://flic.kr/p/Gdg3JC
liz west - https://flic.kr/p/rAbED
http://collagefactory.blogspot.hu/2010/04/brdf-for-diffuseglossyspecular.html
Subscribe if you would like to see more of these! - http://www.youtube.com/subscription_center?add_user=keeroyz
Music: Dat Groove by Audionautix is licensed under a Creative Commons Attribution license (https://creativecommons.org/licenses/by/4.0/)
Artist: http://audionautix.com/
Splash screen/thumbnail design: Felícia Fehér - http://felicia.hu
Károly Zsolnai-Fehér's links:
Facebook → https://www.facebook.com/TwoMinutePapers/
Twitter → https://twitter.com/karoly_zsolnai
Web → https://cg.tuwien.ac.at/~zsolnai/
Оглавление (2 сегментов)
Segment 1 (00:00 - 05:00)
Dear Fellow Scholars, this is Two Minute Papers with Károly Zsolnai-Fehér. If you are new here, this is a series about research with the name Two Minute Papers, but let's be honest here. It's never two minutes. We are going to talk about two really cool papers that help us create physically based material models from photographs that we can use in our light simulation programs. Just as a note, these authors, Miika and Jaakko have been on a rampage for years now and have popped so many fantastic papers each of which I was blown away by. For instance, earlier, we talked about their work on Gradient Domain Light Transport, brilliant piece of work, I've put a link in the description box, make sure to check it out! So the problem we're trying to solve is very simple to understand: the input is a photograph of a given material somewhere in our vicinity, and the output is a bona fide physical material model that we can use in our photorealistic rendering program. We can import real world materials in our virtual worlds, if you will. Before we proceed, let's define a few mandatory terms: A material is diffuse if incoming light from one direction is reflected equally in all This means that they look the same from all directions. White walls and matte surfaces are excellent examples of that. A material we shall consider specular if incoming light from one direction is reflected back to one direction. This means that if we turn our head a bit, we will see something different. For instance, the windshield of a car, water and reflections in a mirror can be visualized with a specular material model. Of course, materials can also be a combination of both. For instance, car paint, our hair and skin are all combinations of these material models. Glossy materials are midway between the two where the incoming light from one direction is reflected to not everywhere equally, and not in one direction, but a small selected set of directions. They change a bit when we move our head, but not that much. In the Two-Shot Capture paper, a material model is given by how much light is reflected and absorbed by the diffuse and specular components of the material, and something that we call a normal map, which captures the bumpiness of the material. Other factors like glossiness and anisotropy are also recorded, but we shall focus on the diffuse and the specular parts. The authors ask us to grab our phone for two photographs of a material to ensure a high-quality reconstruction procedure: one with flash, and one without. And the question immediately arises: why two images? Well, the image without flash can capture the component that looks the same from all directions, this is the diffuse component, and the photograph with flash can capture the specular component because we can see how the material handles specular reflections. And it is needless to say, the presented results are absolutely fantastic. So, first paper, two images, one material model. And therein lies the problem, which they tried to address in the second paper. If a computer looks at such an image, it doesn't know which part of one photograph is the diffuse and which is the specular reflection. However, I remember sitting in the waiting room of a hospital while reading the first paper, and this waiting room had a tiled glossy wall, and I was thinking that one image should be enough, because if I look at something, I can easily discern what the diffuse colors are, and which part is the specular reflection of something else. I don't need multiple photographs for that. I can also immediately see how bumpy it is, even from one photograph, I don't need to turn my head around. This is because we, humans have not a mathematical, but an intuitive understanding of the materials we see around us. So can we explain the same kind of understanding of materials to a computer somehow? Can we do it with only one image? And the answer is, yes we can, and, hopefully, we already feel the alluring call of neural networks. We can get a neural network that was trained on a lot of different images to try to guess what these material reflectance parameters should look like. However, the output should not be one image, but multiple images with the diffuse and specular reflectance informations, and a normal map to describe the bumpiness of this surface. Merely throwing a neural network at this problem, is however, not sufficient. There needs to be some kind of conspiracy between these images, because real materials are not arbitrarily put together. If one of these images is smooth, or has interesting features somewhere, the others have to follow it in some way. This "some way" is mathematically quite challenging to formulate, which is a really cool part
Segment 2 (05:00 - 06:00)
of the paper. This conspiracy part is a bit like if we had 4 criminals testifying at a trial, where they try to sell their lies, and to maintain the credibility of their made up story, they have previously had to synchronize their lies so they line up correctly. The paper contains neat tricks to control the output of the neural network and create these conspiracies across these multiple image outputs that yield a valid and believable material model. And the results, are again, just fantastic. Second paper, one image, one material model. It doesn't get any better than that. Spectacular, not specular... spectacular piece of work. The first paper is great, but the second is smoking hot, by all that is holy, I am getting goosebumps. If you are interested in hearing a bit more about light transport and are not afraid of some mathematics, we recently recorded my full course on this at the Technical University of Vienna, the entirety of which is freely available for everyone. There is a link for it in the video description box, make sure to check it out! Thanks for watching, and for your generous support, and I'll see you next time!