SORRY weblink not attached before...
See it below!!
http://features.cgsociety.org/story_custom.php?story_id=5434&page=1
"~ Love conquers all, but if love doesn't do it, try hard work ~."
http://www.google.com/profiles/samson13.
Face it... Fight it
SAMSON :-)
Click on the link for the webpage access
"We thought the renders looked pretty real, but when you put on the glasses, you believe it even more because now you can see both sides of the face at the same time because your eyes are far enough apart to take in that information." - Guy Williams, VFX Supervisor
The initial plan was to create a blue jungle but found that though it looked beautiful, it felt contrived. The nail in the coffin was that the frames became too monochromatic in their lighting. If, for example, the day was overcast the light was predominantly blue, as was the radiosity coming from the ground.
The result was blue light with blue highlights with blue reflections on blue skin "and it all got pretty simplistic. It looked real, but had a surreal quality to it, so we decided to get more accents of color back into the jungle, shift parts back to being green so you could have a blue sky filled with green bounce with a yellow key, and it gave a nice balance of light in the scene."
They used Paint Effects to add the final detail, defining the paint effects rules, then would convert it all back to a renderable asset. They had an entire team of people whose job was to do paint effects dressing that sometimes was almost as heavy as the shot. "It made for really lush jungle scenes and gave us so much ability to put so much ability to put so much visual detail into the frame that your mind started to cave in and accept it as being real."
Then there was the biolume. Saindon offered some insight to that process. "We knew the colors, the ideas that Cameron was looking for. Cameron's passion and extensive knowledge of underwater photography is where he draws a lot of his inspiration, so we did a lot's of reference on underwater plants and animals to figure out what the biolume nighttime on Pandora would look like.
We spent a lot of time figuring out each individual plant, then put them together to see what different colors would look like combined. We came up with some very interesting plants but once you make a forest out of this, it was a little crazy; it looked like Las Vegas on steroids. We learned that you had to be very subtle with it. If you overdressed every plant you get something that's hard to watch
And how do you light such an environment, with reflections off the plants and transitions through the leaves to make this huge jungle that completely self-shadows, without doing it traditionally with shadow maps which would have been a complete nightmare? "We had to come up with things like, how do we calculate occlusion, how do we encode that in a relatively efficient manner? Those were the challenges," said Stables.
Imagine a character with lights and shadows, each increasing the complexity. Now imagine the same with a jungle of thousands of trees. "We primarily used image-based lighting, high dynamic range images (HDRI), many of which were shot around New Zealand.
We would accent with area lighting, etc., as needed. By being able to efficiently calculate all the shadowing, all the occlusion for a large environment, then use HDRI to put the illumination in, it enabled us to light very big environments very realistically." The next step was how to render. "We used a couple of techniques," said Saindon. "We did spherical harmonics for building in the lighting for all the different assets so we could look up lighting direction on the plants and get occlusions and lighting directions efficiently and quickly."
Another was stochastic pruning in which, for example, a million-polygon tree close to camera is easy to render, but further away it gets smaller in the buckets within RenderMan, and becomes very difficult to render. "So we used stochastic pruning in which we keep the silhouette of the object but we threw away faces as it moved into the distance." Jon Allitt wrote it as a test and worked so efficiently most TD's didn't even realize it was doing a tremendous amount of work for them.
Using this method, a million poly tree could drop to 20 or 30 polygons. It's a similar process as LOD (Level of Detail), but with LOD you have to build the individual levels of detail and during the transitional period both heavy and lighter models are in the scene and can be heavier to render.
With stochastic pruning, as long as the model is built in such a way that geometry can be thrown away, it can look at the bounding box and determine what can be discarded, and it just throws geometry away as the bounding box gets smaller.
There were bugs in every shot, enough to fill any jungle, virtual or otherwise. "Cameron asked how many bugs we had, saying ?I only see about 20.' We told them there were 20,000," said Williams, with a smile. "Cameron laughed that he had to have 20,000 bugs to see 20." It was because the bugs were incredibly small, ranging from the size of a wasp down to a mosquito, and would recede into the distance by about 300 feet.
Many were hidden by the plants and the rest were hidden in the dark, so you could only see the ones flying through the god rays. Cameron had them boost the bug count to 60,000 so he could see 60 at a time. Luckily, they rendered quickly.
"The only problem we had with the bugs is every now and then one would fly right into the lens of the camera. Some camera hog would loiter in front of the lens and hurt your stereo space and your perception of the shot." Because of how Weta does their compositing, "painting out stuff is not too easy to do, so we had to fix it in Renders. We didn't render with holdout mattes; we rendered using depth sorting information using deep opacities. The advantage of that is we could render the character again and again, even if we changed the motion, and not have to rerender the jungle. But the disadvantage is we had to carry around extra data sets of distance from the camera and use that to stitch it together. If you ever wanted to paint out a bug, it was really hard to paint out depth information. You could remove the bug but you would still see its ghost in the fog or in the depth of field, all of which keys off the deep opacities."
The camera, informally nicknamed the Steering Wheel or Virtual Camera, was like a steering wheel that had an LCD monitor. It doesn't actually have a camera, it only has motion control data so as Cameron looked through what was equivalent to the viewfinder, he saw Pandora, the Na'vi, the Avatars, in real time and with no lag. It allowed him to treat the process like a live action shoot. The Steering Wheel was created specifically for this film, as was much of the technology, including the practical 3D camera.
The Virtual Art Department (VAD) would come up with an environment for a certain scene, then Cameron would use the Steering Wheel to walk around the environment with his art department crew and point out needed adjustments, like moving a particular tree so he could shoot from a preferred angle. The VAD artists sat on stage with him so they could immediately move the assets on Cameron's request. It allowed him to scout the location of the digital set with the camera using similar methodologies as if it were a live-action shoot, and the equivalent of telling a Greens Department how he wanted to dress out the scene. It was like a live action movie with all the benefits of a CG movie.
"Doing everything in 3D had a big impact on our methodology," said Saindon. Because Avatar is a 3D movie all the fluid sims and explosions had to be in volumetrics, from fire and smoke to water. "Usually we shoot everything; we go to the back lot, shoot FX and slap them in.
In this movie I don't think we used any practical effects, everything we did was done in 3D and rendered. We've pushed the dynamic effects a lot further on this movie than we have in the past," Saindon explained. "We could no longer cheat, we couldn't use actual footage.
"Jim was very smart at the interoccular and convergence. In a wide shot he set it very narrow; in wide shots you don't see a huge amount of 3D. In a close shot, he would set the IO a little wider. He didn't go crazy with the convergence, he didn't try to pull things off the screen. He set your eye where it needed to be for the shot and didn't try to go crazy with the 3D gags. I always felt 3D was a gimmick, I figured if the movie wasn't good enough to hold up in 2D I wouldn't bother. I have to say the 3D that Jim came up with is drastically better than anything I've seen."
I have to agree. When, on the first sortie, they leave the human outpost called Hellsgate and enter the jungle, pilot Trudy Chacon, played by the feisty actor Michelle Rodriguez, comments, "You should see your faces." I am sure I wore the same expression. It makes you wonder which side of the gate heaven and hell is on.
No matter, I'm moving to Pandora. Please forward my mail.
Avatar Official Link
Weta Digital
Wayne Stables, VFX Supervisor
Eric Saindon, VFX Supervisor
Guy Williams, VFX Supervisor
Jon Allitt
Discuss this article on CGTalk
"~ Love conquers all, but if love doesn't do it, try hard work ~."
http://www.google.com/profiles/samson13.
Face it... Fight it
SAMSON :-)
--
You received this message because you are subscribed to the Google Groups "Samsongroup" group.
To post to this group, send email to samsongroup@googlegroups.com.
To unsubscribe from this group, send email to samsongroup+unsubscribe@googlegroups.com.
For more options, visit this group at http://groups.google.com/group/samsongroup?hl=en.