|
![](/i/fill.gif) |
"Edouard Poor" <pov### [at] edouard info> wrote:
> All I'm doing is moving the camera to a random offset each render. The offset is
> chosen from the iris image I've superimposed in the bottom right. This one is a
> pentagram, but any bitmap can be chosen.
Interesting technique. My approach in the past was a physical aperture + focal
blur, or an ugly patch to change the sampling. I was surprised how long it
took even for simple spheres, as you have here.
> so that the final composition of all the effects is faster than a single render
> with all the elements turned on
With a single render you can use adaptive sampling much like focal blur, but
then you lose out on some of the effects you've mentioned. Is it really
faster? Maybe it's the *combination* of effects that slows it down in a single
render. Interesting. If nothing else, this method is sure easy enough to
parallelize!
> The picture shows that my code still has some issues - out of focus elements in
> certain areas (towards the edge of the picture near the axes) have some odd
> distortions. I'm going to have to track down the problem on those...
If you mean the lower left, I think it looks alright. The distortion seems
natural due to the camera angle, and if the specular reflection is distorted,
then a convolution with the aperture image will show the same effect.
- Ricky
Post a reply to this message
|
![](/i/fill.gif) |