|
"How do I turn animation on? I have used the clock
-variable
in my scene, but POV-Ray still only calculates one frame."
The easiest way is to just specify the appropriate command line parameter
on the command line or in the command line field in the rendering settings
menu (in the Windows version). For example, if you want to create 20 frames,
type this: +kff20
This will create 20 frames with the clock
variable going
from 0 to 1. The
other command line parameters are found in the POV-Ray documentation.
Ken Tyler has also another good solution for this:
In the directory that you installed POV-Ray into you will find a subdirectory
called scenes and another inside that called animate. You will find several
example files showing you how to write your scene to use the clock variable.
You will still need to activate POV-Ray's animation feature by using an
.ini
file with the correct info or with command line
switches. I personaly
like to use the ini file method. If you try this open the master
povray.ini
file from the tools menu and add the following lines:
;clock=1 ;Initial_Frame=1 ;Final_Frame=20 ;Cyclic_Animation = on ;Subset_Start_Frame=6 ;Subset_End_Frame=9
Save the file and close it. When you need to use the animation feature
simply go in and edit the povray.ini
file and uncomment out the functions
you want to use. At a minimum you will need to use the
initial_frame
and
final_frame
option to make it work. Once you have stopped
rendering your
series of frames be sure to comment out the clock variables in the ini
file. After you have rendered a series of individual frames you will
still need to compile them into the animation format that you wish to
use such as AVI or MPEG. See my »links pages
for programs that can help you do this. POV-Ray has no internal ability to
do this for you except on the Macintosh platform of the program.
The Mac version normally doesn't use .ini
files and lacks any
command line, but uses a completely graphical interface instead. To
activate animation, choose the render settings item from the Edit menu
(right under "Preferences", it will be titled "FILENAME Settings",
where FILENAME is the name of your file), click on the Animation tab,
and enter the needed information in the text boxes.
Short answer: The only way to run POV-Ray on multiple processors is to run several copies of POV-Ray.
Long answer:
Making a program use multiple threads is not as trivial as it may sound. Here are some reasons why it is quite difficult to make with POV-Ray:
An excellent article about the issue can be found at » http://www.acm.org/tog/resources/RTNews/html/rtnv12n2.html#art3.
Here is an answer from John M. Dlugosz with useful tips:
The POV-Ray rendering engine is a single thread of execution, so when run on a dual Pentium Pro (running NT4) the CPU indicator only goes up to about 50%. POV doesn't use more than half the available power on the machine.
That's the basic issue, though to quibble a bit it's not exactly true: the rendering engine soaks up one whole CPU, but the editor runs on its own thread, and operating system functions (writing to the file, updating the display, network activity, system background tasks) run on different threads. This gives a little bit of a bonus, and the system uses as much as 54% of available MIPS when watching it. More importantly, the machine is still highly responsive, and editing or other applications continue on without being sluggish.
But for a long render, it's annoying to have one CPU be mostly idle. What can be done to cut rendering time in half (from 20 hours down to 10, for example)?
The simplest thing is to run two copies of POV on the machine. Have one copy render the top half, and the other render the bottom half. Then paste the halves together in your picture editor.
One thing to watch out for: don't just fire up two copies and point them at the same INI file and image file. They will overwrite each other's output and make a big mess. Instead, you must make sure each is writing to a different file.
For moderate renders, you might let one copy chug away on the long render, and use a second copy interactivly to continue development in POV.
"Is there a way to generate a wireframe output image from a POV scene file?"
Short answer: No.
Long answer:
You have to understand the difference between a modeller like 3D-Studio and POV-Ray in the way they handle objects. Those modellers always use triangle meshes (and some modellers use also NURBS which can be very easily converted into triangles). Triangle meshes are extremely simple to represent in a wireframe format: Just draw a line for each triangle side.
However, POV-Ray handles most of the objects as mathematical entities, not triangle meshes. When you tell POV-Ray to create a sphere, POV-Ray only handles it as a point and a radius, nothing else (besides the possible matrix transform applied to it). POV-Ray only has a notion of the shape of the object as a mathematical formula (it can calculate the intersection of a line and the sphere).
For wireframe output there should be a way to convert that mathematical representation of the object into actual triangles. This is called tesselation.
For some mathematical objects, like the sphere, the box, etc, tesselation is quite trivial. For other entities, like CSG difference, intersection, etc, it's more difficult (although not impossible). For other entities it's completely impossible: infinite non-flat surfaces like paraboloids and hyperboloids (well, actually it is possible if you limit the size of the surface to a finite shape; still the amount of triangles that needs to be created would be extremely high).
There have been lots of discussions about incorporating tesselation into POV-Ray. But since POV-Ray is just a renderer, not a modeller, it doesn't seem to be worth the efforts (adding tesselation to all the primitives and CSG would be a huge job).
(Of course tesselation could give some other advantages, like the ability to fake non-uniform transformations to objects like most triangle mesh modellers do...)
If you just want fast previews of the image, you can try to use the quality parameter of POV-Ray. For example setting quality to 0 (+q0) can give a very fast render. See also the rendering speed question.
"Can I specify variable IOR for an object? Is there any patch that can do this? Is it possible?"
Short answer: No.
Long answer:
There are basically two ways of defining variable IOR for an object: IOR changing on the surface of the object and IOR changing throughout inside the object.
The first one is physically incorrect. For uniform IOR it simulates physical IOR quite correctly since for objects with uniform density the light bends at the surface of the object and nowhere else. However if the density of the object is not uniform but changes throughout its volume, the light will bend inside the object, while travelling through it, not only on the surface of the object.
This is why variable IOR on the surface of the object is incorrect and the possibility of making this was removed in POV-Ray 3.1.
From this we can deduce that a constant IOR is kind of property of the surface of the object while variable IOR is a property of the interior of the object (like media in POV-Ray). Of course the physically correct interpretation of this phenomenon is that IOR is always a property of the whole object (ie. its interior), not only its surface (and this is why IOR is now a property of the interior of the object in POV-Ray); however, the effect of a constant IOR has effect only at the surface of the object and this is what POV-Ray does when bending the rays.
The correct simulation for variable IOR, thus, would be to bend the ray inside the object depending on the density of the interior of the object at each point.
This is much harder to do than one may think. The reasons are similar to why non-uniform transformations are too difficult to calculate reasonably (as far as I know there exists no renderer that calculates true non-uniform transformations; mesh modellers just move the vertices, they don't actually transform the object; a true non-uniform transformation would bend the triangles). Moreover: Non-uniform transformations can be faked if the object is made of many polygons (you can move the vertices as most mesh modellers do), but you can't fake a variable IOR in this way.
Variable IOR is (mostly) impossible to calculate analytically (ie. in a mathematically exact way) at least in a reasonable time. The only way would be to calculate it numerically (usually by super-sampling).
Media in POV-Ray works in this way. It doesn't even try to analytically solve the color of the media, but supersamples the media along the ray and averages the result. This can be pretty inaccurate as we can see with the media method 1 (the only one which was supported in POV-Ray 3.1). However some tricks can be used to make the result more accurate without having to spend too much time, for example antialiasing (which is used by the media method 3 in POV-Ray 3.5). This is a quite easy calculation because the ray is straight, POV-Ray knows the start and end points of the ray and it knows that it doesn't intersect with anything along the ray (so it doesn't have to make ray-object intersection calculations while supersampling).
Variable IOR is, however, a completely different story. Here the program would have to shoot a LOT of rays along the path of the bending light ray. For each ray it would have to make all the regular ray-object intersection calculations. It's like having hundreds or thousands of transparent objects one inside another (with max_trace_level set so high that the ray will go through all of them). You can easily test how slow this is. It's VERY slow.
One could think that "hey, why not just shoot a few tens of rays and then use some kind of antialiasing to get the fine details, like in media method 3".
Well, it might work (I have never seen it tested), but I don't think it will help much. The problem is the inaccuracy of the supersampling (even when using antialiasing). In media it's not a big problem; if a very small shadowed area in the media is not detected by the supersampling process, the result will not differ very much from the correct one (since the shadowed area was so small it would have diminished the brightness of that ray just a bit but no more) and it will probably still look good.
With IOR this isn't anymore true. With IOR even very, very small areas may have very strong effect in the end result, since IOR can drastically change the direction of the ray thus making the result completely different (even very small changes can have great effect if the object behind the current refracting object is far away).
This can have disastrous effects. The ior may change drastically from pixel
to pixel almost at random, not to talk from frame to frame in an animation.
To get a more or less accurate result lots of rays would be needed; just a
few rays is not enough. And shooting lots of rays is an extremely slow
process.
Photon mapping uses forward raytracing (ie. sending rays from light sources) calculate reflecting and refracting light (aka. caustics), which is a new feature in POV-Ray 3.5.
The following is from the homepage of the developer (Nathan Kopp):
"My latest fun addition to POV is the photon map. The basic goal of this implementation of the photon map is to render true reflective and refractive caustics. The photon map was first introduced by Henrik Wann Jensen. It is a way to store light information gathered from a backwards ray-tracing [sic] step in a data structure independent from the geometry of a scene."
It is surprisingly fast and efficient. How is this possible when forward raytracing is so inefficient? For several reasons:
As you have seen, for the photon mapping to work in an acceptable way, you have to tell the program which objects you want to reflect/refract light and which you don't. This way you can optimize a lot the photon mapping step.
|