The basic rendering pipeline is usually diagrammed as a series of transformations through several identifiable spaces. An object is defined in its own space, its object space. Objects are transformed into positions and orientations in world space. This is the same space in which light sources are placed as well as the observer or camera. The observer parameters include position, view direction (sometimes specified with a center of interest in which case the view direction is the vector from the observer to the center of interest) and tilt.
Object space and world space are typically right-handed coordinate systems. For purposes of our discussions, the world space coordinate system will have the positive x-axis pointing to the right, the positive y-axis pointing up and the z-axis points away from the observer representing depth. This is different from some application areas in which the x-y plane is the ground plane and z is altitude. It doesn't make any difference which convention is adopted as long as the animator and programmer agree on the convention.
In preparation for the perspective transformation, objects are usually transformed from world space to eye space in which the eye is at the origin looking down the positive z-axis in right-handed space. The perspective transformation transforms object definitions from eye space to image space. Visible extents in image space are usually standardized into the minus one to plus one range in x and y, and from zero to one in z. Image space is then scaled and translated into screen space by making the visible ranges in x and y coincide with the display coordinates in screen space; z can be left alone. So that we have:
object space -> world space -> eye space -> image space -> screen space
Ray tracing differs from the above sequence of transformations in that the act of tracing rays through space implicitly accomplishes the perspective transformation. If the rays are constructed in world space based on pixel coordinates (the usual approach) then the progression through spaces for ray tracing reduces to:
object space -> world space -> screen space
In either case, animation is typically produced by modifying the object space to world space transformation of objects over time or, in the case of walk-throughs or fly-bys, by transforming the observer in world space. Object space to world space transformations can be kept in 4x4 transformation matrices. The 4x4 matrix is the smallest matrix that can represent all of the relevant transformation and, because it is a square matrix, has a good chance at having a computable inverse (in fact, in our case of transformations, it always has a computable inverse) and can be concatenated with other transformations matrices to produce compound transformations while still maintaining 4x4ness. The 4x4 identity matrix has zeros everywhere except along its diagonal wherein lies ones.
As a review, computer graphics typically pre multiplies a 4x1 column matrix which represents a point, by the 4x4 transformation matrix to produce a transformed 1x4 matrix. Except when doing perspective transformation, the fourth element of the column point vector is unity and the first three elements are the x, y and z coordinates of the point. The translation matrix has x, y, and z translation values as the first three values of the fourth column. The uniform scale matrix that scales by 'S' is the identity matrix with 1/S as the fourth element of the 4th column. Non-uniform scale is the identity matrix with Sx, Sy, and Sz as the first three elements along the diagonal. Shearing is merely a rotation followed by a non-uniform scale.
Alternatively, a 3x3 rotation matrix and a translation vector can be used to control the object to world space transformation. The 4x4 matrix referred to above is simply the 3x3 in the upper left corner of the 4x4 with the transpose of the translation vector as the row underneath it with a column of [0 0 0 1] completing the values on the right side of the matrix.
In some books, row point matrices are postmultiplied by 4x4 transformation matrices which is the transpose of the 4x4 matric described above. Everything mentioned still applies; all matrices using this convention are merely transposes of the previous convention, including the vectors.
For the non-ray tracing case, the transformations which take an object from its defining space into the screen have the following characteristics:
|Object Transform||A series of scales, rotations, and translations which are specified by the user to place a copy of the object in world space|
|World-to-Eye Transfrom||A rigid body transformation which transforms the observer position to the origin and the view vector to align with the positive z-axis in left-handed space.|
|Perspective Matrix||This produces Z*tan(theta) in the fourth component of a homogeneous representation. In addition, the visible range of x and y values are mapped into the range [-1 +1] and the visible range in z is mapped into the range [0 +1].|
|Perspective Divide||Each point produced by the Perspective Matrix multiplication has a non-unit fourth component which represents the perspective transform. Dividing each point by its W component (of an [x y z w] representation] completes the perspective transformation.|
|Image to Screen Space Mapping||This maps the visible range in x and y produced above ([-1 +1]) to the screen coordinate syste, [0 500] for example.|
Consider the case of the moon orbiting the earth. For sake of simplicity, let's assume that the center of the earth is at the origin, and initially the moon data is defined with its center at the origin. We have three approaches that could be taken and will illustrate various effects of roundoff error.
First the moon data could be transformed out to its orbit position, let's say (r,0,0). For each frame of animation, we could apply a delta y-axis transformation matrix to the moon's points where each delta represents the angle it move in one frame. Roundoff errors will accumulate in the world-space object points. Points which began as coplanar will no longer be coplanar. This can have undesirable effects, especially in display algorithms which linearly interpolate values along a surface.
The second approach is to build a y-axis transformation matrix that will take the object space points into their current world-space points. For each frame, we concatenate a delta y-axis transformation matrix with the current transformation matrix and then apply that resultant matrix to the moon's points. Roundoff error will accumulate in the transformation matrix. Over time, the matrix will deviate from representing a rigid transformation. Shearing effects will begin to creep into the transformation and angles will cease to be preserved.
The third approach is add the delta value to an accumulating angle variable and then build the y-axis rotation matrix from that angle parameter. This would then be concatenated with the x-axis translation matrix and the resultant matrix would be applied to the original moon points in object space. In this case, roundoff error will accumulate in the angle variable and the angle of rotation may begin to deviate from what is desired. This may have undesirable effects when trying to coordinate motions, but the transformation matrix, which is built anew every frame, will not accumulate any errors itself. The transformation will always represent a valid rigid transformation with planarity and angles being preserved.
Usually the default orientation of the observer is 'head up'. That is, the observer usually has his up vector in the plane formed by the view vector and the y-axis. This, of course, is undefined for straight up and straight down views and must be dealt with as special cases or simply avoided. This default orientation means that if the observer has a fixed center of interest and the observer's position arcs directly over the center of interest, then just before and just after being directly overhead, the observer's up vector will instantaneously rotate by almost one hundred eighty degrees.
In addition to the observer position and orientation, the field of view has to be specified. This includes a viewing angle, hither clipping distance, and yon clipping distance.