Displaying 3D Polygon Animations
API
Games use software known as an API (Application Programme Interface) which is a set of tools for building software applications a good API makes it easier for the software to be developed as it gives you all the building blocks you need to make the software you're trying to make
Most Operating Systems such as Windows provide an API so that programmers can make applications for that operating system although API's are for programmers they benefit the users as applications made on the same API will probably have similar interfaces
Direct 3D
Direct 3D is an API designed for manipulating and displaying 3D objects it was developed by Microsoft Direct 3D provides programmers with a way for programmers to utilize any graphics card in a pc and use it to display objects almost all pc's are compatible with Direct 3D
Open GL
Open GL was developed by silicon graphics in the early 90's and has become one of the most widely used graphics api's in the world it is very similar to Direct 3D however it is open source meaning anyone can edit the code and do what they want with it whereas Direct 3D limits you to what Microsoft says what you can do
Graphics Pipeline
The graphics pipeline is the way that a computer transferees the mathematical data that it has on the object into the object that we see on the screen the 3D graphics pipeline typically takes a 3D object when it's in data and converts it into a 2D raster image Open GL and direct 3D both have very similar graphics pipelines
Stages of the graphics pipeline
First the scene is created out of geometric primitive shapes this is usually done using triangles as they are good for this as they always exist on a single plane
At this stage anything that can not be seen by the camera is not shown and discarded
Scan Conversion or Rasterization
This is the stage where the 2D image is converted into a raster format and the resonating pixel values are determined from now on
Texturing, Fragment Shading
At this stage the pre pixels are assigned a colour based values interloped based on vertices during rasterization from a texture that's in memory or from a shader programme
Display
At this point the final coloured pixels are displayed on the screen
API
Games use software known as an API (Application Programme Interface) which is a set of tools for building software applications a good API makes it easier for the software to be developed as it gives you all the building blocks you need to make the software you're trying to make
Most Operating Systems such as Windows provide an API so that programmers can make applications for that operating system although API's are for programmers they benefit the users as applications made on the same API will probably have similar interfaces
Direct 3D
Direct 3D is an API designed for manipulating and displaying 3D objects it was developed by Microsoft Direct 3D provides programmers with a way for programmers to utilize any graphics card in a pc and use it to display objects almost all pc's are compatible with Direct 3D
Open GL
Open GL was developed by silicon graphics in the early 90's and has become one of the most widely used graphics api's in the world it is very similar to Direct 3D however it is open source meaning anyone can edit the code and do what they want with it whereas Direct 3D limits you to what Microsoft says what you can do
Graphics Pipeline
The graphics pipeline is the way that a computer transferees the mathematical data that it has on the object into the object that we see on the screen the 3D graphics pipeline typically takes a 3D object when it's in data and converts it into a 2D raster image Open GL and direct 3D both have very similar graphics pipelines
Stages of the graphics pipeline
First the scene is created out of geometric primitive shapes this is usually done using triangles as they are good for this as they always exist on a single plane
Modelling and Transformation
This stage transforms the local objects consternates into the 3D world coordinate system
Camera Transformation
Next it transforms the 3D world Coordinates into 3D Camera Coordinates with the camera as the origin
Lighting
Illuminates according to the lighting and reflectiveness of the object for example if the room as pitch black the objects will be seen as black
Project Transformation
This stage transfares the 3D coordinates into a 2D view of the camera a object further away from the camera looks smaller and one's that are closer up look larger this is caused by the x and y coordinates of each of the objects being divided by it's z coordinate (this represents it's distance from the camera) in orthographic projection objects retain there original size regardless of distance from the camera
Clipping
Scan Conversion or Rasterization
This is the stage where the 2D image is converted into a raster format and the resonating pixel values are determined from now on
Texturing, Fragment Shading
At this stage the pre pixels are assigned a colour based values interloped based on vertices during rasterization from a texture that's in memory or from a shader programme
Display
At this point the final coloured pixels are displayed on the screen
Sources:
https://encrypted-tbn0.gstatic.com/images?q=tbn:ANd9GcRGqZWF84sbKCSWaWHpQl4RCXWdRUIzP7lMmDgZLQ5Hfc9Lt-z7
No comments:
Post a Comment