ParaView/UsersGuide/Recommendations: Difference between revisions

From KitwarePublic
Jump to navigationJump to search
mNo edit summary
(Replaced content with "{{ParaView/Template/DeprecatedUsersGuide}}")
 
(2 intermediate revisions by one other user not shown)
Line 1: Line 1:
'''1.4.1 Avoiding Data Explosion'''
{{ParaView/Template/DeprecatedUsersGuide}}
 
The pipeline model that ParaView presents is very convenient for exploratory visualization. The loose coupling between components provides a very flexible framework for building unique visualizations, and the pipeline structure allows you to tweak parameters quickly and easily.
 
The downside of this coupling is that it can have a larger memory footprint. Each stage of this pipeline maintains its own copy of the data. Whenever possible, ParaView performs shallow copies of the data so that different stages of the pipeline point to the same block of data in memory. However, any filter that creates new data or changes the values or topology of the data must allocate new memory for the result. If ParaView is filtering a very large mesh, inappropriate use of filters can quickly deplete all available memory. Therefore, when visualizing large data sets, it is important to understand the memory requirements of filters.
 
Please keep in mind that the following advice is intended only for when dealing with very large amounts of data and the remaining available memory is low. When you are not in danger of running out of memory, ignore all of the following advice.
 
When dealing with structured data, it is absolutely important to know what filters will change the data to unstructured. Unstructured data has a much higher memory footprint, per cell, than structured data because the topology must be explicitly written out. There are many filters in ParaView that will change the topology in some way, and these filters will write out the data as an unstructured grid, because that is the only data set that will handle any type of topology that is generated. The following list of filters will write out a new unstructured topology in its output that is roughly equivalent to the input. These filters should never be used with structured data and should be used with caution on unstructured data.
 
* '''Append Datasets'''
* '''Append Geometry'''
* '''Clean'''
* '''Clean to Grid'''
* '''Connectivity'''
* '''D3'''
* '''Delaunay 2D/3D'''
* '''Extract Edges'''
* '''Linear Extrusion'''
* '''Loop Subdivision'''
* '''Reflect'''
* '''Rotational Extrusion'''
* '''Shrink'''
* '''Smooth'''
* '''Subdivide'''
* '''Tessellate'''
* '''Tetrahedralize'''
* '''Triangle Strips'''
* '''Triangulate'''
 
Technically, the '''Ribbon''' and '''Tube''' filters should fall into this list. However, as they only work on 1D cells in poly data, the input data is usually small and of little concern.
 
This similar set of filters also output unstructured grids, but they also tend to reduce some of this data. Be aware though that this data reduction is often smaller than the overhead of converting to unstructured data. Also note that the reduction is often not well balanced. It is possible (often likely) that a single process may not lose any cells. Thus, these filters should be used with caution on unstructured data and extreme caution on structured data.
 
* '''Clip'''
* '''Decimate'''
* '''Extract Cells by Region'''
* '''Extract Selection'''
* '''Quadric Clustering'''
* '''Threshold'''
 
Similar to the items in the preceding list, '''Extract Subset''' performs data reduction on a structured data set, but also outputs a structured data set. So the warning about creating new data still applies, but you do not have to worry about converting to an unstructured grid.
 
This next set of filters also outputs unstructured data, but it also performs a reduction on the dimension of the data (for example 3D to 2D), which results in a much smaller output. Thus, these filters are usually safe to use with unstructured data and require only mild caution with structured data.
 
* '''Cell Centers'''
* '''Contour'''
* '''Extract CTH Fragments'''
* '''Extract CTH Parts'''
* '''Extract Surface'''
* '''Feature Edges'''
* '''Mask Points'''
* '''Outline (curvilinear)'''
* '''Slice'''
* '''Stream Tracer'''
 
These filters do not change the connectivity of the data at all. Instead, they only add field arrays to the data. All the existing data is shallow copied. These filters are usually safe to use on all data.
 
* '''Block Scalars'''
* '''Calculator'''
* '''Cell Data to Point Data'''
* '''Curvature'''
* '''Elevation'''
* '''Generate Surface Normals'''
* '''Gradient'''
* '''Level Scalars'''
* '''Median'''
* '''Mesh Quality'''
* '''Octree Depth Limit'''
* '''Octree Depth Scalars'''
* '''Point Data to Cell Data'''
* '''Process Id Scalars'''
* '''Random Vectors'''
* '''Resample with dataset'''
* '''Surface Flow'''
* '''Surface Vectors'''
* '''Texture Map to...'''
* '''Transform'''
* '''Warp (scalar)'''
* '''Warp (vector)'''
 
This final set of filters are those that either add no data to the output (all data of consequence is shallow copied) or the data they add is generally independent of the size of the input. These are almost always safe to add under any circumstances (although they may take a lot of time).
 
* '''Annotate Time'''
* '''Append Attributes'''
* '''Extract Block'''
* '''Extract Datasets'''
* '''Extract Level'''
* '''Glyph'''
* '''Group Datasets'''
* '''Histogram'''
* '''Integrate Variables'''
* '''Normal Glyphs'''
* '''Outline'''
* '''Outline Corners'''
* '''Plot Global Variables Over Time'''
* '''Plot Over Line'''
* '''Plot Selection Over Time'''
* '''Probe Location'''
* '''Temporal Shift Scale'''
* '''Temporal Snap-to-Time-Steps'''
* '''Temporal Statistics'''
 
There are a few special case filters that do not fit well into any of the previous classes. Some of the filters, currently '''Temporal Interpolator''' and '''Particle Tracer''', perform calculations based on how data changes over time. Thus, these filters may need to load data for two or more instances of time, which can double or more the amount of data needed in memory. The '''Temporal Cache''' filter will also hold data for multiple instances of time. Also keep in mind that some of the temporal filters such as the temporal statistics and the filters that plot over time may need to iteratively load all data from disk. Thus, it may take an impractically long amount of time even if does not require any extra memory.
 
The '''Programmable Filter''' is also a special case that is impossible to classify. Since this filter does whatever it is programmed to do, it can fall into any one of these categories.
 
 
'''1.4.2 Culling Data'''
 
 
When dealing with large data, it is clearly best to cull out data whenever possible, and the earlier the better. Most large data starts as 3D geometry and the desired geometry is often a surface. As surfaces usually have a much smaller memory footprint than the volumes that they are derived from, it is best to convert to a surface soon. Once you do that, you can apply other filters in relative safety.
 
A very common visualization operation is to extract isosurfaces from a volume using the contour filter. The '''Contour''' filter usually outputs geometry much smaller than its input. Thus, the '''Contour''' filter should be applied early if it is to be used at all. Be careful when setting up the parameters to the '''Contour''' filter because it still is possible for it to generate a lot of data. This obviously can happen if you specify many isosurface values. High frequencies such as noise around an isosurface value can also cause a large, irregular surface to form.
 
Another way to peer inside of a volume is to perform a '''Slice''' on it. The '''Slice''' filter will intersect a volume with a plane and allow you to see the data in the volume where the plane intersects. If you know the relative location of an interesting feature in your large data set, slicing is a good way to view it.
 
If you have little a-priori knowledge of your data and would like to explore the data without paying the memory and processing time for the full data set, you can use the '''Extract Subset''' filter to subsample the data. The subsampled data can be dramatically smaller than the original data and should still be well load balanced. Of course, be aware that you may miss small features if the subsampling steps over them and that once you find a feature you should go back and visualize it with the full data set.
 
There are also several features that can pull out a subset of a volume: '''Clip''', '''Threshold''', '''Extract Selection''', and '''Extract Subset''' can all extract cells based on some criterion. Be aware, however, that the extracted cells are almost never well balanced; expect some processes to have no cells removed. Also, all of these filters with the exception of '''Extract Subset''' will convert structured data types to unstructured grids. Therefore, they should not be used unless the extracted cells are of at least an order of magnitude less than the source data.
 
When possible, replace the use of a filter that extracts 3D data with one that will extract 2D surfaces. For example, if you are interested in a plane through the data, use the '''Slice''' filter rather than the '''Clip''' filter. If you are interested in knowing the location of a region of cells containing a particular range of values, consider using the '''Contour''' filter to generate surfaces at the ends of the range rather than extract all of the cells with the '''Threshold''' filter. Be aware that substituting filters can have an effect on downstream filters. For example, running the '''Histogram''' filter after '''Threshold''' will have an entirely different effect then running it after the roughly equivalent '''Contour''' filter.

Latest revision as of 18:37, 14 January 2015

PAGE DELETED
The Users Guide has been moved from the Wiki to The ParaView Guide. Please use the history if you want to access the old version of this document.