Learning the ropes¶
In the tutorials, we will introduce the PySPH framework in the context of the examples provided. Read this if you are a casual user and want to use the framework as is. If you want to add new functions and capabilities to PySPH, you should read The PySPH framework. If you are new to PySPH however, we highly recommend that you go through this document and the next tutorial (A more detailed tutorial).
Recall that PySPH is a framework for parallel SPH-like simulations in Python. The idea therefore, is to provide a user friendly mechanism to set-up problems while leaving the internal details to the framework. All examples follow the following steps:
The tutorials address each of the steps in this flowchart for problems with increasing complexity.
The first example we consider is a “patch” test for SPH formulations for incompressible fluids in elliptical_drop_simple.py. This problem simulates the evolution of a 2D circular patch of fluid under the influence of an initial velocity field given by:
The kinematical constraint of incompressibility causes the initially circular patch of fluid to deform into an ellipse such that the volume (area) is conserved. An expression can be derived for this deformation which makes it an ideal test to verify codes.
Taking a look at the example (see elliptical_drop_simple.py), the first several lines are imports of various modules:
from numpy import ones_like, mgrid, sqrt from pysph.base.utils import get_particle_array from pysph.solver.application import Application from pysph.sph.scheme import WCSPHScheme
This is common for most examples and it is worth noting the pattern of the
PySPH imports. Fundamental SPH constructs like the kernel and particle
containers are imported from the
base subpackage. The framework
related objects like the solver and integrator are imported from the
solver subpackage. Finally, we import from the
sph subpackage, the
physics related part for this problem.
The organization of the
pysph package is given below.
Organization of the
PySPH is organized into several sub-packages. These are:
pysph.base: This subpackage defines the
pysph.base.particle_array.ParticleArray, the various SPH Kernels, the nearest neighbor particle search (NNPS) code, and the Cython code generation utilities.
pysph.sph: Contains the various SPH equations, the Integrator related modules and associated integration steppers, and the code generation for the SPH looping.
pysph.sph.wccontains the equations for the weakly compressible formulation.
pysph.sph.solid_mechcontains the equations for solid mechanics and
pysph.sph.mischas miscellaneous equations.
pysph.solver: Provides the
pysph.solver.application.Applicationand a convenient way to interact with the solver as it is running.
pysph.parallel: Provides the parallel functionality.
pysph.tools: Provides some useful tools including the
pysphscript CLI and also the data viewer which is based on Mayavi.
pysph.examples: Provides many standard SPH examples. These examples are meant to be extended by users where needed. This is extremely handy to reproduce and compare SPH schemes.
Functions for loading/generating the particles¶
The code begins with a few functions related to obtaining the exact solution for the given problem which is used for comparing the computed solution.
A single new class called
EllipticalDrop which derives from
pysph.solver.application.Application is defined. There are several
methods implemented on this class:
initialize: lets users specify any parameters of interest relevant to the simulation.
create_scheme: lets the user specify the
pysph.sph.scheme.Schemeto use to solve the problem. Several standard schemes are already available and can be readily used.
create_particles: this method is where one creates the particles to be simulated.
create_scheme are mandatory for without
them SPH would be impossible. The rest (and other methods) are optional. To
see a complete listing of possible methods that one can subclass see
create_particles method looks like:
class EllipticalDrop(Application): # ... def create_particles(self): """Create the circular patch of fluid.""" dx = self.dx hdx = self.hdx ro = self.ro name = 'fluid' x, y = mgrid[-1.05:1.05+1e-4:dx, -1.05:1.05+1e-4:dx] x = x.ravel() y = y.ravel() m = ones_like(x)*dx*dx h = ones_like(x)*hdx*dx rho = ones_like(x) * ro u = -100*x v = 100*y # remove particles outside the circle indices =  for i in range(len(x)): if sqrt(x[i]*x[i] + y[i]*y[i]) - 1 > 1e-10: indices.append(i) pa = get_particle_array(x=x, y=y, m=m, rho=rho, h=h, u=u, v=v, name=name) pa.remove_particles(indices) print("Elliptical drop :: %d particles" % (pa.get_number_of_particles())) self.scheme.setup_properties([pa]) return [pa]
The method is used to initialize the particles in Python. In PySPH, we use a
ParticleArray object as a container for particles of a given
species. You can think of a particle species as any homogenous entity in a
simulation. For example, in a two-phase air water flow, a species could be
used to represent each phase. A
ParticleArray can be conveniently
created from the command line using NumPy arrays. For example
>>> from pysph.base.utils import get_particle_array >>> x, y = numpy.mgrid[0:1:0.01, 0:1:0.01] >>> x = x.ravel(); y = y.ravel() >>> pa = sph.get_particle_array(x=x, y=y)
would create a
ParticleArray, representing a uniform distribution
of particles on a Cartesian lattice in 2D using the helper function
get_particle_array() in the base subpackage. The
get_particle_array_wcsph() is a special version of this suited to
ParticleArrays in PySPH use flattened or one-dimensional arrays.
ParticleArray is highly convenient, supporting methods for
insertions, deletions and concatenations. In the
function, we use this convenience to remove a list of particles that fall
outside a circular region:
where, a list of indices is provided. One could also provide the indices in the
form of a
cyarray.carray.LongArray which, as the name suggests, is
an array of 64 bit integers.
The particle array also supports what we call strided properties where you may associate multiple values per particle. Normally the stride length is 1. This feature is convenient if you wish to associate a matrix or vector of values per particle. You must still access the individual values as a “flattened” array but one can resize, remove, and add particles and the strided properties will be honored. For example:
>>> pa.add_property(name='A', data=2.0, default=-1.0, stride=2)
Will create a new property called
'A' with a stride length of 2.
Any one-dimensional (NumPy) array is valid input for PySPH. You can generate this from an external program for solid modelling and load it.
PySPH works with multiple ParticleArrays. This is why we actually return a list in the last line of the get_circular_patch function above.
create_particles always returns a list of particle arrays even if
there is only one. The method
adds any properties needed for the particular scheme being used.
Setting up the PySPH framework¶
As we move on, we encounter instantiations of the PySPH framework objects.
In this example, the
pysph.sph.scheme.WCSPH scheme is created in
create_scheme method. The
WCSPHScheme internally creates other
basic objects needed for the SPH simulation. In this case, the scheme
instance is passed a list of fluid particle array names and an empty list of
solid particle array names. In this case there are no solid boundaries. The
class is also passed a variety of values relevant to the scheme and
simulation. The kernel to be used is created and passed to the
configure_solver method of the scheme. The
pysph.sph.integrator.EPECIntegrator is used to integrate the
particle properties. Various solver related parametes are also setup.
def create_scheme(self): s = WCSPHScheme( ['fluid'], , dim=2, rho0=self.ro, c0=self.co, h0=self.dx*self.hdx, hdx=self.hdx, gamma=7.0, alpha=0.1, beta=0.0 ) dt = 5e-6 tf = 0.0076 s.configure_solver(dt=dt, tf=tf) return s
As can be seen, various options are configured for the solver, including initial damping etc. The scheme is responsible for:
- setting up the actual equations that describe the interactions between particles (see SPH equations),
- setting up the kernel (SPH Kernels) and integrator (Integrator related modules) to use for the simulation. In this case a default cubic spline kernel is used.
- setting up the Solver (Module solver), which marshalls the entire simulation.
For a more detailed introduction to these aspects of PySPH please read, the
A more detailed tutorial tutorial which provides greater detail on these.
However, by simply creating the
WCSPHScheme and creating the particles,
one can simulate the problem.
The astute reader may notice that the
EllipticalDrop example is
Application. This makes it easy to pass command
line arguments to the solver. It is also important for the seamless parallel
execution of the same example. To appreciate the role of the
Application consider for a moment how might we write a parallel
version of the same example. At some point, we would need some MPI imports and
the particles should be created in a distributed fashion. All this (and more)
is handled through the abstraction of the
Application which hides
all this detail from the user.
Running the example¶
In the last two lines of the example, we instantiate the
class and run it:
if __name__ == '__main__': app = EllipticalDrop() app.run()
Application takes care of creating the particles, creating the
solver, handling command line arguments etc. Many parameters can be
configured via the command line, and these will override any parameters setup
in the respective
create_* methods. For example one may do the following
to find out the various options:
$ pysph run elliptical_drop_simple -h
If we run the example without any arguments it will run until a final time of 0.0075 seconds. We can change this example to 0.005 by the following:
$ pysph run elliptical_drop_simple --tf=0.005
When this is run, PySPH will generate Cython code from the equations and
integrators that have been provided, compiles that code and runs the
simulation. This provides a great deal of convenience for the user without
sacrificing performance. The generated code is available in
~/.pysph/source. If the code/equations have not changed, then the code
will not be recompiled. This is all handled automatically without user
intervention. By default, output files will be generated in the directory
If we wish to utilize multiple cores we could do:
$ pysph run elliptical_drop_simple --openmp
$ mpirun -np 4 pysph run elliptical_drop_simple
This will automatically parallelize the run using 4 processors. In this example doing this will only slow it down as the number of particles is extremely small.
Visualizing and post-processing¶
You can view the data generated by the simulation (after the simulation
is complete or during the simulation) by running the
command. To view the simulated data you may do:
$ pysph view elliptical_drop_simple_output
If you have Mayavi installed this should show a UI that looks like:
For more help on the viewer, please run:
$ pysph view -h
On the user interface, the right side shows the visualized data. On top of it there are several toolbar icons. The left most is the Mayavi logo and clicking on it will present the full Mayavi user interface that can be used to configure any additional details of the visualization.
On the bottom left of the main visualization UI there is a button which has the text “Launch Python Shell”. If one clicks on this, one obtains a full Python interpreter with a few useful objects available. These are:
>>> dir() ['__builtins__', '__doc__', '__name__', 'interpolator', 'mlab', 'particle_arrays', 'scene', 'self', 'viewer'] >>> particle_arrays['fluid'].name 'fluid'
particle_arrays object is a dictionary of ParticleArrayHelpers
which is available in
interpolator is an instance of
pysph.tools.mayavi_viewer.InterpolatorView that is used by the
viewer. The other objects can be used to script the user interface if desired.
Note that the
particle_arrays can be indexed by array name or index.
Here is an example of scripting the viewer. Let us say we have two particle arrays, ‘boundary’ and ‘fluid’ in that order. Let us say, we wish to make the boundary translucent, then we can write the following:
b = particle_arrays['boundary'] b.plot.actor.property.opacity = 0.2
This does require some knowledge of Mayavi and scripting with it. The plot
attribute of the
a Glyph instance from Mayavi. It is useful to use the record feature
of Mayavi to learn more about how best to script the view.
The viewer will always look for a
mayavi_config.py script inside the
output directory to setup the visualization parameters. This file can be
created by overriding the
customize_output method. See the dam break 3d
example to see this being used. Of course, this file can also be created
Loading output data files¶
from pysph.solver.utils import load data = load('elliptical_drop_100.hdf5') # if one has only npz files the syntax is the same. data = load('elliptical_drop_100.npz')
When opening the saved file with
load, a dictionary object is returned.
The particle arrays and other information can be obtained from this
particle_arrays = data['arrays'] solver_data = data['solver_data']
particle_arrays is a dictionary of all the PySPH particle arrays.
You may obtain the PySPH particle array,
fluid, like so:
fluid = particle_arrays['fluid'] p = fluid.p
p is a numpy array containing the pressure values. All the saved particle
array properties can thus be obtained and used for any post-processing task.
solver_data provides information about the iteration count, timestep
and the current time.
A good example that demonstrates the use of these is available in the
post_process method of the
Data from the solver can also be interpolated using the
pysph.tools.interpolator.Interpolator class. Here is the simplest
example of interpolating data from the results of a simulation onto a fixed
grid that is automatically computed from the known particle arrays:
from pysph.solver.utils import load data = load('elliptical_drop_output/elliptical_drop_100.npz') from pysph.tools.interpolator import Interpolator parrays = data['arrays'] interp = Interpolator(list(parrays.values()), num_points=10000) p = interp.interpolate('p')
p is now a numpy array of size 10000 elements shaped such that it
interpolates all the data in the particle arrays loaded.
interp.y are numpy arrays of the chosen
p. To visualize this we may simply do:
from matplotlib import pyplot as plt plt.contourf(interp.x, interp.y, p)
It is easy to interpolate any other property too. If one wishes to explicitly set the domain on which the interpolation is required one may do:
xmin, xmax, ymin, ymax, zmin, zmax = 0., 1., -1., 1., 0, 1 interp.set_domain((xmin, xmax, ymin, ymax, zmin, zmax), (40, 50, 1)) p = interp.interpolate('p')
This will create a meshgrid in the specified region with the specified number of points.
One could also explicitly set the points on which one wishes to interpolate the data as:
interp.set_interpolation_points(x, y, z)
x, y, z are numpy arrays of the coordinates of the points on which
the interpolation is desired. This can also be done with the constructor as:
interp = Interpolator(list(parrays.values()), x=x, y=y, z=z)
There are some cases, where one may require a higher order interpolation or
gradient approximation of the property. This can be done by passing a
method for interpolation to the interplator as:
interp = Interpolator(list(parrays.values()), num_points=10000, method='order1')
Currently, PySPH has three method of interpolation namely
order1 is set as method then one can get the
higher order interpolation or it’s derivative by just passing an extra
argument to the interpolate method suggesting the component. To get
derivative in x we can do as:
px = interp.interpolate('p', comp=1)
Here for comp=0, the interpolated property is returned and 1, 2, 3 will return gradient in x, y and z directions respectively.
For more details on the class and the available methods, see
In addition to this there are other useful pre and post-processing utilities described in Miscellaneous Tools for PySPH.
Viewing the data in an IPython notebook¶
PySPH makes it relatively easy to view the data inside an IPython notebook with minimal additional dependencies. A simple UI is provided to view the saved data using this interface. It requires jupyter, ipywidgets and ipympl. Currently, a 2D and 3D viewer are provided for the data. Here is a simple example of how one may use this in a notebook. Inside a notebook, one needs the following:
%matplotlib ipympl from pysph.tools.ipy_viewer import Viewer2D viewer = Viewer2D('dam_break_2d_output')
viewer has many useful methods:
viewer.show_info() # prints useful information about the run. viewer.show_results() # plots any images in the output directory viewer.show_log() # Prints the log file.
The most handy one is the one to perform interactive plots:
This shows a simple ipywidgets based UI that uses matplotlib to plot the data
on the browser. The different saved snapshots can be viewed using a convenient
slider. The viewer shows both the particles as well as simple vector plots.
This is convenient when one wishes to share and show the data without
requiring Mayavi. It does require pysph to be installed in order to be able to
load the files. It is mandatory to have the first line that sets the matplotlib
There is also a 3D viewer which may be used using
Viewer3D instead of the
Viewer2D above. This viewer requires ipyvolume to be installed.
A slightly more complex example¶
The first example was very simple. In particular there was no post-processing of the results. Many pysph examples also include post processing code in the example. This makes it easy to reproduce results and also easily compare different schemes. A complete version of the elliptical drop example is available at elliptical_drop.py.
There are a few things that this example does a bit differently:
- It some useful code to generate the exact solution for comparison.
- It uses a
Gaussiankernel and also uses a variety of different options for the solver (see how the
configure_solveris called) for various other options see
EllipticalDropclass has a
post_processmethod which optionally post-process the results generated. This in turn uses a couple of private methods
- The last line of the code has a call to
app.post_process(...), which actually post-processes the data.
This example is therefore a complete example and shows how one could write a useful and re-usable PySPH example.
Application has several more methods that can be used in
additional contexts, for example one may override the following additional
add_user_options: this is used to create additional user-defined command line arguments. The command line options are available in
self.optionsand can be used in the other methods.
consume_user_options: this is called after the command line arguments are parsed, and can be optionally used to setup any variables that have been added by the user in
add_user_options. Note that the method is called before the particles and solver etc. are created.
create_domain: this is used when a periodic domain is needed.
create_inlet_outlet: Override this to return any inlet an outlet objects. See the
There are many others, please see the
documentation to see these. The order of invocation of the various methods is
also documented there.
There are several examples that ship with PySPH, explore these to get a better idea of what is possible.
Debugging when things go wrong¶
When you attempt to run your own simulations you may run into a variety of errors. Some errors in setting up equations and the like are easy to detect and PySPH will provide an error message that should usually be helpful. If this is a Python related error you should get a traceback and debug it as you would debug any Python program.
PySPH writes out a log file in the output directory, looking at that is sometimes useful. The log file will usually tell you the kernel, integrator, NNPS, and the exact equations and groups used for a simulation. This can be often be very useful when sorting out subtle issues with the equations and groups.
Things get harder to debug when you get a segmentation fault or your code just crashes. Even though PySPH is implemented in Python you can get one of these if your timestep is too large or your equations are doing strange things (divide by zero, taking a square root of a negative number). This happens because PySPH translates your code into a lower-level language for performance. The following are the most common causes of a crash/segfault:
- The particles have “blown up”, this can happen when the accelerations are very large. This can also happen when your timestep is very large.
- There are mistakes in your equations or integrator step. Divide by zero, or some quantity was not properly initialized – for example if the particle masses were not correctly initialized and were set to zero you might get these errors. It is also possible that you have made some indexing errors in your arrays, check all your array accesses in your equations.
Let us see how we can debug these. Let us say your code is in
you can do the following:
$ python example.py --pfreq 1 --detailed-output
In this case, the
--pfreq 1 asks pysph to dump output at every timestep.
By default only specific properties that the user has requested are saved.
--detailed-output dumps every property of every array. This includes
all accelerations as well. Viewing this data with the
pysph view command
makes it easy to see which acceleration is causing a problem.
Sometimes even this is not enough as the particles diverge or the code blows
up at the very first step of a multi-stage integrator. In this case, no output
would be generated. To debug the accelerations in this situation one may
define a method called
pre_step in your
pysph.solver.application.Application subclass as follows:
class EllipticalDrop(Application): # ... def pre_step(self, solver): solver.dump_output()
What this does is to ask the solver to dump the output right before each
timestep is taken. At the start of the simulations the first accelerations
have been calculated and since this output is now saved, one should be able to
debug the accelerations. Again, use
--detailed-output with this to look at
the accelerations right at the start.