Gaussian is an electronic structure modeling program that Yale has licensed for its HPC clusters. To see available versions of Guassian on the cluster, run:
module avail gaussian
Running Gaussian on the Cluster
When running Gaussian, it is recommended that users request exclusive access to allocated nodes (e.g., by requesting 8 cpus per node) and that they specify the largest possible memory allocation for the number of nodes requested. In addition, in most cases, the scratch storage location (set by the environment variable
GAUSS_SCRDIR) should be on the local parallel file system of the cluster, rather than in the user’s home directory.
Before running Gaussian, you must set up a number of environment variables. This is accomplished most easily by loading the Gaussian module file using:
module load Apps/Gaussian
To run Gaussian interactively, you need to create an interactive session on a compute node. You could start an interactive session using 4 cores for 4 hours using
srun --pty -c 4 -p interactive -t 4:00:00 bash
See our Slurm documentation for more detailed information on requesting resources for interactive jobs.
Gaussian in parallel
To run Gaussian in parallel, you may find it simplest to use the script
par_g09 instead of executing Gaussian directly.
par_g09 takes 1 optional argument and is used as shown here:
par_g09 [num_proc_shared] < g09_input > g09_output
When included in a submission script, par_g09 will set the following Gaussian input variables:
%LindaWorkers: List of nodes that will act as parallel workers during parallel portions of the Gaussian computations. The list is created automatically based on the nodes allocated to the particular job. Workers communicate using the Network Linda system provided with the Gaussian software.
%NProcShared: Number of shared-memory processes per Linda worker.
It is highly recommended that users run 1 Linda worker per node with
%NProcShared=8 on the Omega cluster. (This is the default behavior of
par_g09.) However, users may easily override the value of
%NProcShared by supplying the optional num_proc_shared argument to
par_g09. (Of course, both variables may be overridden by including them in the Gaussian input file.)
On the Grace cluster, there are up to 20 processors and 128 GB of memory per node, so it may often work well to use just a single node with many shared-memory processors, in which case you can simply modify the Gaussian input file and skip the use of the
par_g09 script. If you do wish to use multiple nodes on the Grace cluster, please pay careful attention to the Slurm parameters you use, particular to the setting of
--mem-per-cpu to ensure that each of the nodes you request has sufficient resources available.
In connection with Gaussian, we have also installed GaussView 5, Gaussian Inc.'s advanced and powerful graphical interface for Gaussian. With GaussView, you can import or build the molecular structures that interest you; set up, launch, monitor and control Gaussian calculations; and retrieve and view the results, all without ever leaving the application. GaussView 5 includes many new features designed to make working with large systems of chemical interest convenient and straightforward. It also provides full support for all of the new modeling methods and features in Gaussian 09.
In order to use GaussView, you must run an X Server on your desktop or laptop, and you must enable X forwarding when logging into the cluster. See our X11 forwarding documentation for instructions.
Loading the module file for Gaussian sets up your environment for GaussView as well. Then you can type the command
gv. GaussView 5 may not be compatible with certain versions of the X servers you may run on your desktop or laptop. If you encounter problems, these can often be overcome by starting GaussView with the command
gv -mesagl or