Computer Physics Communications Program LibraryPrograms in Physics & Physical Chemistry |

[Licence| Download | New Version Template] aenz_v1_0.tar.gz(1466 Kbytes) | ||
---|---|---|

Manuscript Title: MCMC^{2}: a Monte Carlo code for Multiply-Charged Clusters | ||

Authors: David A. Bonhommeau, Marie-Pierre Gaigeot | ||

Program title: MCMC^{2} | ||

Catalogue identifier: AENZ_v1_0Distribution format: tar.gz | ||

Journal reference: Comput. Phys. Commun. 184(2013)873 | ||

Programming language: Fortran 90 with MPI extensions for parallelization. | ||

Computer: x86 and IBM platforms. | ||

Operating system: - CentOS 5.6 Intel Xeon X5670 2.93 GHz, gfortran + MPICH2.
- CentOS 5.3 Intel Xeon E5520 2.27 GHz, gfortran/g95/pgf90 + MPICH2.
- Red Hat Enterprise 5.3 Intel Xeon X5650 2.67 GHz, gfortran + IntelMPI.
- IBM Power 6 4.7 GHz, xlf + PESS (IBM parallel library).
| ||

Has the code been vectorised or parallelized?: Yes, parallelized using MPI extensions.
Number of CPUs used: 2 to ~ 40 but the code enables the use of 999 CPUs if desired. | ||

RAM: 10-20 Mb | ||

Keywords: Monte Carlo simulations, Coarse-grained models, Charged clusters, Charged droplets, Electrospray ionisation, Parallel tempering, Parallel charging. | ||

Classification: 23. | ||

Nature of problem:We provide a general parallel code to investigate structural and thermodynamic properties of multiply charged clusters. | ||

Solution method:Parallel Monte Carlo methods are implemented for the exploration of the configuration space of multiply charged clusters. Two parallel Monte Carlo methods were found appropriate to achieve such a goal: the Parallel Tempering method, where replicas of the same cluster at different temperatures are distributed among different CPUs, and Parallel Charging where replicas (at the same temperature) having different particle charges or numbers of charged particles are distributed on different CPUs. | ||

Restrictions:The current version of the code uses Lennard-Jones interactions, as the main cohesive interaction between spherical particles, and electrostatic interactions (charge-charge, charge-induced dipole, induced dipole-induced dipole, polarisation). The Monte Carlo simulations can only be performed in the NVT ensemble in the present code. | ||

Unusual features:The Parallel Charging methods, based on the same philosophy as Parallel Tempering but with particle charges and number of charged particles as parameters instead of temperature, is an interesting new approach to explore energy landscapes. Splitting of the simulations is allowed and averages are accordingly updated. | ||

Running time:The running time depends on the number of Monte Carlo steps, cluster size, and type of interactions selected (eg, polarisation turned on or off, and method used for calculating the induced dipoles). Typically a complete simulation can last from a few tens of minutes or a few hours for small clusters ( N ≤ 100, not including polarisation interactions), to one week for large clusters (N ≥ 1000 not including polarisation interactions), and several weeks for large clusters (N ≥ 1000) when including polarisation interactions. A restart procedure has been implemented that enables a splitting of the simulation accumulation phase. |

Disclaimer | ScienceDirect | CPC Journal | CPC | QUB |