Welcome

Welcome to our website. Our group, based in the Department of Chemistry and Materials Science at Aalto University, Finland is led as Principal Investigator by Dr. Miguel Caro. We carry out academic research on computer simulation of materials using atomistic models. Browse the website to find out more about our work.

The group at the PsiK Conference in Lausanne in August 2022 (Max and Yanzhou were at the conference but are missing from the picture). See older pictures here.
Announcements and latest news
Understanding the structure of amorphous carbon and silicon with machine learning atomistic modeling
Understanding the structure of amorphous carbon and silicon with machine learning atomistic modelingMiguel CaroMarch 6, 2023When we think about semiconductors the first ones to come to mind are silicon, germanium and the III-Vs (GaAs, InP, AlN, GaN, etc.), in their crystalline forms. In fact, the degree of crystallinity in these materials often dictates the quality of the devices that can be made with them. As an example, dislocation densities as low as 1000/cm2 can negatively affect the performance of GaAs LEDs. For reference, within a perpendicular section of (111)-oriented bulk GaAs, that’s a bit less than one in a trillion (1012) “bad” primitive unit cells. Typical overall defect densities (e.g., including point defects) are much higher, but still the number of “good” atoms is thousands of times larger than the number of “bad” atoms. With this in mind, one might be boggled to find out amorphous semiconductors can have useful properties of their own, despite being the equivalent of a continuous network of crystallographic defects. On the way to understand the properties of an amorphous material, our first pit stop is understanding its atomic-scale structure. For this purpose, computational atomistic modeling tools are particularly useful, since many of the techniques commonly used to study the atomistic structure of crystals are not applicable to amorphous materials, precisely because they rely on the periodicity of the crystal lattice. Unfortunately, one of the most used computational approaches for materials modeling, density functional theory (DFT), is computationally too expensive to study the sheer structural complexity in real amorphous materials, in particular long-range structure for which many thousands or even millions of atoms need to be considered. The introduction and popularization in recent years of machine learning (ML) based atomistic modeling and, in particular, ML potentials (MLPs), has enabled for the first time realistic studies of amorphous semiconductors with accuracy close to that of DFT. As two of the most important such materials, amorphous silicon (a-Si) and amorphous carbon (a-C) have been the target of much of these early efforts. In a recent Topical Review in Semiconductor Science and Technology (provided Open Access via this link), I have tried to summarize our attempts to understand the structure of a-C and a-Si and highlight how MLPs allow us to peek at the structure of these materials and draw the connection between this structure and the emerging properties. The discussion is accompanied by a general description of atomistic modeling of a-C and a-Si and a brief introduction to MLPs, and it could be interesting to the material scientist curious about the modeling of amorphous materials or the DFT practicioner curious about what MLPs can do that DFT can’t. At this stage, the field is still evolving (fast!) and I expect(/hope) this review will become obsolete soon, as more accurate and CPU-efficient atomistic ML techniques become commonplace. Especially, I expect the description of a-Si and a-C to rapidly evolve from the study of pure samples to the more realistic (and chemically complex) materials, containing unintentional defects as well as chemical functionalization. All in all, I am excited to witness in which direction the field will steer in the next 5-10 years. I promise to do my part and wait at least that much before writing the next review on the topic! [...] Read more...
Automated X-ray photoelectron spectroscopy (XPS) prediction for carbon-based materials: combining DFT, GW and machine learning
Automated X-ray photoelectron spectroscopy (XPS) prediction for carbon-based materials: combining DFT, GW and machine learningMiguel CaroJuly 13, 2022The details of this work are now published (open access ) in Chem. Mater. and our automated prediction tool is available from nanocarbon.fi/xps. Many popular experimental methods for determining the structure of materials rely on the periodic repetition of atomic arrangements present in crystals. A common example is X-ray diffraction. For amorphous materials, the lack of periodicity renders these methods impractical. Core-level spectroscopies, on the other hand, can give information about the distribution of atomic motifs in a material without the requirement of periodicity. For carbon-based materials, X-ray photoelectron spectroscopy (XPS) is arguably the most popular of these techniques. In XPS a core electron is excited via absorption of incident light and ejected out of the sample. Core electrons occupy deep levels close to the atomic nuclei, for instance 1s states in oxygen and carbon atoms. They do not participate in chemical bonding since they are strongly localized around the nuclei and lie far deeper in energy than valence electrons. Because core electrons lie so deep in the potential energy well, energetic X-ray light is required to eject them out of the sample. In XPS, the light source is monochromatic, which means that all the X-ray photons have around the same energy, $h \nu$. When a core electron in the material absorbs one of these photons with enough energy to leave the sample, we can measure its kinetic energy and work out its binding energy (BE) as $\text{BE} = E_\text{kin} – h \nu$. After collecting many of these individual measurements, a spectrum of BEs will appear, because each core electron has a BE that depends on its particular atomic environment. For instance, a core electron from a C atom bonded only to other C atoms has a lower BE than a C atom that is bonded also to one or more O atoms. And even the details of the bonding matter: a core electron from a “diamond-like” C atom (a C atom surrounded by four C neighbors) has a higher BE than that coming from a “graphite-like” C atom (which has only three neighbors). Therefore, the different features, or “peaks” in the spectrum can be traced back to the atomic environments from which core electrons are being excited, giving information about the atomic structure of the material. This is illustrated in the featured image of this post (the one at the top of this page). What makes XPS so attractive for computational materials physicists and chemists like us is that it provides a direct link between simulation and experiment that can be exploited to 1) validate computer-generated model structures and 2) try to work out the detailed atomic structure of an experimental sample. 1) is more or less obvious; 2) is motivated by the fact that experimental analysis of XPS spectra is usually not straightforward because features that come from different atomic environments can overlap on the spectrum (i.e., they coincidentally occur at the same energy). In both cases, computational XPS prediction requires two things. First, a computer-generated atomic structure. Second, an electronic-structure method to compute core-electron binding energies. While candidate structural models can be made with a variety of tools (a favorite of ours is molecular dynamics in combination with machine-learning force fields), a standing issue with computational XPS prediction is the accuracy of the core-electron BE calculation. Even BE calculations based on density-functional theory, the workhorse of modern ab initio materials modeling, lack satisfactory accuracy. A few years ago my colleague Dorothea Golze, with whom I used to share an office during our postdoc years at Aalto University, started to develop highly accurate techniques for core-electron BE determination based on a Green’s function approach, commonly referred to as the GW method. These GW calculations can yield core-electron BEs at unprecedented accuracy, albeit at great computational cost. In particular, applying this method to atomic systems with more than a hundred atoms is impractical due to CPU costs. This is where machine learning (ML) can come in handy. Four years ago, around the time when Dorothea’s code was becoming “production ready”, I was just getting started in the world of ML potentials from kernel regression, using the Gaussian approximation potential (GAP) method developed by my colleagues Gábor Csányi and Albert Bartók ten years before. I also had prior experience doing XPS calculations based on DFT for amorphous carbon (a-C) systems. Then the connection was clear. GAPs work by constructing the total (potential) energy of the system as the optimal collection of individual energies. This approximation is not based on physics (local atomic energies are not a physical observable) but is necessary to keep GAPs computationally tractable. However, the core-electron BE is a local physical observable, and thus idally suited to be learned using the same mathematical tools that make GAP work. In essence, the BE is expressed as a combination of mathematical functions which feed on the details of the atomic environment of the atom in question (i.e., how the nearby atoms are arranged). Coincidentally, it was also around that time that our national supercomputing center, CSC, was deploying the first of their current generation of supercomputers, Puhti. They opened a call for Grand Challenge proposals for computational problems that require formidable CPU resources. It immediately occurred to me that Dorothea and I should apply for this to generate high-quality GW data to train an ML model of XPS for carbon-based materials. Fortunately, we got the grant, worth around 12.5M CPUh, and set out to make automated XPS a reality. But even formidable CPU resources are not enough to satisfy the ever-hungry GW method, so we had to be clever about how to construct an ML model which required as few GW data points as possible. We did this in two complementary ways. One the one hand, we used data-clustering techniques that we had previously employed to classify atomic motifs in a-C to select the most important (or characteristic) atomic environments out of a large database of computer-generated structures containing C, H and O (“CHO” materials). On the other hand, we came up with an ML model architecture which combined DFT and GW data. This is handy because DFT data is comparatively cheap to generate (it’s not cheap in absolute terms!) and we can learn the difference between a GW and a DFT calculation with a lot less data than we need to learn the GW predictions directly. So we can construct a baseline ML model using abundant DFT data and refine this model with scarce and precious GW data. And it works! An overview of the database of CHO materials used to identify the optimal training points (triangles). Four years and many millions of CPUh later, our freely available XPS Prediction Server is capable of producing an XPS spectrum within seconds for any CHO material, whenever the user can provide a computer-generated atomic structure in any of the common formats. Even better, these predicted spectra are remarkably close to those obtained experimentally. This opens the door for more systematic and reliable validation of computer-generated models of materials and a better integration of experimental and computational materials characterization. We hope that these tools will become useful to others and to extend them to other material classes and spectroscopies in the near future. [...] Read more...
Academy of Finland EuroHPC funding awarded to the group
Academy of Finland EuroHPC funding awarded to the groupMiguel CaroFebruary 5, 2022We have been awarded Academy of Finland funding within the framework of the EuroHPC research program, which aims to support the transition to (pre-)exascale high-performance computing (HPC) platforms and quantum computing, among others. Miguel Caro will lead the ExaFF (“Exascale-ready machine learning force fields“) consortium as Principal Investigator and Consortium Coordinator. This project is a collaboration with Andrea Sand’s group at Aalto University and CSC. The Academy of Finland has granted us 1,011,240 EUR for this project, out of which 435,108 EUR correspond to the Caro group. The objective of the project is to improve the accuracy and speed of Gaussian approximation potentials through the development of the TurboGAP code and the underlying algorithms. One of the priorities is to adapt them to efficient GPU execution, with a focus on exploiting the computing power provided by the CSC-hosted LUMI pre-exascale HPC system. The faster code and new functionality will be used to model new materials for battery applications and semiconductors under heavy radiation environments (e.g., circuit components aboard satellites). This is great news for our group, since it secures the resources necessary to continue the development of the TurboGAP code in the short term (the funding runs for 3 years, from 2022 until 2024). See the funding decision on the Academy of Finland’s website. [...] Read more...
Structure and properties of nanoporous carbons
Structure and properties of nanoporous carbonsMiguel CaroJanuary 4, 2022Nanoporous carbons are an emerging class of materials with important applications in energy storage. In particular, the ability of graphitic carbons to intercalate ions is exploited in commercial Li-ion batteries where the anode is (typically) made of graphite and Li ions become electrostatically bound to the carbon host as the battery is charged: electrons are stored in the graphitic matrix as part of an electron-ion pair. When the battery discharges, these electrons are injected into the external circuit (providing power) and the Li ion is released into the electrolyte traveling to the cathode, usually made of a transition-metal oxide, where it also intercalates. This cycle repeats itself as the battery is charged and discharged, with the Li ion traveling back and forth between the anode and cathode. Li is relatively scarce on the Earth’s crust and the mid-to-long term supply of Li needed to cover the rapidly increasing demand for Li-ion batteries is in jeopardy. The Li-intercalation process could in principle also be applied to more Earth-abundant ions, like K and Na, thus providing a reduction in the cost of ion batteries and ensuring future supply of raw materials. These are required to scale up the use of ion batteries and to make it affordarble for domestic and industrial applications. Unfortunately, Na and K do not intercalate in graphite as favorably as Li does, with Na-intercalated graphite deemed thermodynamically unstable, and in all cases incurring strong dimensional changes between charge and discharge. These dimensional changes pose risks to the mechanical stability of the material and the device containing it, with the associated safety concerns. Nanoporous carbons are an obvious alternative to graphite for ion intercalation because the pores, interstitial voids between disordered graphitic planes, can be made within a range of sizes, all larger than the usual interplanar spacing in graphite. Thus, nanoporous carbons can in principle accommodate larger ions, including Na and K, which motivates their study and structural characterization. However, as is often the case with amorphous and disordered materials, experimental characterization can be challenging, since common characterization techniques employed to characterize crystals cannot be used. In the case of nanoporous carbons, experimental characterization of pore sizes and shapes is very complicated. Nanoporous carbons are made of interlinked graphitic planes, where sp3 motifs provide the material with three-dimensional rigidity not present in graphite. Reprinted from Wang et al. Chem. Mater. (2022). With this background in mind, we decided to study the microscopic structure and mechanical properties of nanoporous carbon using state-of-the-art atomistic modeling techniques based on machine learning interatomic potentials. These techniques provide, for the first time, the required combination of accuracy and computational efficiency to study nanoporous carbons where the size of the simulation box does not constrain the size of the pores that can be studied. Our results are now published in Chemistry of Materials: Y. Wang, Z. Fan, P. Qian, T. Ala-Nissila, M.A. Caro; Structure and Pore Size Distribution in Nanoporous Carbon. Chem. Mater. (2022). Link to journal’s website. Open Access  PDF from the publisher. We started out by training a Gaussian approximation potential (GAP) for carbon based on the database developed by Deringer and Csányi [Phys. Rev. B 95, 094203 (2017)]. This new potential [10.5281/zenodo.5243184] achieves better accuracy and speed than the earlier version and can accurately predict the defect formation energies in graphitic carbon. Different accuracy tests on the new a-C GAP. Reprinted from Wang et al. Chem. Mater. (2022). Getting the relative formation energies right is critical for obtaining the correct topology of the complicated network of carbon rings within curved graphitic sheets. In particular, the relative abundance of 5-rings and 7-rings will determine the curvature and thus pore morphology in the material. With this new potential, we carried out large-scale simulations of graphitization with the TurboGAP code developed in our group, using a melt-graphitize-anneal protocol, akin to that by de Tomas et al. [Carbon 109, 681 (2016)], but now with larger systems (more than 130,000 atoms) and the accuracy provided by the new GAP. Generation protocol for nanoporous carbon. Reprinted from Wang et al. Chem. Mater. (2022). With these simulations we managed to generate realistic nanoporous carbon structures within a wide range of mass densities (0.5 to 1.7 g/cm3), and characterized in detail their short-, medium- and long-range order. For instance, these simulations reveal hexagonal motifs to be the dominant structural block in these materials (as expected) followed by 5-rings, then 7-rings and, in much smaller quantities, larger and smaller ring structures, with almost no density dependence for the most common motifs. Ring-size distribution. Reprinted from Wang et al. Chem. Mater. (2022). The pore sizes, the main target of this study, show clearly defined unimodal distributions determined by the overall mass density of the material. This means that the pore sizes and morphologies are relatively homogeneous for a given sample. Pore-size distribution. Reprinted from Wang et al. Chem. Mater. (2022). Finally, a useful result of our study is a library of nanoporous carbon structures freely available to the community and amenable to future studies on the properties of this interesting and important class of carbon materials. This study would have not been possible without the hard work and dedication of our PhD student Yanzhou Wang and the help of the other coauthors, as well as the support provided by the Academy of Finland and the CPU time and other computational resources provided by CSC and Aalto University’s Science IT project. [...] Read more...
Adding Van der Waals corrections to GAP force fields
Adding Van der Waals corrections to GAP force fieldsMiguel CaroAugust 7, 2021This post is about the following paper (and the figures are reproduced from it and are copyright of the American Physical Society): “Machine learning force fields based on local parametrization of dispersion interactions: Application to the phase diagram of C60“, by Heikki Muhli, Xi Chen, Albert P. Bartók, Patricia Hernández-León, Gábor Csányi, Tapio Ala-Nissila, and Miguel A. Caro. Phys. Rev. B 104, 054106 (2021). A preprint is openly available from the arXiv. Adding dispersion (or van der Waals ) corrections to density functional theory (DFT) has been a very active area of research in the past 10-15 years. DFT is a mean field theory and dynamical effects, such as the effects of fluctuating charge distributions on energy and forces, are notoriously missing. The leading term in these vdW corrections is the London dispersion force, which decays as the sixth power of the interatomic distance (Vij ∝ C6 / rij6). While this power law indicates that these interactions quickly become very small with distance for any given two atoms, they are usually attractive. This means that the collective effect of vdW interactions can lead to interesting emerging physical phenomena. An important example of this is how individual graphene layers interact with one another to form graphite. The first vdW correction scheme to achieve widespread adoption was the D2 method, due to Stefan Grimme (Grimme, J. Comp. Chem. 27, 1787 ). In this method, the effective C6 coefficient that parametrizes the London dispersion formula is given as a function of the atomic species and a damping function was introduced to switch off the dispersion correction at short interatomic distances. Soon, more sophisticated approaches arose that improved the accuracy of vdW corrections, by incorporating both 1) more information about the dependence of the effective C6 coefficient on the local atomic environment and 2) improved damping functions that bridge the transition between the long-range (London-type) correlation energy and the short-range DFT correlation energy. The Tkatchenko-Scheffler correction scheme (Tkatchenko and Scheffler, Phys. Rev. Lett. 102, 073005 ) is one such approach, which relies on an “atom in a molecule” approximation and relates the effective C6 coefficients and damping length scales to the Hirshfeld partitioning of the charge density. In parallel to these efforts on dispersion-correction schemes, the community has made huge advances in the past 10 years on machine learning interatomic interactions, usually from DFT data. These machine learning (ML) force fields are normally based on a local (atom-wise) decomposition of the total energy, to keep the simulation problem tractable. This is fine for strong covalent and repulsion interactions, which are typically short ranged. For instance, in carbon materials the covalent (“bonded”) part of the total energy of an atomistic system can be accurately learned with local atomic descriptors that are “blind” beyond 4-5 Angstrom. But vdW interactions are long ranged and, depending on the level of detail one aims at capturing, the “local” atomic environment relevant to vdW interactions is of the order of 15-20 Angstrom. This is bad news, because the computational cost of ML force fields scales as the cube of this distance, known as “cutoff”. This means that a ML force field with a 20 Angstrom cutoff is approximated 64 times as expensive, computationally, as another ML force field with a 5 Angstrom cutoff. Less obvious, but equally severe, limitations of “brute forcing” the learning of long-range interactions include the explosion of the size of configuration space with the cutoff, which requires exponentially more data for training these models. Making ML potentials two orders of magnitude slower is an unacceptable tradeoff for including vdW corrections. The approach we, and also others before us, took in our recent paper is to machine learn the Hirshfeld volumes, which are a function of the local atomic environment with locality similar to the covalent part of the total energy. Then, these Hirshfeld volumes are used to parametrize the Tkatchenko-Scheffler implementation of the London dispersion equation, which is computationally cheap to evaluate. In addition, we took care of efficiently coupling the covalent and vdW parts of the calculation, such that the overhead due to the Hirshfeld volume computation is very small, and the overall vdW-enabled force field is, for typical calculations, only between 20% and 50% more expensive than the force field without vdW corrections. A happy consequence of our ML implementation is that forces can be computed more accurately than the reference method, because the gradients of the Hirshfeld volumes can be computed generally with ML, whereas the DFT implementation cannot easily compute these terms (which are typically missing from DFT calculations). We implemented these corrections in the GAP and TurboGAP codes. As a proof of concept, we trained a new GAP for carbon and fine tuned it specifically to simulate C60, a carbon material entirely made of C60 molecules. We could simulate large systems with thousands of atoms for relatively long MD times, and could chart the metastable high temperature/high pressure phase diagram of C60, observing the transformations to graphitic carbon and amorphous carbon taking place in this material under specific thermodynamic conditions (see feature image at the top of this post). Database used to train the new C60 GAP force field This work is the culmination of a long effort by our group’s PhD student Heikki Muhli, that started more than 2 years ago during his MSc thesis project and continued as part of our COMPEX project, funded by the Academy of Finland, with CPU time provided by CSC and Aalto University’s Science IT project. It is also part of our continuous effort to develop and improve the TurboGAP code, which we hope to be able to launch within the next few months (the code’s development version is available online for the brave). For this work we collaborated with our group’s other PhD student Patricia Hernández-León, Aalto Univerty’s Xi Chen and Tapio Ala-Nissila, and GAP authors Albert Bartók (Warwick) and Gábor Csányi (Cambridge). And this is just the beginning. Newer vdW correction schemes worry about the role of “many-body” physics on vdW corrections, where the leading effect is how many-body interactions affect the effective C6 coefficients (Otero-de-la-Roza, LeBlanc and Johnson, Phys. Chem. Chem. Phys. 22, 8266 ). One of these many-body dispersion methods is the so-called “many-body dispersion” method (MBD; early bird gets the worm), which also feeds on Hirshfeld charge density partitioning (Tkatchenko, DiStasio Jr., Car and Scheffler. Phys. Rev. Lett. 108, 236402 ). With our new infrastructure to couple GAPs to simpler, locally parametrized, force fields, we will be able in the near future to incorporate many-body dispersion effects, as well as other long-range interactions, such as electrostatics. [...] Read more...