Dear all,
this topic is something that is bugging my mind for some time now and I would like to get some feedback from the small community that this forum is. As this is just me ranting and asking for other people opinions, feel free to ignore if you’re not interested.
To give some context, I’ve been trained in molecular simulation (MS) “the old way” with a general statistical physics background. By “old way”, I mean I worked on the MD and MC code developed in-house by my labs during my PhD, both of them written in Fortran, just like (most of) our post-processing codes. We even plotted data using xmgrace, and I was the one who doomed some of our traditional workflows by introducing Python scripting in the lab. But I knew the Nosé-Hoover and MC equations nearly by heart and some of my favorite statistical physics books were the ones from Hoover.
As weird as it might sound, I discovered a lot of the advanced state-of-the-art software and techniques after completing my PhD, like Metadynamics, collective variables, advanced coarse-graining protocols etc, some of this nice work developed by the same people here on the forum. Modern machine-learning is baffling compared to what I used to do. Some of these methods rely on heavy thermodynamics or statistical physics analysis so I am not surprised I sometime have a hard time wrapping my head around it, some of them I still do not quite fully understand.
Now I have the feeling that I am becoming an old fart in the sense that there is a very wide variety of software around. I used VMD for a long time and switched to OVITO, I am now a proficient LAMMPS user (or so do I think), but I can’t help to wonder if the current state of those software is good for people to grasp MS. Don’t get me wrong, I love the work that is put in these, and I am helping here for a reason, but I sometime feel that even I rely too much on them and that this blunts my creativity with regard to analyses.
For people arriving in the field, in the same way that it is very hard to teach quantum dynamics without some classical mechanics basis (and even then, it is still hard to teach quantum mechanics), I feel as if some of the comfy software masks some of the required basic understanding of what MD is and does. I’ve been quite surprised by seeing people doing material science doing only “0K energy minimization” in MD and comparing computed values with real life material. In the same way, I saw some papers published with only nice “illustrative figures” from OVITO but not much more, as if quantitative analysis was not useful, necessary or even obtainable from simulations. It is also sometime given with the bare minimum, and other time computational information, like forcefield, are barely present at all. This makes me actually go “WTF?” sometimes. Would anyone find acceptable to publish an experimental paper with only an illustrative picture of there sample? I know that the publishing systems has quite some issues with peer review, greedy publishers and the astronomical number of papers published (most of them ending up being NOT read), but I can’t help but feel that there is also some issues in the way of doing/presenting MS.
I have a feeling that some people are leaved on there own into doing some MS without proper training, and just taking the software, sometimes barely reading the doc and “fiddle around” as best as they can. There is, as I said, a lot to know and master to do proper MS nowadays. So are there still dinosaurs teaching the basics to newcomers? Are they still relevant? Is there any way to hop in directly using new tools and technique without missing some relevant parts?
Now I know that there is, in science training, a lot of “learning by doing” and getting proper guidance from teachers you copy the workflow/thinking process of. Beside @akohlmey saying it (rightfully) a countless number of time, I agree with Michael Polanyi[1] on this topic. But I wonder about other people opinion here on the current state of MS understanding given the current state of software development, if my feeling is shared or if that’s not the case.
As a side note, I know that people posting on this forum represent a biased sample with a distribution having, I suppose, high density at the two ends of the “beginner to proficient user” spectrum. And this is why I am suspending my own judgment on this question.
Who just said essentially the same thing but in BIG BOOKS with a lot of pages and footnotes. ↩︎