I think it looks correct. I get the same files with differences. I think most are due to differences in the eigensolver and natual frequencies. See discussion on that topic here:
What kind of google cloud machines are you going to run on? To see speedup you will need to use two dedicated machines with all the available cpus on each .
As for the google cloud machines, I have yet to decide what exactly I need. So far I just wanted to try your install procedure, so I set up an instance group where the number of instances could change from 1 to 10. The instances were E2 machines, the boot disk was 50 GB (oneAPI Base required at least 23 GB). At the moment the execution time is not the priority, but I need a lot of memory. My desktop at home has only 64 GB of memory, and it is not enough for my problem.
So I have run a buckling job in a google cloud cluster at last: one instance, custom, N2 machines, 8 vCPU, 120 GB RAM, 96 GB boot disk, at least 95 GB RAM was used, and it took about 100 minutes.
That is good. But if just running on one instance then probably no need to run the mpi version. The regular executable with pardiso solver should work just as well.
Both the MPI and non-MPI version can use all the processors. The only difference is the MPI version can use multiple hosts also. So if you install option 2 below then you are limited to all the processors on one host.
(1) Spooles ( Not recommended. 2-3X slower than Pardiso and can not solve models with more than a million degrees of freedom )
(2) Pardiso ( Must have the Intel compiler. If not, it is available for free from Intel.com. Does not require administrative privelages )
(3) Pardiso MPI ( Same requirements as above, but needs HPC kit also. Only works with v2.18 . Will update it for newer versions if there is interest. )
I might note I have run problems requiring more memory than I have by increasing my available page file size and using an SSD for the page file. I have a very fast SSD, but watching I/O bandwidth has indicated that is probably not necessary. Most of the FEM problems only access modest portions of the memory at one time or at nearly the same time. For my 64 GB machine working virtual memory set sizes up to 160% of physical memory have been usable. Larger starts to have a lot of paging slow things down a lot. Pastix does not use memory as efficiently as Pardiso, so problems don’t work as well when they get larger than physical memory, and they reach that limit sooner. I have not been using 18 yet.
I updated the script so 2.22 should now install for the single multicore machine. But the MPI version still will only work with 2.18. It will take some work to update that to 2.22. First I would need to see the use-case, i.e what features do you need in 2.22 that are not in 2.18? And what speedup are you getting on your cluster.
I don’t have a usable cluster currently. Just a bunch of old workstations. i am currently making the budget for next year and would like to invest a bit in hardware.
For the use case. I currently did make a try to get a workflow with openradioss and cubit and want to rebuild the example in calculix. The Johnson cook material is only available with 2.21.
The more important thing is i need a bit more power for running fsi problems using openfoam, calculix and precice.