So you did the experiments because the dead code was still in the source.
Where and how did this benefit and where didn't it?
On Fri, May 14, 2021 at 7:55 AM David Cerutti <dscerutti.gmail.com> wrote:
> The slide attached shows how one could draw lines down the long axis of the
> pencil prisms I've been advocating to cut the unit cell into and make eight
> voronoi volumes into which to place the particles. Four layers of such
> subdivisions will create 32 subregions in a hexagonal prism which is at
> least (1/2)(cutoff + buffer) thick in all directions, dividing the space
> into at least 4.6x the density of accept/reject mask subregions as the
> current pmemd for a cubic cell, or at least 5.7x the density that the
> current pmemd achieves in an octahedral cell. You probably don't need the
> full 8x density that you would get with another cubic partitioning at half
> the cell width. I would expect 32-bit masks to be more palatable for the
> GPU registers as well.
>
> Gustavo, I'm not sure it is possible to inquire about the accuracy without
> being drawn into a discussion about incentives. I expect that Desmond-GPU
> (which has to be considered separately from Desmond-Anton) produces stable
> simulations that give accurate results in the context of the force field.
>
> Dave
>
>
> On Fri, May 14, 2021 at 10:17 AM Gustavo Seabra <gustavo.seabra.gmail.com>
> wrote:
>
> > Thanks Dave for posting the link. I've been following the discussions
> > trying to figure out what they were talking about ;-)
> >
> > I find the discussion very interesting though, and I see the benchmarks
> are
> > really impressive, in terms of speed, and I do now want to enter into
> > motive considerations here.
> >
> > However, I do have one question that does not seem to be addressed in the
> > benchmarks. How precise are the calculations? With such speed, should it
> > not be possible to calculate a number of physical quantities and compare
> to
> > experimental values? IMHO, that is the only way to assess if the "tricks"
> > used are really good science. Do they break anything, or lead to the
> wrong
> > behavior of the system?
> >
> > Ultimately, I'd rather wait longer to get correct results. But of course,
> > that is also valid for Amber benchmarks.
> >
> > All the best,
> > --
> > Gustavo Seabra.
> >
> >
> > On Thu, May 13, 2021 at 6:03 PM David Cerutti <dscerutti.gmail.com>
> wrote:
> >
> > > For reference, here are the benchmarks that I think people are talking
> > > about:
> > > Desmond-GPU Performance as of April 2021.pdf (deshawresearch.com)
> > > <
> > >
> >
> https://www.deshawresearch.com/publications/Desmond-GPU%20Performance%20as%20of%20April%202021.pdf
> > > >
> > >
> > > Desmond uses a different long-ranged summation, the "U-series" which
> was
> > a
> > > bit of a dance of the seven veils and then turned out to be very
> similar
> > to
> > > other P3M techniques, SPME included. The U-series was the way they go
> to
> > > 8fs between updates to the long-ranged component of the electrostatics.
> > > Regardless of what it is, though, I'll say that my own experiences in
> > > multiple time stepping (see mdgx GB) don't leave much room to go higher
> > > than 5fs in any component of the force. Long ago, circa 2015 their
> DHFR
> > > benchmark was much faster than Amber (the 50% Scott is alluding to),
> > which
> > > seems to be a ratio they have maintained over the years, but it's now
> > more
> > > in line with the rest of the benchmarks--one can compute the number of
> > > atoms moved by the code in a given time and see that the STMV case is,
> > > indeed, moving substantially more than DHFR. It's pretty impressive
> that
> > > they can do ten million atoms, but of course that's more of a stunt (I
> > > would have been more inclined to do eight virions in a cube). That
> said,
> > > the Desmond folks do some pretty wild orchestration of how many fmafs
> and
> > > other arithmetic ops they can pull out of each cycle, so while their
> > > numbers may be tweaked according to any given standard my feeling is
> that
> > > "sales" are not a big incentive for them to cook the books.
> > >
> > > You can surely get more performance out of pmemd on the smaller systems
> > if
> > > you multiply the systems it simulates at one time. 2300ns per day with
> > > DHFR on one of the top-end Ampere cards shouldn't be out of the
> question.
> > > This should be one of the highest priorities in any renovations to the
> > > engine, as most pharma outfits study problems of 25-50k atoms, must run
> > > many windows before getting a single answer, and always have more
> > compounds
> > > to test than GPUs to do it. What I would also suggest is that anything
> > > happening to pmemd's CUDA component is stuck behind some very old
> Fortran
> > > code, with Pieces of a System flying around in a manner that's almost
> as
> > > depressing as the film with Vanessa Kirby. Rebuild the 100k lines of
> > > Fortran in C++ with accessible, well-engineered structs that are hard
> to
> > > break. Topologies, coordinates, and simulation protocols can all be
> > > structs passed around and created or destroyed as needed by a protocol.
> > > Give them each pointer structs that can be copied to the GPU in a
> manner
> > > analogous to cSim today, or preferably as multiple, focused pointer
> > structs
> > > that become kernel arguments when the actual kernel is launched (the
> > > long-ranged electrostatic kernel doesn't need to know about the bonded
> > > parameter constants, for example--a Prmtop struct can have multiple
> > pointer
> > > substructures tailored for different parts of the force calculation).
> > Make
> > > the kernels for producing work units operate on arrays of such structs,
> > so
> > > that a force kernel will seamlessly stride from one system to the next
> as
> > > it plays its part in any given time step. You should const as much as
> > > possible but const auto may be something to use sparingly, so that new
> > > developers will become better immersed in the actual nuts and bolts of
> > the
> > > code by seeing the actual data types. That will give upcoming
> > > graduate students more to work with and help them to understand the
> CUDA
> > > code as something much more C / C++ -like.
> > >
> > > Don't gnash your teeth over what DE Shaw's guys have achieved. The
> > things
> > > that drive sales are utility and unique capabilities, two things that
> > Amber
> > > has done pretty well with despite being the product of a handful of
> > > research groups who mostly prefer to see everyone staying in their
> > > respective lanes. Standardize what a "topology" is and make a clean,
> > > efficient, extensible tool for creating systems. That should be the
> > first
> > > stop for anyone thinking of adding new components to the force field
> or a
> > > free energy protocol. Document the hell out of everything. Stop
> relying
> > > on one Bob, or Scott, or me, or Taisung, or Scott again to
> > > MakeABetterEngine.cu. That needs to be a community activity, and it
> will
> > > improve the employment prospects of your students to have them involved
> > in
> > > professional python / C++ / CUDA programming. Be honest about your
> > > benchmarks and make a new section of the website as an exposition of
> > > Amber's free energy capabilities. It shouldn't take five years for
> > > advertising that doesn't support the group interest to be taken off the
> > > website, or for a researcher with unique ideas and much stronger
> > > associations to the consortium to finally get priority over an
> > > undergraduate who left the group years earlier. Even an academic
> > > organization with $350,000 annual revenue shouldn't continue to rely
> on a
> > > former member to donate his time and money just to keep their CI up and
> > > running, regardless of his generosity in doing so. The DE Shaw Group
> is
> > a
> > > professional organization of extremely talented, probably overworked
> > > individuals united by their goals of advancing molecular simulations.
> > Stop
> > > comparing the benchmarks unless you want to start comparing the
> > > organizations.
> > >
> > > Dave
> > >
> > >
> > > On Thu, May 13, 2021 at 4:48 PM Scott Le Grand <varelse2005.gmail.com>
> > > wrote:
> > >
> > > > To me, it's a sales trick until they demonstrate numerical stability
> to
> > > the
> > > > level Ross and I did with SPFP and SPDP. Have they? But even if it's
> > not
> > > > that stable, at least customers can make an informed choice with such
> > > data,
> > > > no? Also, how often are they rebuilding the neighbor list? Is it a
> > fixed
> > > > interval like GROMACS or is there a skin test?
> > > >
> > > > I am rethinking all this currently and I have friends who think
> > Neighbor
> > > > lists are obsolete if we move to higher timesteps and larger nonbond
> > > > cutoffs, but that brings us to how do we handle exclusions and
> that's a
> > > > rabbit hole. But... Coincidentally, SPFP's perfect force conservation
> > can
> > > > let you add and subtract them if you cap their magnitudes or use some
> > > > variant of softcore to control dynamic range. But are they doing
> > anything
> > > > like this? Details are everything!
> > > >
> > > > On Thu, May 13, 2021 at 1:39 PM Michael R Shirts <
> > > > Michael.Shirts.colorado.edu> wrote:
> > > >
> > > > > > and they skipped calculating the Ewald Sum every other iteration
> > > > (thanks
> > > > > Adrian!).
> > > > >
> > > > > In their semi-defense, IIRC, their default on all DESMOND
> simulations
> > > for
> > > > > a while has been to do multiple timestepping of forces, including
> > Ewald
> > > > sum
> > > > > every other timestep. It's not entirely clear to me if this is
> > > > sufficiently
> > > > > accurate, and they definitely should make that clearer that they
> are
> > > > doing
> > > > > something different, but it's a valid approach (that more people
> > should
> > > > be
> > > > > investigating!) and it's not just a sales trick. Not that there
> > aren't
> > > > > also sales tricks out there.
> > > > >
> > > > > Best,
> > > > > ~~~~~~~~~~~~~~~~
> > > > > Michael Shirts
> > > > > Associate Professor
> > > > > michael.shirts.colorado.edu
> > > > > http://www.colorado.edu/lab/shirtsgroup/
> > > > > Phone: (303) 735-7860
> > > > > Office: JSCBB C123
> > > > > Department of Chemical and Biological Engineering
> > > > > University of Colorado Boulder
> > > > >
> > > > >
> > > > > On 5/13/21, 1:27 PM, "Scott Le Grand" <varelse2005.gmail.com>
> > wrote:
> > > > >
> > > > > So, we're all getting our knickers in a bunch over an Apples to
> > > > Oranges
> > > > > Desmond to AMBER performance comparison.
> > > > >
> > > > > Please don't...
> > > > >
> > > > > They cheated, because that's what they do to keep their
> investors
> > > > > happy.
> > > > > They used a 32^3 grid, and they skipped calculating the Ewald
> Sum
> > > > every
> > > > > other iteration (thanks Adrian!). Rather than get upset here,
> > point
> > > > and
> > > > > laugh at DE Shaw et al. that they are afraid to go head to head
> > > with
> > > > > AMBER,
> > > > > and if they do (and they won't because they're chicken bawk
> bawk
> > > > > bawk), we
> > > > > have the people to address that as well.
> > > > >
> > > > > At our end, there's a ~50% or so performance deficit in AMBER
> 20
> > we
> > > > > need to
> > > > > fix. I've already fixed 2/3 of that building PMEMD 2.0 (770
> > ns/day
> > > > > DHFR 2
> > > > > fs already). Let them prance about with their greasy kids stuff
> > > > > desperate
> > > > > approximations and cheats, SPFP remains performance and
> accuracy
> > > with
> > > > > compromise and if they want to pick a fight with SPFP, make
> them
> > do
> > > > the
> > > > > work to demonstrate equivalent numerical stability (spoilers:
> > they
> > > > > won't
> > > > > because they can't but oh the bellyacheing and handwaving they
> > are
> > > > > willing
> > > > > to do, just watch).
> > > > >
> > > > > Scott
> > > > > _______________________________________________
> > > > > AMBER-Developers mailing list
> > > > > AMBER-Developers.ambermd.org
> > > > > http://lists.ambermd.org/mailman/listinfo/amber-developers
> > > > >
> > > > > _______________________________________________
> > > > > AMBER-Developers mailing list
> > > > > AMBER-Developers.ambermd.org
> > > > > http://lists.ambermd.org/mailman/listinfo/amber-developers
> > > > >
> > > > _______________________________________________
> > > > AMBER-Developers mailing list
> > > > AMBER-Developers.ambermd.org
> > > > http://lists.ambermd.org/mailman/listinfo/amber-developers
> > > >
> > > _______________________________________________
> > > AMBER-Developers mailing list
> > > AMBER-Developers.ambermd.org
> > > http://lists.ambermd.org/mailman/listinfo/amber-developers
> > >
> > _______________________________________________
> > AMBER-Developers mailing list
> > AMBER-Developers.ambermd.org
> > http://lists.ambermd.org/mailman/listinfo/amber-developers
> >
> _______________________________________________
> AMBER-Developers mailing list
> AMBER-Developers.ambermd.org
> http://lists.ambermd.org/mailman/listinfo/amber-developers
>
_______________________________________________
AMBER-Developers mailing list
AMBER-Developers.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber-developers
Received on Fri May 14 2021 - 09:00:04 PDT