Hi,
If it costs us nothing then why not scale PMEMD beyond 999,999
atoms. Someone out there might want to do 1MM+ atom simulation with
the AMBER program suite! Kennie
On 4 Dec 2007, at 2:14 PM, Robert Duke wrote:
> Hello folks!
> I am working hard on high-scaling pmemd code, and in the course of
> the work it became clear to me, due to large async i/o buffer and
> other issues, that going to very high atom counts may require a
> bunch of extra work, especially on certain platforms (BG/L in
> particular...). I posed the question below to Dave Case; he
> suggested I bounce it off the list, so here it is. The crux of the
> matter is how people feel about having an MD capability in pmemd
> for systems bigger than 999,999 atoms in the next release. Please
> respond to the dev list if you have strong feelings in either
> direction.
> Thanks much! - Bob
>
> ----- Original Message ----- From: "Robert Duke" <rduke.email.unc.edu>
> To: "David A. Case" <case.scripps.edu>
> Sent: Tuesday, December 04, 2007 8:45 AM
> Subject: How many atoms?
>
>
>> Hi Dave,
>> Just thought I would pulse you about how strong the desire is to
>> go above 1,000,000 atom systems in the next release. I personally
>> see this as more an advertising issue than real science; it's hard
>> to get good statistics/good science on 100,000 atoms let alone
>> 10,000,000 atoms. However, we do have competition. So the prmtop
>> is not an issue, but the inpcrd format is, and one thing that
>> could be done is to move to supporting the same type of flexible
>> format in the inpcrd as we do in the new-style prmtop. Tom D. has
>> an inpcrd format in amoeba that would probably do the trick; I can
>> easily read this in pmemd but not yet write it (I actually have
>> pulled the code out - left it in the amoeba version of course,
>> but can put it back in as needed). I ask the question now because
>> I am hitting size issues already on BG/L on something like
>> cellulose. Some of this I can fix; some of it really is more
>> appropriately fixed by running on 64 bit memory systems where
>> there actually is a multi-GB physical memory. The problem is
>> particularly bad with some new code I am developing, due to
>> extensive async i/o and requirements for buffers that at least
>> theoretically could be pretty big (up to natom possible; by
>> spending a couple of days writing really complicated code I can
>> actually handle this in small amounts of space with effectively no
>> performance impact - but it is the sort of thing that will be
>> touchy and require additional testing). Anyway, I do want to
>> gauge the desire to move up past 999,999 atoms, and make the point
>> that on something like BG/L, it would actually require a lot more
>> work to be able to run multi-million atom problems (basically got
>> to go back and look at all the allocations, make them dense rather
>> than sparse by doing all indexing through lists, allow for
>> adaptive minimal i/o buffers, etc. etc. - messy stuff, some of it
>> sourcing from having to allocate lots of arrays dimensioned by
>> natom).
>> Best Regards - Bob
>
Professor Kenneth M. Merz, Jr.
Department of Chemistry
Quantum Theory Project
2328 New Physics Building
PO Box 118435
University of Florida
Gainesville, Florida 32611-8435
e-mail: merz.qtp.ufl.edu
http://www.qtp.ufl.edu/~merz
Phone: 352-392-6973
FAX: 352-392-8722
Cell: 814-360-0376
Received on Wed Dec 05 2007 - 06:07:35 PST