2022-01-28 GCN charge models meeting notes

 Date

Jan 28, 2022

 Participants

  • @Jeffrey Wagner

  • @Simon Boothroyd

  • @David Mobley

  • @Michael Shirts

  • @Michael Gilson

  • @Owen Madin

  • @Christopher Bayly

  • Bill Swope

  • @John Chodera

  • @Joshua Horton

  • @Pavan Behara

  • @Willa Wang

  • Yu-Tang Lin

  • Yuanqing Wang

 Goals

  •  

 Discussion topics

Item

Notes

Item

Notes

Slide 1

  • MG – In other discussions, there was a statement that “ELF10” would be in Rosemary. What’s the plan?

    • SB – The plan is for as much to move forward independently as possible. So the biopolymer force should move forward with !LF10 AM1BCC. If and when the GCN is done and validated, if Rosemary isn’t out yet, then it will

  • CB – SB, you said that, with Rosemary, we could have these two charge models dropped in. So in case A, the two charge models (explicit AM1BCC/library charges and GCN charges) converge and we can swap them out. But in the other, we need to fit all the valence terms so we need to pick the charge model before we start training.

    • JC – So, the hope is that the two converge, and if they don’t, then we won’t incorporate the GCN in Rosemary.

  • MS – So, this is specifically a GCN fit to AM1BCC ELF10?

    • SB – There are some choices, I’ll go over them. The MVP is a GCN fit to AM1, and we apply SMARTS based BCCs on top of it. The goal is to target AM1, but it’s possible that we could do RESP charges since they’re all kinda based on the same thing.

    • CB – Agree. It’s my hope that the charge models will be convergent, and that future work can bring in RESP charges and better stuff as it comes.

  •  

Slide 3

  • OM – Is there a decision on whether the network would be trained to AM1 and apply BCCs on top, or train to AM1BCC?

    • SB – My preference is to learn AM1, and then apply BCCs on top.

  • JC – Bronze medal is pretty ambitious - If the GCN can reproduce AM1 charges within inter-toolkit differences then I think that’s a win.

    • SB – Agree, I would just mention that I think our ultimate metric of success should be HFEs, not just charge differences.

  • CB – I’d like to introduce the idea of a “platinum medal”, where we fit the GCN to reproduce ESP, not just RESP point charges.

    • SB – I’ll answer that on the next slide.

    • CB – My idea here is to fit to a more fundamental “ground truth” from QM. The RESP point charges are already squeezed through an information bottleneck.

    • SB – I’d say that we would have already seen this… Since they’re all trying to reproduce ESPs it should all be the same thing.

    • CB – In practice we saw numerical instabilities in the early 1990s fitting to ESPs, and that’s why we had to add restraints. During that work, I found that there wasn’t a clear unique solution to the assignment of point charges, and charges would become large in magnitude. And I think vsites would fix a lot of this.

    • DC – Seconding numerical instabilities

    • DC – When you talk about RESP charges, are you thinking RESP2?

    • SB – RESP2 would be super expensive. When I was computing ESP data from the molecules from the industry benchmark set, I was finding that 8 cores would go through about 5 molecules per day.

    • DC – So if it takes a long time to even do them in vaccuum, then we’ll need BCC corrections, since we won’t have the even MORE expensive implicit solvent calcs.

    • SB – If we fit to polarizability then we should be in good shape

  •  

  •  

Slide 4

  • JW (not asked in interest of time) – Re: “RESP charges are expensive” - We sent a biggish dataset through QCA a few weeks ago that did about 50k mols. What was the differece between that and these expensive sets?

Slide 5

  • CB – AM1 is just supposed to be cheap. Maybe we don’t plan for this work to be “finished” - We could initially just aim to make a GCN for AM1, but we could gradually mix in RESP data to the training set as it becomes available.

    • JC – Agree. Could have a scaling parameter.

    • SB – That’s a good idea.

Slide 7

  • SB will add links to mentioned datasets

    • PB –

  • SB – I’m not sure where to put these datasets. Not sure that they’d be appropriate in qca-dataset-submission.

    • JW and SB will follow up offline

  • CB – As an industry guy, one thing that’s emerged in the cheminformatics world, is that the enamine isn’t actually that chemically diverse. The Riniker and Bleiziffer sets will probably capture the widest range of chemistry.

  • JC – The NCI250k is also a good idea.

  • DM – Yeah, I think the diversity of these should probably be Riniker > OpenFF industry > Enamine.

Slide 9

 

Slide 14

  • JC – Should the GCN include stereochemistry? Mols with two stereocenters will have different electron densities

    • SB – It’ll be really hard to encode the meaning of stereo for a GCN

    • JC – It could have an effect for just close neighbors

    • JW – I’d hazard a guess that bond stereo is way more important than atom centered stereo. So maybe we can get most of the gains by avoid the complications of atom centered stereo by just defining the relationship between the 4 atoms that are cis or trans to each other across a double bond.

    • SB – We could look into this by having a carefully designed training set.

  •  

Slide 16

  • JC – One tough thing is that we can get very different performances based on different random seeds/initial weights. Also the choice of number of layers is kinda arbitrary. Also vanishing gradient problem.

  • SB – I’ll chat with Yuanqing Wang about some of these details

 

  • MS – What is our decision on BCC optimization to experiment?

    • JC – That’s on the silver medal goals

    • SB – It depends on whether we’re fitting to AM1 or RESP….?

    • JC – Are we going to train to QM and condensed phase simultaneously?

    • MS – So it looks like we should reoptimize BCCs with some moderate priors.

    • SB – We had looked at this, basically “how do we dial in the weights?”. This was because we had found that QM was dominating the optimization. But even when balanced well, for example in the GAFF2 paper, the gains were really modest. So I’m most optimistic about vsites

    • JC – I think host-guest will be the most important thing. What’s the status of Jeff S’s work?

      • JW – I think Jeff S had some results on this front, refitting GBSA parameters. IIRC the values changes a lot when trained on host-guest data, but they did improve in performance on predicting host-guest binding. Given the magnitude of change, I’m not sure whether the results would be expected to generalize past host-guest. But this is a hazy recollection - It would be best to ask him directly or look at meeting notes/slides.

 Action items

 Decisions