You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am deveping a model that uses your dataset with docking energies. We realized that the docking energy from the AutoDock Vina can have quite a significant fluctuation depending on the input optimized molecule geometry, random seed, and the hyperparameters chosen for the docking process. Therefore, this makes it hard for us to reproduce the set of reward/energies obtained from your dataset.
Since the verification method can be unstable, instead, we would like to further study your model by looking at some of the top candidates genrated from GFlownet trained on this dataset.
In the paper, you described that $10^{6}$ candidates were generated and the top-1000 were analyzed in terms of their average reward.
Could you please kindly provide the $10^{6}$ or just the top-1000 generated ones along with their scores?
Thank you very much!
Haote
The text was updated successfully, but these errors were encountered:
Hi Haote, I'm not sure I still have the original results saved from the 2021 paper. I'll check.
Do note that the top-1000 scores we report are strictly in terms of the pretrained proxy's score on them, i.e. we did not rerun docking for the compounds we found. The intention with this benchmark wasn't really to solve candidate generation for docking, but rather just to have a hard enough benchmark to work against. You're correct that AutoDock produces stochastic results.
In terms of AutoDock hyperparameters, I'm assuming you've already seen how we define the geometry here and how we call vina here?
Dear Authors,
I am deveping a model that uses your dataset with docking energies. We realized that the docking energy from the AutoDock Vina can have quite a significant fluctuation depending on the input optimized molecule geometry, random seed, and the hyperparameters chosen for the docking process. Therefore, this makes it hard for us to reproduce the set of reward/energies obtained from your dataset.
Since the verification method can be unstable, instead, we would like to further study your model by looking at some of the top candidates genrated from GFlownet trained on this dataset.
In the paper, you described that$10^{6}$ candidates were generated and the top-1000 were analyzed in terms of their average reward.$10^{6}$ or just the top-1000 generated ones along with their scores?
Could you please kindly provide the
Thank you very much!
Haote
The text was updated successfully, but these errors were encountered: