-
Notifications
You must be signed in to change notification settings - Fork 199
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Reproducibility of benchmark evaluation #280
Comments
Hi @Bae-SungHan
We don’t have a number for a single seed, but 0.69 does not sound completely off the range.
CLI is not for automation. My comment is not about quality of prediction, it is about healthy engineering practices - you’re given a nice pythonic API with detailed examples so please use the right tool for the job.
I guess since numbers on published predictions match, you have working code. Sanity check: run against some other baseline with known performance, like RF2AA. If you want someone to validate code - please approach posebusters’ authors. Closing that issue.
This example looks wrong - chai-1 accepts ligands as SMILEs; CCDs are accepted as modified AAs in proteins.
Simple sanity check: check that uniprot sequences are paired, and for other DBs there is no pairing. Another remark is to look at per-complex predictions, and split into bins: number of ligands, max number of copies for a ligand, max number of copies for protein, max sequence length, total number of tokens - it may reveal a helpful pattern on where the discrepancies may come from. |
@arogozhnikov Thank you for kind response! |
Hi. 6 protein chains 5 protein chains (I thought they're subset of 6 protein chains from original cif, but they're not) Thank you |
Hi @Bae-SungHan,
you can also see this in the prediction below (orange is our prediction of crop that you linked to vs grey - ground truth assembly 1 from PDB). The second COM ligand is not part of the crop since it is quite far from the one that was arbitrarily selected. If you want to re-predict a crop using chai-1, you should not use interface for fasta inputs, but directly interact with ![]() |
Thank you for your kind explanation!
Is there anything wrong with the method I mentioned above?
Can you give me some advice for proper benchmark test to get a similar score in the paper? Thank you! |
Hi
I've been trying to reproduce PoseBusters benchmark score you reported, but unfortunately I haven't been able to do so yet. So I would like to ask if I could get help with this.
Here's my process to infer PoseBusters dataset locally;
Below is the code I used;
For the case of
8DP2_UMA
, the output of above is like;I wrote CCD code for the cofactor and SMILES for ligand of interest.
For the case of
8DP2_UMA
, the resulting input fasta is like;Got JackHmmer MSA on the latest version of UniProt, UniRef90, and MGnify DB and JackHmmer argument
-N 1 -E 0.0001 --incE 0.0001 --F1 0.0005 --F2 0.00005 --F3 0.0000005
Overall JackHmmer MSA process referred to the code in alphafold3
After getting a3m files for each DB, I used
merge_a3m_in_directory
method, to finalize the MSA to the Chai inputInfered with
chai fold
command.I modified
--num-trunk-recycles
arguments to 4, as you mentioned in the technical report.I got only 5 sample outputs with single seed.
Calculated ligand RMSD as I described in the Reproduce PoseBusters score #210 (comment) ranked with confidence score.
With steps as I described above, the resulting success rate (RSMD <=2.0) was only about 0.69, much lower than the score you reported, 0.77.
Even considering that I only inferred with single seed, not 5, there's quite big gap so I think there must be problem with my inference process.
Since I successfully reproduced the success rate with calculating ligand RMSD of results in https://chaiassets.com/chai-1/paper/assets/posebusters_predictions.zip, I don't think it's a problem with how to measure ligand rmsd.
Reasons I suspect is three
Unfortunately, I couldn't get any more information about these.
If you can give some advises for PoseBusters reproducing steps, it will be great help for me.
Thank you!
The text was updated successfully, but these errors were encountered: