Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The performance of DDSM for unconditional DNA generation #2

Open
Zehui127 opened this issue Oct 18, 2023 · 2 comments
Open

The performance of DDSM for unconditional DNA generation #2

Zehui127 opened this issue Oct 18, 2023 · 2 comments

Comments

@Zehui127
Copy link

Zehui127 commented Oct 18, 2023

Dear Team,

I have been working on developing the generative model for DNA sequences. For a fair comparison, I compare different algorithms in the unconditional generation case. It seems that DDSM fails to capture the motif distribution in the unconditional DNA sequence generation case. By unconditional generation, I mean the transcription profile is not supported as conditions.

I wonder if you have tried to use DDSM for unconditional DNA sequence generation and what is the expected result.

PS: I tried both time dilation and without dilation, and the generated samples don't seem to be capturing the motif distribution of input sequences. The training script is available.

Best,
Zehui

@PavelAvdeyev
Copy link
Collaborator

Dear Zehui,

If you are using training hyperparameters provided by the paper, you will get sub-optimal results on unconditional generation since our training setups are not optimized for unconditional generation. Therefore, it would be hard to compare with DDSM rigorously on unconditional generation. If you have to perform a comparison on unconditional generation, one option could be fine-tuning the conditional generation model provided by us for unconditional generation. You can also train DDSM for unconditional generation from scratch but it will probably require some hyperparameter tweaks.

@Zehui127
Copy link
Author

Dear @PavelAvdeyev ,

Thanks for your response on this. I indeed train from stretch for unconditional generation. But it seems to be having issue in terms of the quality of the generation sequences. I notice one of the potential issue is that the scoreNet used in the current code is relatively small, we will do some incremental change on the score net and see if it will work.

Best,
Zehui Li

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants