-
Notifications
You must be signed in to change notification settings - Fork 97
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Shared memory issue while running Protenix example #74
Comments
@rakeshr10 it's similar issue to #2 |
@zhangyuxuann Thanks for your reply. Putting shared memory worked but I was wondering if there is a way to run without using shared memory. I need to do large scale runs by spinning up pods on Kubernetes, so using /dev/shm might become an issue. Another question is there a way to download the model weights and point it to the inference script instead of downloading it every time a new pod is spun up for inference. It would also be helpful if you could provide some guidance on how to use the deepspeed evoformer attention kernel without compiling it every time a new pod is spun up. I noticed that the ninja workers are being called to compile evoformer code when it does not find the compiled cuda kernel in /root/.cache folder, this ends up taking a lot of time even though the inference is faster. |
if you build your image from docker, just copy the above modified builder.py like
|
@zhangyuxuann Thanks for your reply. I tried some of your suggestions and noticed these.
|
|
@zhangyuxuann Thanks for your reply. I was able to add fastfold_layer_norm_cuda.so in the path and now it works. I noticed the inference settings are same as in the picture so won't change them. Regarding my first question what I mentioned was when setting --num_workers 0 the inference does not require shared memory when running on GPU. My question was since I am setting --num_workers 0 does it cause any issue while running protenix predict command on a multisequence json file which I am assuming is batch inference and the json file contains multiple sequences and msa files. I tried running the colabfold_msa.py script on a single multiple sequence pairs containing fasta file. While the colabfold_search generates multiple a3m files, the a3m processor does not convert all a3m files into corresponding pairing and nonpairing a3m files for running protenix inference on multiple a3m files. Also it would be great if this script could take the fasta and a3m files and converts them into json inputs file for running protenix inference. |
The issue may be caused by the fasta format, can you provide the input file, to reproduce the issue. |
This is the fasta file with text extension. It has multiple pairs of sequences which can be submitted as a batch sequence search job to mmseqs using colabfold_search command to do msa generation for all pairs simultaneously. It will be good if all of the a3m files generated for all pairs using the colabfold_search command can be converted to protenix compatible a3m and json files using colabfold_msa.py script |
@JinyuanSun Were you able to look into this issue and request for the script to handle multiple a3m files. |
The easiest approach would be to split the file into individual FASTA sequence files and then use a
|
Hi, I tried to run the example I get the following error. How can I resolve this memory issue?
protenix predict --input examples/example.json --out_dir ./output --seeds 101
The text was updated successfully, but these errors were encountered: