adding support for multimers via residue_index hack#25
adding support for multimers via residue_index hack#25sokrypton wants to merge 3 commits intoHeliXonProtein:mainfrom
Conversation
|
I'm not sure if I implemented RoPE() residue index offset correctly. Setting residue index offset for relative-positional embedding works, but if you touch RoPE, it crashes! Here is what I tried for RoPE(): # default
position = torch.arange(total_length).cuda()# custom residue index (with 200 offset between chains)
position = residue_index# repeating residue index
position = torch.arange(subunit_length).repeat(subunits).cuda() |
|
Hi @sokrypton, Thanks for this info, we'll take a look into it. Best |
|
Hi Dr. @sokrypton, Would you care to give us an example input for this pr? |
|
For hetero-dimer: For homo-oligomer: |
|
Hi Dr. @sokrypton, We have reproduced your results and agree that it is really peculiar. We are going to take a deeper look into the reason for this. As for the training set of the language model. We directly take them from Uniref 50 sequences. |
|
I guess... since the LM was trained on single-chains. There is no reason to expect it to generalize to proteins it hasn't seen before, especially protein multimers. I suspect it sometimes works for protein multimers when the multi-chain protein looks like a multi-domain protein in the LM training set. |
|
Hi, We have been looking into this issue and found one problem with RoPE not necessarily regarding this problem, namely that it is symmetric, which may not be what we want and we are phasing it out soon. But still this problem persists and we do not really have an idea yet. |
|
@sokrypton @RuiWang1998 As a hungry user, are the problems with the current branch technical or scientific? |


No description provided.