-
Notifications
You must be signed in to change notification settings - Fork 6
Open
Description
Dear authors,
Thank you for your great work! From my understanding, the training pipeline involves first pre-training the TA-Tok tokenizer and de-tokenizer, followed by jointly training the full MLLM. I've found the code for the MLLM training using the released TA-Tok checkpoint. Would it be possible to also release the code for pre-training the TA-Tok components themselves?
Btw, I'm also wondering how many GPU hours it takes to train TA-Tok themselves?
Metadata
Metadata
Assignees
Labels
No labels