Skip to content

Question Regarding TA-Tok Training Code #7

@nihaotian1

Description

@nihaotian1

Dear authors,

Thank you for your great work! From my understanding, the training pipeline involves first pre-training the TA-Tok tokenizer and de-tokenizer, followed by jointly training the full MLLM. I've found the code for the MLLM training using the released TA-Tok checkpoint. Would it be possible to also release the code for pre-training the TA-Tok components themselves?

Btw, I'm also wondering how many GPU hours it takes to train TA-Tok themselves?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions