Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ITG Loss Adjustment #225

Open
QVQZZZ opened this issue Dec 16, 2024 · 0 comments
Open

ITG Loss Adjustment #225

QVQZZZ opened this issue Dec 16, 2024 · 0 comments

Comments

@QVQZZZ
Copy link

QVQZZZ commented Dec 16, 2024

Thank you for your excellent work! Recently, I have been using the pretrain script and checkpoint you provided to perform post-pretraining on my dataset. My goal is vector retrieval. I have a few questions I'd like to ask:

  1. I noticed that the performance of the text embeddings is relatively lower compared to other llm baselines, but it offers faster inference speed. Do you think replacing the text model (which also means discarding the checkpoint) would be beneficial? Would the code modifications required for the replacement be significant?
  2. Using the pretrain script, I have currently frozen the ITG loss. For my task, do you think it is necessary to include the ITG loss? If I include some images and image descriptions generated by other large models (e.g., 4o) and combine them with the ITG loss, could this also benefit vector retrieval? Perhaps I should include the ITG loss in the early stages of training and freeze it later?
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant