Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

multi-GPU support for mine_hard_negatives #2967

Open
wants to merge 8 commits into
base: v3.1-release
Choose a base branch
from

Conversation

alperctnkaya
Copy link

Added support for multi-GPU encoding in sentence embeddings with model.encode_multi_process

tomaarsen and others added 8 commits September 26, 2024 12:50
UKPLab#2957)

Required to avoid issues with loading expecting local files, but the root is kept on the Hub to avoid downloading the entire repository (with multiple model files, etc.)
Added support for multi-GPU encoding in sentence embeddings with model.encode_multi_process
…#2951)

* Added the possibility of masking the prompts if the tokenizer is left-padded.

* Simplify code

* Remove unrelated changes

* Move prompt_mask into the Transformer model

* Added query and corpus prompts to Information Retrieval Evaluator

* Fix for failing test

* Fix for pooling when mask is not passed

* Fix device placement for prompt_mask

* Revert left-padding changes

* Revert left-padding changes
…tiple evaluators (UKPLab#2963)

Also rename "loss" to Validation Loss
* Add negatives in CachedGISTEmbedLoss

* Add multi-negative n-tuples as valid inputs in 3 losses

---------

Co-authored-by: Tom Aarsen <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants