Skip to content

Conversation

@jmercat
Copy link
Collaborator

@jmercat jmercat commented Jul 31, 2024

We might not want to merge this because it is hacky and there might be a usage that I don't foresee that could be impacted.

Problem:
Somewhere in llm-foundry or composer, the batches are appended with many stop tokens which makes the eval very slow.

Solution:
In OpenLMforCausalLM tests for 0s at the end of the input_ids batch and remove them. Add back fake logits at the end....

Consideration:
This hack could be implemented in SimpleComposerOpenLMCausalLM instead and would have less risk to have unforeseen impacts.
There shouldn't really be a need for this but I could not understand what was happening in Composer or llm-foundry or find where/why the inputs are appended with 0s.

@jmercat jmercat mentioned this pull request Jul 31, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant