Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Minor features #2249

Open
wants to merge 3 commits into
base: main
Choose a base branch
from
Open

Conversation

artemorloff
Copy link
Contributor

Features:

  • disable fewshot_as_multiturn when apply_chat_template is not passed or num_fewshots=0. Why failing the run? For zero-shot setup multiturn==simple chat template, so no error at all. If chat_template is not enabled, then throw warning and disable multiturn (as long as it is not available without chat_template)
  • pass predict_only into filters apply method. Why? The filters are designed to be used even with additional ML models (reward, for example). Then if one runs lm-eval with predict_only this may mean that the filter is not to be used. No user may customize filters to use predict_only info to manage filters behaviour
  • add filter_device param from cli. There was a TODO about it. If I use another LLM as a filter, I may need to pass device that DIFFERS from one used to run the "main" LLM. Like llm-as-a-judge or LLMs to score the generations
  • disable ensure_ascii for apply_chat_template method of TemplateAPI class. Now cyrillic symbols are stored in a valid form
  • add f1_macro and f1_micro metrics (aggregations in fact) to register to handle multi-class classification tasks
  • new param into model_args for APIs - timeout. When running vLLM server and using lm-eval in OpenAI API mode to make requests into this server, timeout may be increased (like to run Llama-3.1-405B, for me I had lots of connection errors, that have been solved by increasing the timeout param)

@CLAassistant
Copy link

CLAassistant commented Aug 25, 2024

CLA assistant check
All committers have signed the CLA.

lm_eval/__main__.py Outdated Show resolved Hide resolved
@baberabb
Copy link
Contributor

Hey @artemorloff. This generally looks good to me! but I do want to get it reviewed by @haileyschoelkopf, just to check if we are ok with backward compatibility.

@artemorloff
Copy link
Contributor Author

@haileyschoelkopf still actual?

@sytelus
Copy link

sytelus commented Sep 14, 2024

timeout is great new feature!! Can we please merge this?

@artemorloff
Copy link
Contributor Author

@haileyschoelkopf @lintangsutawika hi! seems you have already applied one of the features that I suggest. What do you think about the rest? Something I can add?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants