Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Extracting vLLM metrics #2365

Open
vsmolyakov opened this issue Sep 29, 2024 · 0 comments
Open

Extracting vLLM metrics #2365

vsmolyakov opened this issue Sep 29, 2024 · 0 comments

Comments

@vsmolyakov
Copy link

I'm using a vLLM model with simple_evaluate function on a benchmark task and I'm very much interested in extracing vLLM metrics such as time to first token, time in queue, etc. I've been reading through the source code to see if this is currently supported and I found the following in lm_eval/models/vllm_causallms.py

            for output, context in zip(cont, context):
                generated_text = output.outputs[0].text
                res.append(generated_text)
                ...

I was wondering if the following could be added as an option for vLLM models:

            for output, context in zip(cont, context):
                generated_text = output.outputs[0].text
                generated_metrics = output.metrics        # extracting vLLM metrics
                res.append(generated_text)
                res.append(generated_metrics)             # appending metrics to results
                ...

Thanks! This feature would be greatly appreciated.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant