We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
如题,使用vllm_infer.py进行推理时,怎么返回输出token的prob呢?results = LLM(**engine_args).generate(inputs, sampling_params, lora_request=lora_request) 这个函数应该怎么修改呢? 求助,急急急~~~
results = LLM(**engine_args).generate(inputs, sampling_params, lora_request=lora_request)
No response
The text was updated successfully, but these errors were encountered:
No branches or pull requests
Reminder
Description
如题,使用vllm_infer.py进行推理时,怎么返回输出token的prob呢?
results = LLM(**engine_args).generate(inputs, sampling_params, lora_request=lora_request)
这个函数应该怎么修改呢?求助,急急急~~~
Pull Request
No response
The text was updated successfully, but these errors were encountered: