-
Notifications
You must be signed in to change notification settings - Fork 445
Issues: InternLM/lmdeploy
[Benchmark] benchmarks on different cuda architecture with mo...
#815
opened Dec 11, 2023 by
lvhan028
Open
9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug] generation profile hangs on Mixtral-8x7B-Instruct-v0.1 with pytorch backend
#2948
opened Dec 24, 2024 by
zhulinJulia24
3 tasks
[Bug] Does PytorchEngine Visual Model Support Prefix Caching?
#2789
opened Nov 21, 2024 by
OftenDream
3 tasks
[Bug] pytorch backend 's precision points loss 1.0-2.5 points between main code and v0.6.1 on some models.
#2679
opened Oct 29, 2024 by
zhulinJulia24
3 tasks
[Docs] 问lmdeploy中的w8a8-triton实现是否有 实际llm(如llama2,qwen2)的推理速度加速效果的benchmark测试?
#2567
opened Oct 9, 2024 by
brisker
[Bug] output not consistent with different max_prefill_token_num for long context input on pytorch engine
#2457
opened Sep 12, 2024 by
RunningLeon
3 tasks done
[Bug] use openai server, request get asyncio.exceptions.TimeoutError
#2360
opened Aug 22, 2024 by
wlwqq
3 tasks done
[Bug] PyTorch Engine poor performance compared to vllm
#1449
opened Apr 18, 2024 by
jjjjohnson
2 tasks done
[Docs] got an unexpected keyword argument 'enable_lora'
#1151
opened Feb 19, 2024 by
sleepwalker2017
[Deploy Error] ValueError: If
eos_token_id
is defined, make sure that pad_token_id
is defined.
#522
opened Oct 1, 2023 by
vansin
ProTip!
Find all open issues with in progress development work with linked:pr.