-
Notifications
You must be signed in to change notification settings - Fork 445
Issues: InternLM/lmdeploy
[Benchmark] benchmarks on different cuda architecture with mo...
#815
opened Dec 11, 2023 by
lvhan028
Open
9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug] llama2-70B NTK,输入长度只能支持到8K?
#479
opened Sep 26, 2023 by
sjzhou4
updated Sep 27, 2023
2 tasks done
[Deploy Error] ValueError: If
eos_token_id
is defined, make sure that pad_token_id
is defined.
#522
opened Oct 1, 2023 by
vansin
updated Oct 11, 2023
codellama-Instruct版本地测试运行和服务部署返回的结果不一致
#579
opened Oct 18, 2023 by
for-just-we
updated Oct 19, 2023
1 of 2 tasks
[Feature] may support Mistral-7B recently? whatever Turbomind or Pytorch
backlog
#534
opened Oct 9, 2023 by
varuy322
updated Oct 26, 2023
[Bug] "cos_cached" shape incorrect error.
#664
opened Nov 8, 2023 by
WarrenZhao
updated Nov 8, 2023
2 tasks
支持多长输入 TurboMind supports Qwen-7B, dynamic NTK-RoPE scaling and dynamic logN scaling
#430
opened Sep 19, 2023 by
yuanjie-ai
updated Nov 14, 2023
[Feature] how to open window attention in qwen-14B?
backlog
#638
opened Nov 2, 2023 by
amulil
updated Nov 17, 2023
[Feature] Any Plan to implement INT8 weight-only quantizaton
#703
opened Nov 17, 2023 by
yunzhongyan0
updated Nov 17, 2023
[Bug] The GPU memory doesn't change after changing batch_size
#731
opened Nov 22, 2023 by
hxdbf
updated Nov 22, 2023
2 tasks
【Design Questinon】any plan to decouple batching and cache from llama?
backlog
#476
opened Sep 26, 2023 by
jinuxstyle
updated Nov 23, 2023
[Performance] How many gains does Interactive Inference Mode contribute to the throughput?
#565
opened Oct 17, 2023 by
nullxjx
updated Jan 10, 2024
[Question]请问这里的有关grid norm的计算逻辑有什么理论指导吗?
#1014
opened Jan 22, 2024 by
rightchose
updated Jan 22, 2024
[RFC] Refactor chat template and remove model name from engine config
RFC
#1065
opened Jan 30, 2024 by
AllentDan
updated Feb 5, 2024
[Bug] 使用internlm2-chat-7b 微调后的自制模型,4bit量化后无法使用
#1109
opened Feb 3, 2024 by
zhanghui-china
updated Feb 6, 2024
2 tasks done
Previous Next
ProTip!
no:milestone will show everything without a milestone.