Skip to content

Issues: predibase/lorax

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

Always consider base model (no adapter) to be active enhancement New feature or request
#8 opened Nov 10, 2023 by tgaddair
Panic when adapter cannot be loaded bug Something isn't working
#14 opened Nov 12, 2023 by tgaddair
4 tasks
Add RoPE scaling CLI args enhancement New feature or request good first issue Good for newcomers
#142 opened Dec 19, 2023 by tgaddair
Extend testing enhancement New feature or request good first issue Good for newcomers
#148 opened Dec 22, 2023 by flozi00
Support multiple ranks per SGMV op enhancement New feature or request
#160 opened Jan 4, 2024 by tgaddair
Add per-adapter metrics to /metrics endpoint enhancement New feature or request good first issue Good for newcomers
#203 opened Jan 23, 2024 by tgaddair
Add a small GUI to interact with models easier enhancement New feature or request
#255 opened Feb 17, 2024 by magdyksaleh
Add ChatGLM as base model and support adapters enhancement New feature or request good first issue Good for newcomers
#280 opened Feb 26, 2024 by tgaddair
Include total time to generate tokens in final payload details enhancement New feature or request good first issue Good for newcomers
#326 opened Mar 13, 2024 by martindavis
Add support for control vector adapters per request enhancement New feature or request
#336 opened Mar 17, 2024 by tgaddair
Add in "--adapter-memory-fraction" to docs documentation Improvements or additions to documentation good first issue Good for newcomers
#355 opened Mar 22, 2024 by noyoshi
Support constrained generation of valid Python types enhancement New feature or request
#366 opened Mar 29, 2024 by jeffreyftang
Add support for AQLM quantization enhancement New feature or request
#388 opened Apr 4, 2024 by tgaddair
Misleading/wrong openapi schema in REST API docs for structured output documentation Improvements or additions to documentation
#389 opened Apr 5, 2024 by oscarjohansson94
2 of 4 tasks
Supporting inference with EETQ quantized model enhancement New feature or request
#391 opened Apr 5, 2024 by thincal
Fix PyTorch CUDA version in Docker bug Something isn't working
#438 opened Apr 25, 2024 by tgaddair
Batch inference endpoint (OpenAI compatible) enhancement New feature or request
#448 opened Apr 30, 2024 by tgaddair
Add all launcher args as optional in the Helm charts enhancement New feature or request
#465 opened May 9, 2024 by tgaddair
Support inference on INF2 instance enhancement New feature or request
#477 opened May 20, 2024 by prd-tuong-nguyen
Reject unknown fields from API requests enhancement New feature or request good first issue Good for newcomers
#478 opened May 20, 2024 by noyoshi
can't start my local llama3 model server with docker
#511 opened Jun 12, 2024 by cheney369
3 of 4 tasks
Add echo parameter in request
#518 opened Jun 17, 2024 by dennisrall
ProTip! What’s not been updated in a month: updated:<2025-01-28.