-
Notifications
You must be signed in to change notification settings - Fork 116
Issues: predibase/lorax
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Fail to run Phi-3
bug
Something isn't working
#485
opened May 22, 2024 by
prd-tuong-nguyen
1 of 4 tasks
make install
insufficient for running llama3-8B-Instruct
documentation
#484
opened May 22, 2024 by
fozziethebeat
2 of 4 tasks
Add HTTP status codes to docs
documentation
Improvements or additions to documentation
good first issue
Good for newcomers
#481
opened May 20, 2024 by
noyoshi
When caching adapters, cache the adapter ID + the API token pair
enhancement
New feature or request
good first issue
Good for newcomers
#479
opened May 20, 2024 by
noyoshi
Reject unknown fields from API requests
enhancement
New feature or request
good first issue
Good for newcomers
#478
opened May 20, 2024 by
noyoshi
Support inference on INF2 instance
enhancement
New feature or request
#477
opened May 20, 2024 by
prd-tuong-nguyen
Improve warmup checking for max new tokens when using speculative decoding
bug
Something isn't working
good first issue
Good for newcomers
#474
opened May 17, 2024 by
tgaddair
Bug Report: lorax-launcher failed with --source "s3" for model_id "mistralai/Mistral-7B-Instruct-v0.2"
bug
Something isn't working
#473
opened May 17, 2024 by
donjing
1 of 4 tasks
Ensure api_token is not included in the response on error
bug
Something isn't working
#469
opened May 15, 2024 by
tgaddair
Add all launcher args as optional in the Helm charts
enhancement
New feature or request
#465
opened May 9, 2024 by
tgaddair
Retrieve all lora models from Huggingface hub by base model setting.
enhancement
New feature or request
good first issue
Good for newcomers
#463
opened May 8, 2024 by
svjack
Improve async load for adapters to avoid main thread lockups in server
enhancement
New feature or request
#457
opened May 3, 2024 by
tgaddair
Batch inference endpoint (OpenAI compatible)
enhancement
New feature or request
#448
opened Apr 30, 2024 by
tgaddair
Llama3-8b-Instruct won't stop generating
bug
Something isn't working
#442
opened Apr 27, 2024 by
ekim322
4 tasks
Idefics2 and LLaVA
enhancement
New feature or request
#439
opened Apr 26, 2024 by
joaomsimoes
2 tasks done
Improve the latency of New feature or request
load_batched_adapter_weights
enhancement
#433
opened Apr 22, 2024 by
thincal
Combining multiple LoRA adapters
question
Further information is requested
#429
opened Apr 21, 2024 by
winglian
Inference with AWQ quantized base model + compile enabled results in the <unk> tokens
bug
Something isn't working
#426
opened Apr 19, 2024 by
thincal
4 tasks
Error: Warmup(Generation("'bool' object has no attribute 'dtype'"))
bug
Something isn't working
#422
opened Apr 18, 2024 by
KrisWongz
1 of 4 tasks
Can't run Mistral quantized on T4
enhancement
New feature or request
#417
opened Apr 16, 2024 by
emillykkejensen
2 of 4 tasks
Async client to backoff when model overloaded
enhancement
New feature or request
#412
opened Apr 12, 2024 by
jppgks
Previous Next
ProTip!
Exclude everything labeled
bug
with -label:bug.