-
Notifications
You must be signed in to change notification settings - Fork 30
Issues: intel/llm-on-ray
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Benchmark] Load config from yaml and output results with multiple formats
#82
opened Jan 24, 2024 by
xwu99
[DOC] Add Coding Style Guide and how to format code to CONTRIBUTING.md
#46
opened Jan 10, 2024 by
xwu99
[Serving] Add a table of models and corresponding supported parameters
#51
opened Jan 11, 2024 by
KepingYan
Not able to run inference server for mistral 7b model, mpt-7b model on Ray
#65
opened Jan 18, 2024 by
dkiran1
Getting error while executing query_openai_sdk.py to test the inference
#66
opened Jan 18, 2024 by
dkiran1
[Quantization] Support loading AWQ, GPTQ, GGUF/GGML quantized models
#85
opened Jan 26, 2024 by
xwu99
Support and validate model Mixtral-8x7B
enhancement
New feature or request
#119
opened Feb 23, 2024 by
carsonwang
Support functions/tools in OpenAI API
enhancement
New feature or request
#121
opened Feb 23, 2024 by
carsonwang
Add ipex extra in pyproject.toml to use restricted transformers version
#127
opened Feb 29, 2024 by
jiafuzha
Openai API not allow temperature=0.0 for llama-2-7b-chat-hf
#139
opened Mar 12, 2024 by
yutianchen666
Calculate correct input length for every prompt in a single batch
#222
opened May 14, 2024 by
kira-lin
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.