Recent Releases of openllm
openllm - v0.6.26
What's Changed
- chore(deps): bump openai from 1.70.0 to 1.73.0 in the production-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1175
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1176
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.25...v0.6.26
- Python
Published by github-actions[bot] 10 months ago
openllm - v0.6.24
What's Changed
- chore(deps): bump bentoml from 1.4.7 to 1.4.8 in the pip group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1171
- chore(deps): bump openai from 1.69.0 to 1.70.0 in the production-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1169
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.23...v0.6.24
- Python
Published by github-actions[bot] 11 months ago
openllm - v0.6.21
What's Changed
- chore(deps): bump the production-dependencies group with 2 updates by @dependabot in https://github.com/bentoml/OpenLLM/pull/1157
- feat(types): enable strict mypy by @aarnphm in https://github.com/bentoml/OpenLLM/pull/1158
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1159
- chore(deps): bump openai from 1.66.3 to 1.68.2 in the production-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1160
- chore(deps): bump the actions-dependencies group with 2 updates by @dependabot in https://github.com/bentoml/OpenLLM/pull/1161
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1162
- chore(deps): bump the production-dependencies group with 2 updates by @dependabot in https://github.com/bentoml/OpenLLM/pull/1163
- chore(deps): bump actions/setup-python from 5.4.0 to 5.5.0 in the actions-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1164
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1165
- chore: update instructions for deploy with openllm by @aarnphm in https://github.com/bentoml/OpenLLM/pull/1166
- feat: support --env by @aarnphm in https://github.com/bentoml/OpenLLM/pull/1167
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.20...v0.6.21
- Python
Published by github-actions[bot] 11 months ago
openllm - v0.6.20
What's Changed
- chore(deps): bump openai from 1.61.1 to 1.63.0 in the production-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1146
- docs: Fix deploy command by @Sherlock113 in https://github.com/bentoml/OpenLLM/pull/1147
- chore(deps): bump the production-dependencies group with 2 updates by @dependabot in https://github.com/bentoml/OpenLLM/pull/1149
- chore(deps): bump actions/upload-artifact from 4.6.0 to 4.6.1 in the actions-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1150
- docs: Fix serve command by @Sherlock113 in https://github.com/bentoml/OpenLLM/pull/1148
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1151
- chore(deps): bump the production-dependencies group with 2 updates by @dependabot in https://github.com/bentoml/OpenLLM/pull/1152
- chore(deps): bump actions/download-artifact from 4.1.8 to 4.1.9 in the actions-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1153
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1154
- chore(deps): bump the production-dependencies group with 2 updates by @dependabot in https://github.com/bentoml/OpenLLM/pull/1155
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1156
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.19...v0.6.20
- Python
Published by github-actions[bot] 12 months ago
openllm - v0.6.19
What's Changed
- chore(deps): bump openai from 1.61.0 to 1.61.1 in the production-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1143
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1144
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.18...v0.6.19
- Python
Published by github-actions[bot] about 1 year ago
openllm - v0.6.18
What's Changed
- chore(deps): bump the actions-dependencies group with 2 updates by @dependabot in https://github.com/bentoml/OpenLLM/pull/1132
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1133
- chore(deps): bump openai from 1.59.3 to 1.59.6 in the production-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1131
- chore(deps): bump openai from 1.59.6 to 1.59.8 in the production-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1134
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1135
- chore(deps): bump openai from 1.59.8 to 1.60.1 in the production-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1138
- chore(deps): bump pypa/gh-action-pypi-publish from 1.12.3 to 1.12.4 in the actions-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1139
- chore(deps): bump openai from 1.60.1 to 1.61.0 in the production-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1141
- chore(deps): bump actions/setup-python from 5.3.0 to 5.4.0 in the actions-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1142
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1140
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.17...v0.6.18
- Python
Published by github-actions[bot] about 1 year ago
openllm - v0.6.17
What's Changed
- chore(deps): bump openai from 1.57.4 to 1.58.1 in the production-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1123
- chore(deps): bump actions/upload-artifact from 4.4.3 to 4.5.0 in the actions-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1124
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1125
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1127
- chore(deps): bump openai from 1.58.1 to 1.59.3 in the production-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1126
- docs: Add Llama 3.3 to readme by @Sherlock113 in https://github.com/bentoml/OpenLLM/pull/1128
- fix: respect Python version of bento by @bojiang in https://github.com/bentoml/OpenLLM/pull/1130
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.16...v0.6.17
- Python
Published by bojiang about 1 year ago
openllm - v0.6.16
What's Changed
- chore(deps): bump openai from 1.55.0 to 1.55.3 in the production-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1116
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1117
- chore(deps): bump the production-dependencies group across 1 directory with 2 updates by @dependabot in https://github.com/bentoml/OpenLLM/pull/1120
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1119
- chore(deps): bump the actions-dependencies group with 2 updates by @dependabot in https://github.com/bentoml/OpenLLM/pull/1121
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1122
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.15...v0.6.16
- Python
Published by bojiang about 1 year ago
openllm - v0.6.15
What's Changed
- chore(deps): bump the actions-dependencies group across 1 directory with 4 updates by @dependabot in https://github.com/bentoml/OpenLLM/pull/1106
- chore(deps): bump openai from 1.52.0 to 1.53.0 in the production-dependencies group across 1 directory by @dependabot in https://github.com/bentoml/OpenLLM/pull/1105
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1104
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1109
- chore(deps): bump the actions-dependencies group across 1 directory with 2 updates by @dependabot in https://github.com/bentoml/OpenLLM/pull/1111
- chore(deps): bump the production-dependencies group across 1 directory with 2 updates by @dependabot in https://github.com/bentoml/OpenLLM/pull/1110
- chore(deps): bump openai from 1.54.4 to 1.55.0 in the production-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1113
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1114
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.14...v0.6.15
- Python
Published by github-actions[bot] about 1 year ago
openllm - v0.6.14
What's Changed
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1100
- chore(deps): bump openai from 1.51.2 to 1.52.0 in the production-dependencies group by @dependabot in https://github.com/bentoml/OpenLLM/pull/1099
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.13...v0.6.14
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.6.12
What's Changed
- chore(deps): bump actions/checkout from 4.1.7 to 4.2.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1090
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1091
- chore(deps): bump openai from 1.47.0 to 1.50.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1089
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1094
- chore(deps): bump the actions-dependencies group with 3 updates by @dependabot in https://github.com/bentoml/OpenLLM/pull/1098
- chore(deps): bump openai from 1.50.2 to 1.51.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1095
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.11...v0.6.12
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.6.11
What's Changed
- chore(deps): bump openai from 1.41.0 to 1.42.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1069
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1070
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1075
- chore(deps): bump pypa/gh-action-pypi-publish from 1.9.0 to 1.10.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1074
- chore(deps): bump actions/setup-python from 5.1.1 to 5.2.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1073
- chore(deps): bump actions/upload-artifact from 4.3.6 to 4.4.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1072
- chore(deps): bump openai from 1.42.0 to 1.43.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1071
- chore(deps): bump pypa/gh-action-pypi-publish from 1.10.0 to 1.10.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1079
- chore(deps): bump openai from 1.43.0 to 1.44.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1077
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1080
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1082
- chore(deps): bump openai from 1.44.0 to 1.45.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1081
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1086
- chore(deps): bump pypa/gh-action-pypi-publish from 1.10.1 to 1.10.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1085
- chore(deps): bump openai from 1.45.0 to 1.47.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1084
- docs: Update model table by @Sherlock113 in https://github.com/bentoml/OpenLLM/pull/1087
- docs: Update model by @Sherlock113 in https://github.com/bentoml/OpenLLM/pull/1088
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.10...v0.6.11
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.6.10
What's Changed
- chore(deps): bump openai from 1.38.0 to 1.41.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1065
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1061
- chore(deps): bump actions/upload-artifact from 4.3.5 to 4.3.6 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1060
- fix(cloud): respect user-set BENTOML_HOME by @aarnphm in https://github.com/bentoml/OpenLLM/pull/1067
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1066
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.9...v0.6.10
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.6.9
What's Changed
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1058
- chore(deps): bump actions/upload-artifact from 4.3.4 to 4.3.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1057
- chore(deps): bump openai from 1.37.1 to 1.38.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1056
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.7...v0.6.9
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.6.7
What's Changed
- feat(venv): support pip options by @rickzx in https://github.com/bentoml/OpenLLM/pull/1052
- chore: use uv instead of venv layers by @bojiang in https://github.com/bentoml/OpenLLM/pull/1054
New Contributors
- @rickzx made their first contribution in https://github.com/bentoml/OpenLLM/pull/1052
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.6...v0.6.7
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.6.6
What's Changed
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1043
- chore(deps): bump openai from 1.35.12 to 1.35.13 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1041
- chore(deps): bump softprops/action-gh-release from 2.0.6 to 2.0.8 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1046
- chore(deps): bump openai from 1.35.13 to 1.36.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1045
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1047
- chore(deps): bump openai from 1.36.1 to 1.37.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1048
- docs: Update OpenLLM readme by @Sherlock113 in https://github.com/bentoml/OpenLLM/pull/1051
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.5...v0.6.6
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.6.3
What's Changed
- chore: make the UI link clickable in output by @bojiang in https://github.com/bentoml/OpenLLM/pull/1038
New Contributors
- @bojiang made their first contribution in https://github.com/bentoml/OpenLLM/pull/1038
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.6.2...v0.6.3
- Python
Published by ssheng over 1 year ago
openllm - v0.6.0
We are thrilled to announce the release of OpenLLM 0.6, which marks a significant shift in our project's philosophy. This release introduces breaking changes to the codebase, reflecting our renewed focus on streamlining cloud deployment for LLMs.
In the previous releases, our goal was to provide users with the ability to fully customize their LLM deployment. However, we realized that the customization support in OpenLLM led to scope creep, deviating from our core focus on making LLM deployment simple. With the rise of open source LLMs and the growing emphasis on LLM-focused application development, we have decided to concentrate on what OpenLLM does best - simplifying LLM deployment.
We have completely revamped the architecture to make OpenLLM a tool that simplifies running LLMs as an API endpoint, prioritizing ease of use and performance. This means that 0.6 breaks away from many of the old Python APIs provided in 0.5, emphasizing itself as an easy-to-use CLI tool with cross-platform compatibility for users to deploy open source LLMs.
To learn more about the exciting features and capabilities of OpenLLM, visit our GitHub repository. We invite you to explore the new release, provide feedback, and join us in our mission to make cloud deployment of LLMs accessible and efficient for everyone.
Thank you for your continued support and trust in OpenLLM. We look forward to seeing the incredible applications you will build with the tool.
- Python
Published by ssheng over 1 year ago
openllm - v0.5.7
Installation
bash
pip install openllm==0.5.7
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.5.7
Usage
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.5.6...v0.5.7
- Python
Published by github-actions[bot] over 1 year ago
openllm - OpenLLM: v0.5
OpenLLM has undergone a significant upgrade in its v0.5 release to enhance compatibility with the BentoML 1.2 SDK. The CLI has also been streamlined to focus on delivering the most easy-to-use and reliable experience for deploying open-source LLMs to production. However, version 0.5 introduces breaking changes.
Breaking changes, and the reason why.
After releasing version 0.4, we realized that while OpenLLM offers a high degree of flexibility and power to users, they encountered numerous issues when attempting to deploy these models. OpenLLM had been trying to accomplish a lot by providing support for different backends (mainly PyTorch for CPU inference and vLLM for GPU inference) and accelerators. Although this provided users with the option to quickly test on their local machines, we discovered that this brought a lot of confusion when running OpenLLM locally versus the cloud. The difference between local and cloud deployment made it difficult for users to understand and control the packaged Bento to behave correctly on the cloud.
The motivation for 0.5 is to focus on cloud deployment. Cloud deployments often focus on high throughput and high concurrency serving, and GPU is the most common choice of hardware for achieving high throughput and concurrency serving. Therefore, we simplified backend support to just vLLM which is the most suitable and reliable for serving LLM on GPU on the cloud.
Architecture changes and SDK.
For version 0.5, we have decided to reduce the scope and support the backend that yields the most performance (in this case, vLLM). This means that pip install openllm will also depend on vLLM. In other words, we will currently pause our support for CPU going forward. All interactions with OpenLLM servers going forward should be done through clients (i.e., BentoML's Clients, OpenAI, etc.).
CLI
CLI has now been simplified to openllm start and openllm build
HuggingFace models
openllm start
openllm start will continue to accept HuggingFace model id for supported model architectures:
bash
openllm start microsoft/Phi-3-mini-4k-instruct --trust-remote-code
For any models that requires remote code execution, one should pass in
--trust-remote-code
openllm start will also accept serving from local path directly. Make sure to also pass in --trust-remote-code if you wish to use with openllm start
bash
openllm start path/to/custom-phi-instruct --trust-remote-code
openllm build
In previous versions, OpenLLM would copy the local cache of the models into the generated Bento store, resulting in having two copies of the models on users’ machine. From v0.5 going forward, models won't be packaged with the Bento and will be downloaded into Hugging Face cache first time on deployment.
```bash openllm build microsoft/Phi-3-mini-4k-instruct --trust-remote-code
Successfully built Bento 'microsoft--phi-3-mini-4k-instruct-service:5fa34190089f0ee40f9cce3cafc396b89b2e5e83'.
██████╗ ██████╗ ███████╗███╗ ██╗██╗ ██╗ ███╗ ███╗ ██╔═══██╗██╔══██╗██╔════╝████╗ ██║██║ ██║ ████╗ ████║ ██║ ██║██████╔╝█████╗ ██╔██╗ ██║██║ ██║ ██╔████╔██║ ██║ ██║██╔═══╝ ██╔══╝ ██║╚██╗██║██║ ██║ ██║╚██╔╝██║ ╚██████╔╝██║ ███████╗██║ ╚████║███████╗███████╗██║ ╚═╝ ██║ ╚═════╝ ╚═╝ ╚══════╝╚═╝ ╚═══╝╚══════╝╚══════╝╚═╝ ╚═╝.
📖 Next steps: ☁️ Deploy to BentoCloud: $ bentoml deploy microsoft--phi-3-mini-4k-instruct-service:5fa34190089f0ee40f9cce3cafc396b89b2e5e83 -n ${DEPLOYMENTNAME} ☁️ Update existing deployment on BentoCloud: $ bentoml deployment update --bento microsoft--phi-3-mini-4k-instruct-service:5fa34190089f0ee40f9cce3cafc396b89b2e5e83 ${DEPLOYMENTNAME} 🐳 Containerize BentoLLM: $ bentoml containerize microsoft--phi-3-mini-4k-instruct-service:5fa34190089f0ee40f9cce3cafc396b89b2e5e83 --opt progress=plain ```
For quantized models, make sure to also pass in the --quantize flag during build
bash
openllm build casperhansen/llama-3-70b-instruct-awq --quantize awq
See openllm build --help for more information
Private models
openllm start
For private models, we recommend users to save it to BentoML’s Model store first before using openllm start:
python
with bentoml.models.create(name="my-private-models") as model:
PrivateTrainedModel.save_pretrained(model.path)
MyTokenizer.save_pretrained(model.path)
Note: Make sure to also save your tokenizer in this bentomodel
You can then pass in the private model name directly to openllm start
bash
openllm start my-private-models
openllm build
Similar to openllm start, openllm build will only accept private models from BentoML’s model store:
bash
openllm build my-private-models
What's next?
Currently, OpenAI's compatibility will only have the /chat/completions and /models endpoints supported. We will continue bringing /completions as well as function calling support soon, so stay tuned.
Thank you for your continued support and trust in us. We would love to hear more of your feedback on the releases.
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.5.5
Installation
bash
pip install openllm==0.5.5
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.5.5
Usage
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- feat(models): command-r by @aarnphm in https://github.com/bentoml/OpenLLM/pull/1005
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1007
- chore(deps): bump taiki-e/install-action from 2.33.34 to 2.34.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1006
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.5.4...v0.5.5
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.5.4
Installation
bash
pip install openllm==0.5.4
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.5.4
Usage
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- feat(API): add light support for batch inference by @aarnphm in https://github.com/bentoml/OpenLLM/pull/1004
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.5.3...v0.5.4
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.5.3
Installation
bash
pip install openllm==0.5.3
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.5.3
Usage
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.5.2...v0.5.3
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.5.2
Installation
bash
pip install openllm==0.5.2
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.5.2
Usage
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.5.1...v0.5.2
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.5.0
Installation
bash
pip install openllm==0.5.1
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.5.1
Usage
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.5.0...v0.5.1
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.5.0
Installation
bash
pip install openllm==0.5.0
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.5.0
Usage
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/870
- chore(deps): bump taiki-e/install-action from 2.25.9 to 2.26.18 by @dependabot in https://github.com/bentoml/OpenLLM/pull/899
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/909
- chore(deps): bump github/codeql-action from 3.23.1 to 3.24.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/908
- chore(deps): bump sigstore/cosign-installer from 3.3.0 to 3.4.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/907
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/931
- feat: 1.2 APIs by @aarnphm in https://github.com/bentoml/OpenLLM/pull/821
- chore(deps): bump taiki-e/install-action from 2.26.18 to 2.27.9 by @dependabot in https://github.com/bentoml/OpenLLM/pull/920
- chore(deps): bump next from 13.4.8 to 13.5.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/912
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/935
- chore(deps): bump marocchino/sticky-pull-request-comment from 2.8.0 to 2.9.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/933
- chore(deps): bump aquasecurity/trivy-action from 0.16.1 to 0.18.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/932
- chore(deps): bump docker/setup-buildx-action from 3.0.0 to 3.2.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/941
- chore(deps): bump github/codeql-action from 3.24.3 to 3.24.9 by @dependabot in https://github.com/bentoml/OpenLLM/pull/939
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/942
- fix(compat): use annotated type from
typing_compatby @rudeigerc in https://github.com/bentoml/OpenLLM/pull/943 - docs: Update high-level messaging by @Sherlock113 in https://github.com/bentoml/OpenLLM/pull/949
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/947
- chore(deps): bump aquasecurity/trivy-action from 0.18.0 to 0.19.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/946
- chore(deps): bump taiki-e/install-action from 2.27.9 to 2.32.9 by @dependabot in https://github.com/bentoml/OpenLLM/pull/945
- Update README.md by @parano in https://github.com/bentoml/OpenLLM/pull/964
- chore(deps): bump taiki-e/install-action from 2.32.9 to 2.33.9 by @dependabot in https://github.com/bentoml/OpenLLM/pull/970
- chore(deps): bump sigstore/cosign-installer from 3.4.0 to 3.5.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/954
- chore(deps): bump docker/metadata-action from 5.5.0 to 5.5.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/956
- chore(deps): bump actions/setup-python from 5.0.0 to 5.1.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/955
- chore(deps): bump pypa/gh-action-pypi-publish from 1.8.11 to 1.8.14 by @dependabot in https://github.com/bentoml/OpenLLM/pull/958
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/959
- fix: update correct CompletionOutput object by @aarnphm in https://github.com/bentoml/OpenLLM/pull/973
- chore(deps): bump docker/build-push-action from 5.1.0 to 5.3.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/979
- chore(deps): bump docker/login-action from 3.0.0 to 3.1.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/978
- chore(deps): bump github/codeql-action from 3.24.9 to 3.25.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/977
- chore(deps): bump docker/setup-buildx-action from 3.2.0 to 3.3.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/975
- fix: make sure to respect additional parameters parse by @aarnphm in https://github.com/bentoml/OpenLLM/pull/981
- chore(deps): bump peter-evans/create-pull-request from 5.0.2 to 6.0.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/976
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/980
- chore(deps): bump rlespinasse/github-slug-action from 4.4.1 to 4.5.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/988
- chore(deps): bump softprops/action-gh-release from 1 to 2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/987
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/989
- chore(deps): bump taiki-e/install-action from 2.33.9 to 2.33.22 by @dependabot in https://github.com/bentoml/OpenLLM/pull/985
- chore(deps): bump actions/checkout from 4.1.1 to 4.1.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/984
- chore(deps): bump next from 13.4.8 to 14.1.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/983
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/994
- chore(deps): bump actions/checkout from 4.1.5 to 4.1.6 by @dependabot in https://github.com/bentoml/OpenLLM/pull/993
- chore(deps): bump github/codeql-action from 3.25.3 to 3.25.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/992
- chore(deps): bump aquasecurity/trivy-action from 0.19.0 to 0.20.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/991
- fix(docs): update correct BentoML links by @dennisrall in https://github.com/bentoml/OpenLLM/pull/995
- tests: add additional basic testing by @aarnphm in https://github.com/bentoml/OpenLLM/pull/982
- infra: prepare 0.5 releases by @aarnphm in https://github.com/bentoml/OpenLLM/pull/996
- chore(deps): bump actions/upload-artifact from 3.1.3 to 4.3.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/986
- chore(deps): bump actions/download-artifact from 3.0.2 to 4.1.7 by @dependabot in https://github.com/bentoml/OpenLLM/pull/990
- chore(qol): update CLI options and performance upgrade for build cache by @aarnphm in https://github.com/bentoml/OpenLLM/pull/997
- feat(ci): running CI on paperspace by @aarnphm in https://github.com/bentoml/OpenLLM/pull/998
- chore(deps): bump taiki-e/install-action from 2.33.22 to 2.33.34 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1000
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/1002
New Contributors
- @rudeigerc made their first contribution in https://github.com/bentoml/OpenLLM/pull/943
- @dennisrall made their first contribution in https://github.com/bentoml/OpenLLM/pull/995
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.44...v0.5.0
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.5.0-alpha.15
Installation
bash
pip install openllm==0.5.0-alpha.15
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.5.0-alpha.15
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.5.0-alpha.15 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- chore(deps): bump docker/setup-buildx-action from 3.0.0 to 3.2.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/941
- chore(deps): bump github/codeql-action from 3.24.3 to 3.24.9 by @dependabot in https://github.com/bentoml/OpenLLM/pull/939
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/942
- fix(compat): use annotated type from
typing_compatby @rudeigerc in https://github.com/bentoml/OpenLLM/pull/943 - docs: Update high-level messaging by @Sherlock113 in https://github.com/bentoml/OpenLLM/pull/949
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/947
- chore(deps): bump aquasecurity/trivy-action from 0.18.0 to 0.19.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/946
- chore(deps): bump taiki-e/install-action from 2.27.9 to 2.32.9 by @dependabot in https://github.com/bentoml/OpenLLM/pull/945
- Update README.md by @parano in https://github.com/bentoml/OpenLLM/pull/964
- chore(deps): bump taiki-e/install-action from 2.32.9 to 2.33.9 by @dependabot in https://github.com/bentoml/OpenLLM/pull/970
- chore(deps): bump sigstore/cosign-installer from 3.4.0 to 3.5.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/954
- chore(deps): bump docker/metadata-action from 5.5.0 to 5.5.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/956
- chore(deps): bump actions/setup-python from 5.0.0 to 5.1.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/955
- chore(deps): bump pypa/gh-action-pypi-publish from 1.8.11 to 1.8.14 by @dependabot in https://github.com/bentoml/OpenLLM/pull/958
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/959
- fix: update correct CompletionOutput object by @aarnphm in https://github.com/bentoml/OpenLLM/pull/973
- chore(deps): bump docker/build-push-action from 5.1.0 to 5.3.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/979
- chore(deps): bump docker/login-action from 3.0.0 to 3.1.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/978
- chore(deps): bump github/codeql-action from 3.24.9 to 3.25.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/977
- chore(deps): bump docker/setup-buildx-action from 3.2.0 to 3.3.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/975
- fix: make sure to respect additional parameters parse by @aarnphm in https://github.com/bentoml/OpenLLM/pull/981
- chore(deps): bump peter-evans/create-pull-request from 5.0.2 to 6.0.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/976
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/980
- chore(deps): bump rlespinasse/github-slug-action from 4.4.1 to 4.5.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/988
- chore(deps): bump softprops/action-gh-release from 1 to 2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/987
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/989
- chore(deps): bump taiki-e/install-action from 2.33.9 to 2.33.22 by @dependabot in https://github.com/bentoml/OpenLLM/pull/985
- chore(deps): bump actions/checkout from 4.1.1 to 4.1.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/984
- chore(deps): bump next from 13.4.8 to 14.1.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/983
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/994
- chore(deps): bump actions/checkout from 4.1.5 to 4.1.6 by @dependabot in https://github.com/bentoml/OpenLLM/pull/993
- chore(deps): bump github/codeql-action from 3.25.3 to 3.25.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/992
- chore(deps): bump aquasecurity/trivy-action from 0.19.0 to 0.20.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/991
- fix(docs): update correct BentoML links by @dennisrall in https://github.com/bentoml/OpenLLM/pull/995
- tests: add additional basic testing by @aarnphm in https://github.com/bentoml/OpenLLM/pull/982
- infra: prepare 0.5 releases by @aarnphm in https://github.com/bentoml/OpenLLM/pull/996
- chore(deps): bump actions/upload-artifact from 3.1.3 to 4.3.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/986
- chore(deps): bump actions/download-artifact from 3.0.2 to 4.1.7 by @dependabot in https://github.com/bentoml/OpenLLM/pull/990
- chore(qol): update CLI options and performance upgrade for build cache by @aarnphm in https://github.com/bentoml/OpenLLM/pull/997
- feat(ci): running CI on paperspace by @aarnphm in https://github.com/bentoml/OpenLLM/pull/998
- chore(deps): bump taiki-e/install-action from 2.33.22 to 2.33.34 by @dependabot in https://github.com/bentoml/OpenLLM/pull/1000
New Contributors
- @rudeigerc made their first contribution in https://github.com/bentoml/OpenLLM/pull/943
- @dennisrall made their first contribution in https://github.com/bentoml/OpenLLM/pull/995
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.5.0-alpha.1...v0.5.0-alpha.15
- Python
Published by github-actions[bot] over 1 year ago
openllm - v0.4.44
Installation
bash
pip install openllm==0.4.44
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.44
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.44 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- fix: remove vllm dependency for pytorch bento by @larme in https://github.com/bentoml/OpenLLM/pull/893
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.43...v0.4.44
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.43
Installation
bash
pip install openllm==0.4.43
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.43
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.43 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- fix: limit BentoML version range by @larme in https://github.com/bentoml/OpenLLM/pull/881
- chore: bump up bentoml version to 1.1.11 by @larme in https://github.com/bentoml/OpenLLM/pull/883
- Bump BentoML version in tools by @larme in https://github.com/bentoml/OpenLLM/pull/884
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.42...v0.4.43
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.42
Installation
bash
pip install openllm==0.4.42
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.42
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.42 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- docs: Update opt example to ms-phi by @Sherlock113 in https://github.com/bentoml/OpenLLM/pull/805
- chore(script): run vendored scripts by @aarnphm in https://github.com/bentoml/OpenLLM/pull/808
- docs: README.md typo by @weibeu in https://github.com/bentoml/OpenLLM/pull/819
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/818
- chore(deps): bump docker/metadata-action from 5.3.0 to 5.4.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/814
- chore(deps): bump taiki-e/install-action from 2.22.5 to 2.23.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/813
- chore(deps): bump github/codeql-action from 3.22.11 to 3.22.12 by @dependabot in https://github.com/bentoml/OpenLLM/pull/815
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/825
- chore(deps): bump crazy-max/ghaction-import-gpg from 6.0.0 to 6.1.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/824
- chore(deps): bump taiki-e/install-action from 2.23.1 to 2.23.7 by @dependabot in https://github.com/bentoml/OpenLLM/pull/823
- docs: Add Llamaindex in freedom to build by @Sherlock113 in https://github.com/bentoml/OpenLLM/pull/826
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/836
- chore(deps): bump docker/metadata-action from 5.4.0 to 5.5.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/834
- chore(deps): bump aquasecurity/trivy-action from 0.16.0 to 0.16.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/832
- chore(deps): bump taiki-e/install-action from 2.23.7 to 2.24.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/833
- chore(deps): bump vllm to 0.2.7 by @aarnphm in https://github.com/bentoml/OpenLLM/pull/837
- chore: update discord link by @aarnphm in https://github.com/bentoml/OpenLLM/pull/838
- improv(package): use python slim base image and let pytorch install cuda by @larme in https://github.com/bentoml/OpenLLM/pull/807
- fix(dockerfile): conflict deps by @aarnphm in https://github.com/bentoml/OpenLLM/pull/841
- chore: fix typo in list_models pydoc by @fuzzie360 in https://github.com/bentoml/OpenLLM/pull/847
- docs: update README.md telemetry code link by @fuzzie360 in https://github.com/bentoml/OpenLLM/pull/842
- chore(deps): bump taiki-e/install-action from 2.24.1 to 2.25.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/846
- chore(deps): bump github/codeql-action from 3.22.12 to 3.23.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/844
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/848
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/858
- chore(deps): bump taiki-e/install-action from 2.25.1 to 2.25.9 by @dependabot in https://github.com/bentoml/OpenLLM/pull/856
- chore(deps): bump github/codeql-action from 3.23.0 to 3.23.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/855
- fix: proper SSE handling for vllm by @larme in https://github.com/bentoml/OpenLLM/pull/877
- chore: set stop to empty list by default by @larme in https://github.com/bentoml/OpenLLM/pull/878
- fix: all runners sse output by @larme in https://github.com/bentoml/OpenLLM/pull/880
New Contributors
- @weibeu made their first contribution in https://github.com/bentoml/OpenLLM/pull/819
- @fuzzie360 made their first contribution in https://github.com/bentoml/OpenLLM/pull/847
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.41...v0.4.42
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.41
GPTQ Supports
vLLM backend now support GPTQ with upstream
python
openlml start TheBloke/Mistral-7B-Instruct-v0.2-GPTQ --backend vllm --quantise gptq
Installation
bash
pip install openllm==0.4.41
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.41
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.41 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- docs: add notes about dtypes usage. by @aarnphm in https://github.com/bentoml/OpenLLM/pull/786
- chore(deps): bump taiki-e/install-action from 2.22.0 to 2.22.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/790
- chore(deps): bump github/codeql-action from 2.22.9 to 3.22.11 by @dependabot in https://github.com/bentoml/OpenLLM/pull/794
- chore(deps): bump sigstore/cosign-installer from 3.2.0 to 3.3.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/793
- chore(deps): bump actions/download-artifact from 3.0.2 to 4.0.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/791
- chore(deps): bump actions/upload-artifact from 3.1.3 to 4.0.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/792
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/796
- fix(cli): avoid runtime
__origin__check for older Python by @aarnphm in https://github.com/bentoml/OpenLLM/pull/798 - feat(vllm): support GPTQ with 0.2.6 by @aarnphm in https://github.com/bentoml/OpenLLM/pull/797
- fix(ci): lock to v3 iteration of
actions/artifactsworkflow by @aarnphm in https://github.com/bentoml/OpenLLM/pull/799
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.40...v0.4.41
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.40
Installation
bash
pip install openllm==0.4.40
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.40
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.40 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- fix(infra): conform ruff to 150 LL by @aarnphm in https://github.com/bentoml/OpenLLM/pull/781
- infra: update blame ignore to formatter hash by @aarnphm in https://github.com/bentoml/OpenLLM/pull/782
- perf: upgrade mixtral to use expert parallelism by @aarnphm in https://github.com/bentoml/OpenLLM/pull/783
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.39...v0.4.40
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.39
Installation
bash
pip install openllm==0.4.39
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.39
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.39 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- fix(logprobs): correct check logprobs by @aarnphm in https://github.com/bentoml/OpenLLM/pull/779
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.38...v0.4.39
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.38
Installation
bash
pip install openllm==0.4.38
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.38
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.38 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- fix(mixtral): correct chat templates to remove additional spacing by @aarnphm in https://github.com/bentoml/OpenLLM/pull/774
- fix(cli): correct set arguments for
openllm importandopenllm buildby @aarnphm in https://github.com/bentoml/OpenLLM/pull/775 - fix(mixtral): setup hack atm to load weights from pt specifically instead of safetensors by @aarnphm in https://github.com/bentoml/OpenLLM/pull/776
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.37...v0.4.38
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.37
Installation
bash
pip install openllm==0.4.37
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.37
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.37 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- feat(mixtral): correct support for mixtral by @aarnphm in https://github.com/bentoml/OpenLLM/pull/772
- chore: running all script when installation by @aarnphm in https://github.com/bentoml/OpenLLM/pull/773
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.36...v0.4.37
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.36
Mixtral supports
Supports Mixtral on BentoCloud with vLLM and all required dependencies.
Bento built with openllm now defaults to python 3.11 for this change to work.
Installation
bash
pip install openllm==0.4.36
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.36
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.36 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- feat(openai): supports echo by @aarnphm in https://github.com/bentoml/OpenLLM/pull/760
- fix(openai): logprobs when echo is enabled by @aarnphm in https://github.com/bentoml/OpenLLM/pull/761
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/767
- chore(deps): bump docker/metadata-action from 5.2.0 to 5.3.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/766
- chore(deps): bump actions/setup-python from 4.7.1 to 5.0.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/765
- chore(deps): bump taiki-e/install-action from 2.21.26 to 2.22.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/764
- chore(deps): bump aquasecurity/trivy-action from 0.14.0 to 0.16.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/763
- chore(deps): bump github/codeql-action from 2.22.8 to 2.22.9 by @dependabot in https://github.com/bentoml/OpenLLM/pull/762
- feat: mixtral support by @aarnphm in https://github.com/bentoml/OpenLLM/pull/770
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.35...v0.4.36
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.35
Installation
bash
pip install openllm==0.4.35
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.35
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.35 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- chore(deps): bump pypa/gh-action-pypi-publish from 1.8.10 to 1.8.11 by @dependabot in https://github.com/bentoml/OpenLLM/pull/749
- chore(deps): bump docker/metadata-action from 5.0.0 to 5.2.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/751
- chore(deps): bump taiki-e/install-action from 2.21.19 to 2.21.26 by @dependabot in https://github.com/bentoml/OpenLLM/pull/750
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/753
- fix(logprobs): explicitly set logprobs=None by @aarnphm in https://github.com/bentoml/OpenLLM/pull/757
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.34...v0.4.35
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.34
Installation
bash
pip install openllm==0.4.34
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.34
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.34 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- feat(models): Support qwen by @yansheng105 in https://github.com/bentoml/OpenLLM/pull/742
New Contributors
- @yansheng105 made their first contribution in https://github.com/bentoml/OpenLLM/pull/742
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.33...v0.4.34
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.33
Installation
bash
pip install openllm==0.4.33
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.33
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.33 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.32...v0.4.33
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.32
Installation
bash
pip install openllm==0.4.32
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.32
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.32 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- chore(deps): bump taiki-e/install-action from 2.21.17 to 2.21.19 by @dependabot in https://github.com/bentoml/OpenLLM/pull/735
- chore(deps): bump github/codeql-action from 2.22.7 to 2.22.8 by @dependabot in https://github.com/bentoml/OpenLLM/pull/734
- chore: revert back previous backend support PyTorch by @aarnphm in https://github.com/bentoml/OpenLLM/pull/739
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.31...v0.4.32
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.31
Installation
bash
pip install openllm==0.4.31
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.31
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.31 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- fix(docs): remove invalid options by @aarnphm in https://github.com/bentoml/OpenLLM/pull/733
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.30...v0.4.31
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.30
Installation
bash
pip install openllm==0.4.30
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.30
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.30 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.29...v0.4.30
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.29
Installation
bash
pip install openllm==0.4.29
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.29
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.29 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.28...v0.4.29
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.28
Installation
bash
pip install openllm==0.4.28
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.28
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.28 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- fix(baichuan): supported from baichuan 2 from now on. by @MingLiangDai in https://github.com/bentoml/OpenLLM/pull/728
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.27...v0.4.28
- Python
Published by github-actions[bot] about 2 years ago
openllm - v0.4.26
Installation
bash
pip install openllm==0.4.26
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.26
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.26 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- fix(infra): setup higher timer for building container images by @aarnphm in https://github.com/bentoml/OpenLLM/pull/723
- fix(client): correct schemas parser from correct response output by @aarnphm in https://github.com/bentoml/OpenLLM/pull/724
- feat(openai): chat templates and complete control of prompt generation by @aarnphm in https://github.com/bentoml/OpenLLM/pull/725
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.25...v0.4.26
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.25
Installation
bash
pip install openllm==0.4.25
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.25
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.25 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- fix(openai): correct stop tokens and finish_reason state by @aarnphm in https://github.com/bentoml/OpenLLM/pull/722
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.24...v0.4.25
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.24
Installation
bash
pip install openllm==0.4.24
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.24
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.24 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.23...v0.4.24
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.23
Installation
bash
pip install openllm==0.4.23
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.23
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.23 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- chore: cleanup unused prompt templates by @aarnphm in https://github.com/bentoml/OpenLLM/pull/713
- feat(generation): add support for eostokenid by @aarnphm in https://github.com/bentoml/OpenLLM/pull/714
- fix(ci): tests by @aarnphm in https://github.com/bentoml/OpenLLM/pull/715
- refactor: delete unused code by @aarnphm in https://github.com/bentoml/OpenLLM/pull/716
- chore(logger): fix logger and streamline style by @aarnphm in https://github.com/bentoml/OpenLLM/pull/717
- chore(strategy): compact and add stubs by @aarnphm in https://github.com/bentoml/OpenLLM/pull/718
- chore(types): append additional types change by @aarnphm in https://github.com/bentoml/OpenLLM/pull/719
- fix(base-image): update base image to include cuda for now by @aarnphm in https://github.com/bentoml/OpenLLM/pull/720
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.22...v0.4.23
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.22
Installation
bash
pip install openllm==0.4.22
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.22
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.22 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- refactor: update runner helpers and add maxmodellen by @aarnphm in https://github.com/bentoml/OpenLLM/pull/712
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.21...v0.4.22
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.21
Installation
bash
pip install openllm==0.4.21
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.21
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.21 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/711
- chore(deps): bump taiki-e/install-action from 2.21.11 to 2.21.17 by @dependabot in https://github.com/bentoml/OpenLLM/pull/709
- chore(deps): bump docker/build-push-action from 5.0.0 to 5.1.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/708
- chore(deps): bump github/codeql-action from 2.22.5 to 2.22.7 by @dependabot in https://github.com/bentoml/OpenLLM/pull/707
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.20...v0.4.21
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.20
Installation
bash
pip install openllm==0.4.20
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.20
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.20 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.19...v0.4.20
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.19
Installation
bash
pip install openllm==0.4.19
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.19
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.19 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.18...v0.4.19
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.18
Installation
bash
pip install openllm==0.4.18
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.18
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.18 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- chore: update lower bound version of bentoml to avoid breakage by @aarnphm in https://github.com/bentoml/OpenLLM/pull/703
- feat(openai): dynamic model_type registration by @aarnphm in https://github.com/bentoml/OpenLLM/pull/704
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.17...v0.4.18
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.17
Installation
bash
pip install openllm==0.4.17
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.17
Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.17 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the CHANGELOG.md
What's Changed
- infra: update generate notes and better local handle by @aarnphm in https://github.com/bentoml/OpenLLM/pull/701
- fix(backend): correct use variable for backend when initialisation by @aarnphm in https://github.com/bentoml/OpenLLM/pull/702
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.16...v0.4.17
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.16
Installation
bash
pip install openllm==0.4.16
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.16
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.16 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.16
Find more information about this release in the CHANGELOG.md
What's Changed
- feat(ctranslate): initial infrastructure support by @aarnphm in https://github.com/bentoml/OpenLLM/pull/694
- feat(vllm): bump to 0.2.2 by @aarnphm in https://github.com/bentoml/OpenLLM/pull/695
- feat(engine): CTranslate2 by @aarnphm in https://github.com/bentoml/OpenLLM/pull/698
- chore: update documentation about runtime by @aarnphm in https://github.com/bentoml/OpenLLM/pull/699
- chore: update changelog [skip ci] by @aarnphm in https://github.com/bentoml/OpenLLM/pull/700
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.15...v0.4.16
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.15
Installation
bash
pip install openllm==0.4.15
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.15
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.15 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.15
Find more information about this release in the CHANGELOG.md
What's Changed
- fix(cattrs): strictly lock <23.2 until we upgrade validation logic by @aarnphm in https://github.com/bentoml/OpenLLM/pull/690
- fix(annotations): check library through find_spec by @aarnphm in https://github.com/bentoml/OpenLLM/pull/691
- feat: heuristics logprobs by @aarnphm in https://github.com/bentoml/OpenLLM/pull/692
- chore: update documentation by @aarnphm in https://github.com/bentoml/OpenLLM/pull/693
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.14...v0.4.15
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.14
Installation
bash
pip install openllm==0.4.14
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.14
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.14 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.14
Find more information about this release in the CHANGELOG.md
What's Changed
- fix(dependencies): ignore broken cattrs release by @aarnphm in https://github.com/bentoml/OpenLLM/pull/689
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.13...v0.4.14
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.13
Installation
bash
pip install openllm==0.4.13
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.13
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.13 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.13
Find more information about this release in the CHANGELOG.md
What's Changed
- fix(llm): remove unnecessary check by @aarnphm in https://github.com/bentoml/OpenLLM/pull/683
- examples: improve instructions and cleanup simple API server by @aarnphm in https://github.com/bentoml/OpenLLM/pull/684
- fix(build): lock lower version based on each release and update infra by @aarnphm in https://github.com/bentoml/OpenLLM/pull/686
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.12...v0.4.13
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.12
Installation
bash
pip install openllm==0.4.12
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.12
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.12 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.12
Find more information about this release in the CHANGELOG.md
What's Changed
- fix(envvar): explicitly set NVIDIADRIVERCAPABILITIES by @aarnphm in https://github.com/bentoml/OpenLLM/pull/681
- fix(torch_dtype): correctly infer based on options by @aarnphm in https://github.com/bentoml/OpenLLM/pull/682
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.11...v0.4.12
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.11
Installation
bash
pip install openllm==0.4.11
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.11
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.11 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.11
Find more information about this release in the CHANGELOG.md
What's Changed
- infra: update cbfmt options by @aarnphm in https://github.com/bentoml/OpenLLM/pull/676
- fix(examples): add support for streaming feature by @aarnphm in https://github.com/bentoml/OpenLLM/pull/677
- fix: correct set item for attrs >23.1 by @aarnphm in https://github.com/bentoml/OpenLLM/pull/678
- fix(build): correctly parse default env for container by @aarnphm in https://github.com/bentoml/OpenLLM/pull/679
- fix(env): correct format environment on docker by @aarnphm in https://github.com/bentoml/OpenLLM/pull/680
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.10...v0.4.11
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.10
Installation
bash
pip install openllm==0.4.10
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.10
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.10 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.10
Find more information about this release in the CHANGELOG.md
What's Changed
- fix(runner): remove keyword args for attrs.get() by @jeffwang0516 in https://github.com/bentoml/OpenLLM/pull/661
- fix: update notebook by @xianml in https://github.com/bentoml/OpenLLM/pull/662
- feat(type): provide structured annotations stubs by @aarnphm in https://github.com/bentoml/OpenLLM/pull/663
- feat(llm): respect warnings environment for dtype warning by @aarnphm in https://github.com/bentoml/OpenLLM/pull/664
- infra: makes huggingface-hub requirements on fine-tune by @aarnphm in https://github.com/bentoml/OpenLLM/pull/665
- types: update stubs for remaining entrypoints by @aarnphm in https://github.com/bentoml/OpenLLM/pull/667
- perf: reduce footprint by @aarnphm in https://github.com/bentoml/OpenLLM/pull/668
- perf(build): locking and improve build speed by @aarnphm in https://github.com/bentoml/OpenLLM/pull/669
- docs: add LlamaIndex integration by @aarnphm in https://github.com/bentoml/OpenLLM/pull/646
- infra: remove codegolf by @aarnphm in https://github.com/bentoml/OpenLLM/pull/671
- feat(models): Phi 1.5 by @aarnphm in https://github.com/bentoml/OpenLLM/pull/672
- fix(docs): chatglm support on vLLM by @aarnphm in https://github.com/bentoml/OpenLLM/pull/673
- chore(loading): include verbose warning about trustremotecode by @aarnphm in https://github.com/bentoml/OpenLLM/pull/674
- perf: potentially reduce image size by @aarnphm in https://github.com/bentoml/OpenLLM/pull/675
New Contributors
- @jeffwang0516 made their first contribution in https://github.com/bentoml/OpenLLM/pull/661
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.9...v0.4.10
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.9
Installation
bash
pip install openllm==0.4.9
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.9
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.9 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.9
Find more information about this release in the CHANGELOG.md
What's Changed
- infra: update scripts to run update readme automatically by @aarnphm in https://github.com/bentoml/OpenLLM/pull/658
- chore: update requirements in README.md by @aarnphm in https://github.com/bentoml/OpenLLM/pull/659
- fix(falcon): remove early_stopping default arguments by @aarnphm in https://github.com/bentoml/OpenLLM/pull/660
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.8...v0.4.9
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.8
Installation
bash
pip install openllm==0.4.8
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.8
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.8 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.8
Find more information about this release in the CHANGELOG.md
What's Changed
- docs: update instruction adding new models and remove command docstring by @aarnphm in https://github.com/bentoml/OpenLLM/pull/654
- chore(cli): move playground to CLI components by @aarnphm in https://github.com/bentoml/OpenLLM/pull/655
- perf: improve build logics and cleanup speed by @aarnphm in https://github.com/bentoml/OpenLLM/pull/657
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.7...v0.4.8
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.7
Installation
bash
pip install openllm==0.4.7
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.7
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.7 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.7
Find more information about this release in the CHANGELOG.md
What's Changed
- refactor: use DEBUG env-var instead of OPENLLMDEVDEBUG by @aarnphm in https://github.com/bentoml/OpenLLM/pull/647
- fix(cli): update context name parsing correctly by @aarnphm in https://github.com/bentoml/OpenLLM/pull/652
- feat: Yi models by @aarnphm in https://github.com/bentoml/OpenLLM/pull/651
- fix: correct OPENLLMDEVBUILD check by @xianml in https://github.com/bentoml/OpenLLM/pull/653
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.6...v0.4.7
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.6
Installation
bash
pip install openllm==0.4.6
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.6
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.6 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.6
Find more information about this release in the CHANGELOG.md
What's Changed
- chore: cleanup unused code path by @aarnphm in https://github.com/bentoml/OpenLLM/pull/633
- perf(model): update mistral inference parameters and prompt format by @larme in https://github.com/bentoml/OpenLLM/pull/632
- infra: remove unused postprocess_generate by @aarnphm in https://github.com/bentoml/OpenLLM/pull/634
- docs: update README.md by @aarnphm in https://github.com/bentoml/OpenLLM/pull/635
- fix(client): correct destructor the httpx object boht sync and async by @aarnphm in https://github.com/bentoml/OpenLLM/pull/636
- doc: update adding new model guide by @larme in https://github.com/bentoml/OpenLLM/pull/637
- fix(generation): compatibility dtype with CPU by @aarnphm in https://github.com/bentoml/OpenLLM/pull/638
- fix(cpu): more verbose definition for dtype casting by @aarnphm in https://github.com/bentoml/OpenLLM/pull/639
- fix(service): to yield out correct JSON objects by @aarnphm in https://github.com/bentoml/OpenLLM/pull/640
- fix(cli): set default dtype to auto infer by @aarnphm in https://github.com/bentoml/OpenLLM/pull/642
- fix(dependencies): lock build < 1 for now by @aarnphm in https://github.com/bentoml/OpenLLM/pull/643
- chore(openapi): unify inject param by @aarnphm in https://github.com/bentoml/OpenLLM/pull/645
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.5...v0.4.6
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.5
Installation
bash
pip install openllm==0.4.5
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.5
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.5 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.5
Find more information about this release in the CHANGELOG.md
What's Changed
- refactor(cli): move out to its own packages by @aarnphm in https://github.com/bentoml/OpenLLM/pull/619
- fix(cli): correct set working_dir by @aarnphm in https://github.com/bentoml/OpenLLM/pull/620
- chore(cli): always show available models by @aarnphm in https://github.com/bentoml/OpenLLM/pull/621
- fix(sdk): make sure build to quiet out stdout by @aarnphm in https://github.com/bentoml/OpenLLM/pull/622
- chore: update jupyter notebooks with new API by @aarnphm in https://github.com/bentoml/OpenLLM/pull/623
- fix(ruff): correct consistency between isort and formatter by @aarnphm in https://github.com/bentoml/OpenLLM/pull/624
- feat(vllm): support passing specific dtype by @aarnphm in https://github.com/bentoml/OpenLLM/pull/626
- chore(deps): bump taiki-e/install-action from 2.21.8 to 2.21.11 by @dependabot in https://github.com/bentoml/OpenLLM/pull/625
- feat(cli):
--dtypearguments by @aarnphm in https://github.com/bentoml/OpenLLM/pull/627 - fix(cli): make sure to pass the dtype to subprocess service by @aarnphm in https://github.com/bentoml/OpenLLM/pull/628
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/629
- infra: removing clojure frontend from infra cycle by @aarnphm in https://github.com/bentoml/OpenLLM/pull/630
- fix(torch_dtype): load eagerly by @aarnphm in https://github.com/bentoml/OpenLLM/pull/631
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.4...v0.4.5
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.4
Installation
bash
pip install openllm==0.4.4
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.4
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.4 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.4
Find more information about this release in the CHANGELOG.md
What's Changed
- chore: no need compat workaround for setting cell_contents by @aarnphm in https://github.com/bentoml/OpenLLM/pull/616
- chore(llm): expose quantise and lazy load heavy imports by @aarnphm in https://github.com/bentoml/OpenLLM/pull/617
- feat(llm): update warning envvar and add embedded mode by @aarnphm in https://github.com/bentoml/OpenLLM/pull/618
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.3...v0.4.4
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.3
Installation
bash
pip install openllm==0.4.3
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.3
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.3 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.3
Find more information about this release in the CHANGELOG.md
What's Changed
- feat(server): helpers endpoints for conversation format by @aarnphm in https://github.com/bentoml/OpenLLM/pull/613
- feat(client): support return response_cls to string by @aarnphm in https://github.com/bentoml/OpenLLM/pull/614
- feat(client): add helpers subclass by @aarnphm in https://github.com/bentoml/OpenLLM/pull/615
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.2...v0.4.3
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.2
Installation
bash
pip install openllm==0.4.2
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.2
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.2 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.2
Find more information about this release in the CHANGELOG.md
What's Changed
- refactor(cli): cleanup API by @aarnphm in https://github.com/bentoml/OpenLLM/pull/592
- infra: move out clojure to external by @aarnphm in https://github.com/bentoml/OpenLLM/pull/593
- infra: using ruff formatter by @aarnphm in https://github.com/bentoml/OpenLLM/pull/594
- infra: remove tsconfig by @aarnphm in https://github.com/bentoml/OpenLLM/pull/595
- revert: configuration not to dump flatten by @aarnphm in https://github.com/bentoml/OpenLLM/pull/597
- package: add openllm core dependencies to labels by @aarnphm in https://github.com/bentoml/OpenLLM/pull/600
- fix: loading correct local models by @aarnphm in https://github.com/bentoml/OpenLLM/pull/599
- fix: correct importmodules locally by @aarnphm in https://github.com/bentoml/OpenLLM/pull/601
- fix: overload flattened dict by @aarnphm in https://github.com/bentoml/OpenLLM/pull/602
- feat(client): support authentication token and shim implementation by @aarnphm in https://github.com/bentoml/OpenLLM/pull/605
- fix(client): check for should retry header by @aarnphm in https://github.com/bentoml/OpenLLM/pull/606
- chore(client): remove ununsed state enum by @aarnphm in https://github.com/bentoml/OpenLLM/pull/609
- chore: remove generated stubs for now by @aarnphm in https://github.com/bentoml/OpenLLM/pull/610
- refactor(config): simplify configuration and update start CLI output by @aarnphm in https://github.com/bentoml/OpenLLM/pull/611
- docs: update supported feature set by @aarnphm in https://github.com/bentoml/OpenLLM/pull/612
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.1...v0.4.2
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.1
OpenLLM version 0.4.0 introduces several enhanced features.
Unified API and Continuous Batching support: 0.4.0 brings a revamped API for OpenLLM. Users now can run LLM with two new APIs. Under async context, calls to both
llm.generate_iteratorandllm.generatenow supports continuous batching for the most optimal throughput.await llm.generate_iterator(prompt, stop, **kwargs): one shot generation for any given prompt```python import openllm, asyncio
llm = openllm.LLM("HuggingFaceH4/zephyr-7b-beta")
async def infer(prompt,*kwargs): return await llm.generate(prompt, *kwargs)
asyncio.run(infer("Time is a definition of"))
```
await llm.generate(prompt, stop, **kwargs: stream generation that returns tokens as they become ready```python import bentoml, openllm import openllm
llm = openllm.LLM("HuggingFaceH4/zephyr-7b-beta")
svc = bentoml.Service(name='zephyr-instruct', runners=[llm.runner])
@svc.api(input=bentoml.io.Text(), output=bentoml.io.Text(mediatype='text/event-stream')) async def prompt(inputtext: str) -> str: async for generation in llm.generateiterator(inputtext): yield f"data: {generation.outputs[0].text}\n\n"
```
Internally, the runner can be accessed only with
llm.runnerandllm.runner.generate_iterator.The backend is now automatically inferred based on the presence of
vllmin the environment. However, if you prefer to manually specify the backend, you can achieve this by usingopenllm.LLM("HuggingFaceH4/zephyr-7b-beta", backend='pt').Quantization can also be passed directly to this new
LLMAPI.python llm = openllm.LLM("TheBloke/Mistral-7B-Instruct-v0.1-AWQ", quantize='awq')
Mistral Model: OpenLLM now supports Mistral. To start a Mistral server, simply execute
openllm start mistral.AWQ and SqueezeLLM Quantization: AWQ and SqueezeLLM is now supported with vLLM backend. Simply pass
--quantize [awq|squezzellm]toopenllm startto use AWQ or SqueezeLLM quantization.IMPORTANT: For using AWQ it is crucial that the model weight is already quantized with AWQ. Please look for the model variant on HuggingFace hub for the AWQ version of the model you want to use. Currently, only AWQ with vLLM is fully tested and supported.
General bug fixes: fixed a bug with regards to tag generation. Standalone Bento that use this new API should just work as expected if the model is already exists in the model store.
- For consistency, make sure to run
openllm prune -y --include-bentos
- For consistency, make sure to run
Installation
bash
pip install openllm==0.4.1
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.1
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.1 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.1
Find more information about this release in the CHANGELOG.md
What's Changed
- chore(runner): yield the outputs directly by @aarnphm in https://github.com/bentoml/OpenLLM/pull/573
- chore(openai): simplify client examples by @aarnphm in https://github.com/bentoml/OpenLLM/pull/574
- fix(examples): correct dependencies in requirements.txt [skip ci] by @aarnphm in https://github.com/bentoml/OpenLLM/pull/575
- refactor: cleanup typing to expose correct API by @aarnphm in https://github.com/bentoml/OpenLLM/pull/576
- fix(stubs): update initialisation types by @aarnphm in https://github.com/bentoml/OpenLLM/pull/577
- refactor(strategies): move logics into openllm-python by @aarnphm in https://github.com/bentoml/OpenLLM/pull/578
- chore(service): cleanup API by @aarnphm in https://github.com/bentoml/OpenLLM/pull/579
- infra: disable npm updates and correct python packages by @aarnphm in https://github.com/bentoml/OpenLLM/pull/580
- chore(deps): bump aquasecurity/trivy-action from 0.13.1 to 0.14.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/583
- chore(deps): bump taiki-e/install-action from 2.21.7 to 2.21.8 by @dependabot in https://github.com/bentoml/OpenLLM/pull/581
- chore(deps): bump sigstore/cosign-installer from 3.1.2 to 3.2.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/582
- fix: device imports using strategies by @aarnphm in https://github.com/bentoml/OpenLLM/pull/584
- fix(gptq): update config fields by @aarnphm in https://github.com/bentoml/OpenLLM/pull/585
- fix: unbound variable for completion client by @aarnphm in https://github.com/bentoml/OpenLLM/pull/587
- fix(awq): correct awq detection for support by @aarnphm in https://github.com/bentoml/OpenLLM/pull/586
- feat(vllm): squeezellm by @aarnphm in https://github.com/bentoml/OpenLLM/pull/588
- docs: update quantization notes by @aarnphm in https://github.com/bentoml/OpenLLM/pull/589
- fix(cli): append model-id instruction to build by @aarnphm in https://github.com/bentoml/OpenLLM/pull/590
- container: update tracing dependencies by @aarnphm in https://github.com/bentoml/OpenLLM/pull/591
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.4.0...v0.4.1
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.4.0
Release Highlights
OpenLLM 0.4.0 brings a few revamp feature
Unified API
0.4.0 brings a revamped API for OpenLLM. Users now can run LLM with two new API
await llm.generate_iterator(prompt, stop, **kwargs)await llm.generate(prompt, stop, **kwargs
llm.generate is the one shot generation for any given prompt, whereas llm.generate_iterator is the streaming variant.
```python import openllm, asyncio
llm = openllm.LLM("HuggingFaceH4/zephyr-7b-beta")
async def infer(prompt,*kwargs): return await llm.generate(prompt, *kwargs)
asyncio.run(infer("Time is a definition of")) ```
For using within a BentoML Service, one can do the following
```python import bentoml, openllm import openllm
llm = openllm.LLM("HuggingFaceH4/zephyr-7b-beta")
svc = bentoml.Service(name='zephyr-instruct', runners=[llm.runner])
@svc.api(input=bentoml.io.Text(), output=bentoml.io.Text(mediatype='text/event-stream')) async def prompt(inputtext: str) -> str: async for generation in llm.generateiterator(inputtext): yield f"data: {generation.outputs[0].text}\n\n" ```
Mistral supports
Mistral is now supported with OpenLLM. Simply do openllm start mistral to start a mistral server
AWQ support
AWQ is not supported with both vLLM and PyTorch backend. Simply pass --quantize awq to use AWQ.
[!IMPORTANT] For using AWQ it is crucial that the model weight is already quantized with AWQ. Please look for the model variant on HuggingFace hub for the AWQ version of the model you want to use
General bug fixes
Fixes a bug with regards to tag generation. Standalone Bento that use this new API should just work as expected if the model is already exists in the model store.
For consistency, make sure to run openllm prune -y --include-bentos
Installation
bash
pip install openllm==0.4.0
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.4.0
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.4.0 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.4.0
Find more information about this release in the CHANGELOG.md
What's Changed
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/563
- chore(deps): bump aquasecurity/trivy-action from 0.13.0 to 0.13.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/562
- chore(deps): bump taiki-e/install-action from 2.21.3 to 2.21.7 by @dependabot in https://github.com/bentoml/OpenLLM/pull/561
- chore(deps-dev): bump eslint from 8.47.0 to 8.53.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/558
- chore(deps): bump @vercel/og from 0.5.18 to 0.5.20 by @dependabot in https://github.com/bentoml/OpenLLM/pull/556
- chore(deps-dev): bump @types/react from 18.2.20 to 18.2.35 by @dependabot in https://github.com/bentoml/OpenLLM/pull/559
- chore(deps-dev): bump @typescript-eslint/eslint-plugin from 6.9.0 to 6.10.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/564
- fix : updated client to toggle tls verification by @ABHISHEK03312 in https://github.com/bentoml/OpenLLM/pull/532
- perf: unify LLM interface by @aarnphm in https://github.com/bentoml/OpenLLM/pull/518
- fix(stop): stop is not available in config by @aarnphm in https://github.com/bentoml/OpenLLM/pull/566
- infra: update docs on serving fine-tuning layers by @aarnphm in https://github.com/bentoml/OpenLLM/pull/567
- fix: update build dependencies and format chat prompt by @aarnphm in https://github.com/bentoml/OpenLLM/pull/569
- chore(examples): update openai client by @aarnphm in https://github.com/bentoml/OpenLLM/pull/568
- fix(client): one-shot generation construction by @aarnphm in https://github.com/bentoml/OpenLLM/pull/570
- feat: Mistral support by @aarnphm in https://github.com/bentoml/OpenLLM/pull/571
New Contributors
- @ABHISHEK03312 made their first contribution in https://github.com/bentoml/OpenLLM/pull/532
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.3.14...v0.4.0
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.3.14
Installation
bash
pip install openllm==0.3.14
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.3.14
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.3.14 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.3.14
Find more information about this release in the CHANGELOG.md
What's Changed
- chore(deps): bump taiki-e/install-action from 2.20.15 to 2.21.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/546
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/548
- chore(deps): bump aquasecurity/trivy-action from 0.12.0 to 0.13.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/545
- chore(deps): bump github/codeql-action from 2.22.4 to 2.22.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/544
- fix: update llama2 notebook example by @xianml in https://github.com/bentoml/OpenLLM/pull/516
- chore(deps-dev): bump @types/react from 18.2.20 to 18.2.33 by @dependabot in https://github.com/bentoml/OpenLLM/pull/542
- chore(deps-dev): bump @typescript-eslint/eslint-plugin from 6.8.0 to 6.9.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/537
- chore(deps-dev): bump @edge-runtime/vm from 3.1.4 to 3.1.6 by @dependabot in https://github.com/bentoml/OpenLLM/pull/540
- chore(deps-dev): bump eslint from 8.47.0 to 8.52.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/541
- fix: Max new tokens by @XunchaoZ in https://github.com/bentoml/OpenLLM/pull/550
- chore(inference): update vllm to 0.2.1.post1 and update config parsing by @aarnphm in https://github.com/bentoml/OpenLLM/pull/554
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.3.13...v0.3.14
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.3.13
Installation
bash
pip install openllm==0.3.13
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.3.13
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.3.13 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.3.13
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.3.12...v0.3.13
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.3.12
Installation
bash
pip install openllm==0.3.12
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.3.12
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.3.12 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.3.12
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.3.11...v0.3.12
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.3.10
Installation
bash
pip install openllm==0.3.10
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.3.10
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.3.10 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.3.10
Find more information about this release in the CHANGELOG.md
What's Changed
- chore(deps-dev): bump @typescript-eslint/parser from 6.7.5 to 6.8.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/511
- chore(deps-dev): bump @next/eslint-plugin-next from 13.5.4 to 13.5.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/510
- chore(deps-dev): bump @typescript-eslint/eslint-plugin from 6.7.5 to 6.8.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/509
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/529
- chore(deps): bump actions/checkout from 4.1.0 to 4.1.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/528
- chore(deps): bump github/codeql-action from 2.22.3 to 2.22.4 by @dependabot in https://github.com/bentoml/OpenLLM/pull/527
- chore(deps): bump taiki-e/install-action from 2.20.3 to 2.20.15 by @dependabot in https://github.com/bentoml/OpenLLM/pull/526
- chore(deps-dev): bump eslint-plugin-import from 2.28.1 to 2.29.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/525
- chore(deps-dev): bump turbo from 1.10.15 to 1.10.16 by @dependabot in https://github.com/bentoml/OpenLLM/pull/521
- chore(deps-dev): bump @types/dedent from 0.7.0 to 0.7.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/524
- chore(falcon): Use official implementation by @aarnphm in https://github.com/bentoml/OpenLLM/pull/530
- chore(deps-dev): bump @types/node from 20.5.3 to 20.8.7 by @dependabot in https://github.com/bentoml/OpenLLM/pull/522
- feat: Conversation template by @XunchaoZ in https://github.com/bentoml/OpenLLM/pull/519
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.3.9...v0.3.10
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.3.9
Installation
bash
pip install openllm==0.3.9
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.3.9
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.3.9 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.3.9
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.3.8...v0.3.9
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.3.8
Installation
bash
pip install openllm==0.3.8
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.3.8
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.3.8 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.3.8
Find more information about this release in the CHANGELOG.md
What's Changed
- docs: Add OpenLLM tutorial Google Colab link by @Sherlock113 in https://github.com/bentoml/OpenLLM/pull/497
- chore(deps-dev): bump @types/react-dom from 18.0.6 to 18.2.13 by @dependabot in https://github.com/bentoml/OpenLLM/pull/492
- chore(deps-dev): bump @types/react from 18.2.20 to 18.2.28 by @dependabot in https://github.com/bentoml/OpenLLM/pull/493
- chore(deps-dev): bump typescript from 5.1.6 to 5.2.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/494
- chore(deps-dev): bump @types/node from 20.5.3 to 20.8.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/495
- fix(breaking): remove embeddings and update client implementation by @aarnphm in https://github.com/bentoml/OpenLLM/pull/500
- feat: openai.Model.list() by @XunchaoZ in https://github.com/bentoml/OpenLLM/pull/499
- chore(deps): bump taiki-e/install-action from 2.20.2 to 2.20.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/502
- chore(deps): bump github/codeql-action from 2.22.2 to 2.22.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/501
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/506
- fix: check for parity by @aarnphm in https://github.com/bentoml/OpenLLM/pull/508
- chore(deps-dev): bump prettier-plugin-tailwindcss from 0.5.5 to 0.5.6 by @dependabot in https://github.com/bentoml/OpenLLM/pull/504
- chore(deps-dev): bump @types/node from 20.5.3 to 20.8.6 by @dependabot in https://github.com/bentoml/OpenLLM/pull/503
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.3.7...v0.3.8
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.3.7
Installation
bash
pip install openllm==0.3.7
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.3.7
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.3.7 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.3.7
Find more information about this release in the CHANGELOG.md
What's Changed
- chore(deps): bump taiki-e/install-action from 2.18.9 to 2.18.13 by @dependabot in https://github.com/bentoml/OpenLLM/pull/372
- chore(deps): bump pypa/cibuildwheel from 2.15.0 to 2.16.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/376
- chore(deps): bump github/codeql-action from 2.21.5 to 2.21.8 by @dependabot in https://github.com/bentoml/OpenLLM/pull/378
- chore(deps): bump @mui/material from 5.14.8 to 5.14.9 by @dependabot in https://github.com/bentoml/OpenLLM/pull/368
- chore(deps-dev): bump @edge-runtime/vm from 3.1.0 to 3.1.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/370
- chore(deps): bump @floating-ui/dom from 1.5.1 to 1.5.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/366
- chore(deps-dev): bump @typescript-eslint/parser from 6.4.1 to 6.7.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/365
- chore(deps): bump @vercel/og from 0.5.11 to 0.5.13 by @dependabot in https://github.com/bentoml/OpenLLM/pull/356
- chore(deps-dev): bump postcss from 8.4.23 to 8.4.29 by @dependabot in https://github.com/bentoml/OpenLLM/pull/360
- chore(deps-dev): bump turbo from 1.10.13 to 1.10.14 by @dependabot in https://github.com/bentoml/OpenLLM/pull/364
- chore(bench): add more prompts by @aarnphm in https://github.com/bentoml/OpenLLM/pull/380
- chore(deps): bump @mui/styled-engine from 5.13.2 to 5.14.10 by @dependabot in https://github.com/bentoml/OpenLLM/pull/379
- chore(deps): bump @mui/base from 5.0.0-beta.14 to 5.0.0-beta.16 by @dependabot in https://github.com/bentoml/OpenLLM/pull/384
- chore(deps): bump @babel/runtime from 7.22.10 to 7.22.15 by @dependabot in https://github.com/bentoml/OpenLLM/pull/362
- chore(deps): bump @mui/system from 5.14.8 to 5.14.10 by @dependabot in https://github.com/bentoml/OpenLLM/pull/383
- chore(deps): bump @mui/x-data-grid from 6.13.0 to 6.14.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/361
- chore(deps): bump @mui/private-theming from 5.14.8 to 5.14.10 by @dependabot in https://github.com/bentoml/OpenLLM/pull/382
- chore(deps): bump @mui/icons-material from 5.14.8 to 5.14.9 by @dependabot in https://github.com/bentoml/OpenLLM/pull/367
- chore(deps): bump @mui/utils from 5.14.8 to 5.14.10 by @dependabot in https://github.com/bentoml/OpenLLM/pull/381
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/406
- chore(deps): bump actions/checkout from 4.0.0 to 4.1.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/405
- chore(deps): bump taiki-e/install-action from 2.18.13 to 2.18.16 by @dependabot in https://github.com/bentoml/OpenLLM/pull/404
- chore(deps-dev): bump eslint from 8.47.0 to 8.50.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/400
- chore(deps): bump nextra from 2.11.1 to 2.13.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/393
- chore(deps-dev): bump @types/react from 18.2.20 to 18.2.22 by @dependabot in https://github.com/bentoml/OpenLLM/pull/403
- chore(deps-dev): bump tailwindcss from 3.3.2 to 3.3.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/398
- chore(deps-dev): bump autoprefixer from 10.4.12 to 10.4.16 by @dependabot in https://github.com/bentoml/OpenLLM/pull/397
- chore(deps): bump @vercel/og from 0.5.13 to 0.5.17 by @dependabot in https://github.com/bentoml/OpenLLM/pull/396
- chore(deps-dev): bump @next/eslint-plugin-next from 13.4.19 to 13.5.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/392
- chore(deps): bump @floating-ui/core from 1.4.1 to 1.5.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/402
- chore(deps-dev): bump @svgr/webpack from 2.0.0-alpha.26fa501a to 8.1.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/389
- chore(deps): bump nextra-theme-docs from 2.12.3 to 2.13.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/411
- chore(deps-dev): bump @typescript-eslint/eslint-plugin from 6.7.0 to 6.7.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/414
- chore(deps-dev): bump shadow-cljs from 2.25.4 to 2.25.6 by @dependabot in https://github.com/bentoml/OpenLLM/pull/394
- chore(deps): bump @mui/x-data-grid from 6.14.0 to 6.15.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/395
- chore(deps-dev): bump @typescript-eslint/parser from 6.7.0 to 6.7.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/412
- chore(deps-dev): bump @types/react from 18.2.20 to 18.2.23 by @dependabot in https://github.com/bentoml/OpenLLM/pull/413
- chore(deps-dev): bump postcss from 8.4.29 to 8.4.30 by @dependabot in https://github.com/bentoml/OpenLLM/pull/390
- chore(deps): bump @mui/x-data-grid from 6.15.0 to 6.16.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/434
- chore(deps-dev): bump postcss from 8.4.30 to 8.4.31 by @dependabot in https://github.com/bentoml/OpenLLM/pull/430
- chore(deps): bump pypa/cibuildwheel from 2.16.0 to 2.16.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/440
- chore(deps): bump github/codeql-action from 2.21.8 to 2.21.9 by @dependabot in https://github.com/bentoml/OpenLLM/pull/439
- chore(deps): bump taiki-e/install-action from 2.18.16 to 2.19.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/438
- chore(deps): bump @mui/x-date-pickers from 6.13.0 to 6.16.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/428
- chore(deps-dev): bump @types/react from 18.2.20 to 18.2.24 by @dependabot in https://github.com/bentoml/OpenLLM/pull/427
- chore(deps-dev): bump @types/node from 20.5.3 to 20.8.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/435
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/441
- chore(deps): bump @babel/runtime from 7.22.15 to 7.23.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/436
- chore(deps): bump nextra from 2.13.0 to 2.13.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/425
- feat: PromptTemplate and system prompt support by @MingLiangDai in https://github.com/bentoml/OpenLLM/pull/407
- feat: OpenAI compatible API by @XunchaoZ in https://github.com/bentoml/OpenLLM/pull/417
- chore(deps): bump pypa/cibuildwheel from 2.16.1 to 2.16.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/474
- chore(deps): bump github/codeql-action from 2.21.9 to 2.22.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/473
- chore(deps): bump aws-actions/configure-aws-credentials from 4.0.0 to 4.0.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/471
- chore(deps): bump actions/setup-python from 4.7.0 to 4.7.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/470
- chore(deps): bump taiki-e/install-action from 2.19.2 to 2.20.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/472
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/475
- fix: do not reply on env var for built bento/docker by @larme in https://github.com/bentoml/OpenLLM/pull/477
- fix: asyncio stalling inside notebooks by @xianml in https://github.com/bentoml/OpenLLM/pull/478
- fix: fix client HTTPS by @sauyon in https://github.com/bentoml/OpenLLM/pull/480
- feat: add 1 openllm llama2 notebook demo by @xianml in https://github.com/bentoml/OpenLLM/pull/479
- chore(deps-dev): bump eslint from 8.47.0 to 8.51.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/469
- chore(deps-dev): bump @types/node from 20.5.3 to 20.8.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/467
- chore(deps-dev): bump turbo from 1.10.14 to 1.10.15 by @dependabot in https://github.com/bentoml/OpenLLM/pull/461
- chore(deps): bump nextra from 2.13.1 to 2.13.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/458
- chore(deps): bump nextra-theme-docs from 2.13.1 to 2.13.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/457
- chore(deps-dev): bump @typescript-eslint/eslint-plugin from 6.7.3 to 6.7.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/483
- chore(deps): bump highlight.js from 11.8.0 to 11.9.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/456
- chore(deps-dev): bump @next/eslint-plugin-next from 13.5.2 to 13.5.4 by @dependabot in https://github.com/bentoml/OpenLLM/pull/464
- chore(deps-dev): bump @typescript-eslint/parser from 6.7.3 to 6.7.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/482
- chore(deps-dev): bump prettier-plugin-tailwindcss from 0.5.4 to 0.5.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/455
- chore(deps-dev): bump @types/react-transition-group from 4.4.6 to 4.4.7 by @dependabot in https://github.com/bentoml/OpenLLM/pull/460
- chore(deps-dev): bump @edge-runtime/vm from 3.1.3 to 3.1.4 by @dependabot in https://github.com/bentoml/OpenLLM/pull/466
- chore(deps): bump github/codeql-action from 2.22.0 to 2.22.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/484
- chore(deps): bump taiki-e/install-action from 2.20.1 to 2.20.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/485
- chore(deps-dev): bump @types/react from 18.2.20 to 18.2.28 by @dependabot in https://github.com/bentoml/OpenLLM/pull/486
- chore(deps): bump @mui/x-data-grid from 6.16.0 to 6.16.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/487
- chore(deps): bump @vercel/og from 0.5.17 to 0.5.18 by @dependabot in https://github.com/bentoml/OpenLLM/pull/488
- chore(deps-dev): bump @types/node from 20.5.3 to 20.8.4 by @dependabot in https://github.com/bentoml/OpenLLM/pull/491
- chore(deps): bump @babel/runtime from 7.23.1 to 7.23.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/490
- feat(client): simple implementation and streaming by @aarnphm in https://github.com/bentoml/OpenLLM/pull/256
New Contributors
- @MingLiangDai made their first contribution in https://github.com/bentoml/OpenLLM/pull/407
- @XunchaoZ made their first contribution in https://github.com/bentoml/OpenLLM/pull/417
- @xianml made their first contribution in https://github.com/bentoml/OpenLLM/pull/478
- @sauyon made their first contribution in https://github.com/bentoml/OpenLLM/pull/480
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.3.6...v0.3.7
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.3.6
Installation
bash
pip install openllm==0.3.6
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.3.6
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.3.6 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.3.6
Find more information about this release in the CHANGELOG.md
What's Changed
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/374
- chore(deps): bump peter-evans/create-pull-request from 4.2.4 to 5.0.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/373
- feat: support continuous batching on
generateby @aarnphm in https://github.com/bentoml/OpenLLM/pull/375
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.3.5...v0.3.6
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.3.5
Installation
bash
pip install openllm==0.3.5
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.3.5
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.3.5 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.3.5
Find more information about this release in the CHANGELOG.md
What's Changed
- fix: set default serialisation methods by @aarnphm in https://github.com/bentoml/OpenLLM/pull/355
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.3.4...v0.3.5
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.3.4
Installation
bash
pip install openllm==0.3.4
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.3.4
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.3.4 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.3.4
Find more information about this release in the CHANGELOG.md
What's Changed
- dos: fix typo by @Sherlock113 in https://github.com/bentoml/OpenLLM/pull/305
- fix(serving): vllm bad num_gpus by @alanpoulain in https://github.com/bentoml/OpenLLM/pull/326
- fix(serialisation): vllm ignore by @aarnphm in https://github.com/bentoml/OpenLLM/pull/324
- feat: continuous batching with vLLM by @aarnphm in https://github.com/bentoml/OpenLLM/pull/349
- fix(prompt): correct export extra objects items by @aarnphm in https://github.com/bentoml/OpenLLM/pull/351
New Contributors
- @alanpoulain made their first contribution in https://github.com/bentoml/OpenLLM/pull/326
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.3.3...v0.3.4
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.3.3
Installation
bash
pip install openllm==0.3.3
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.3.3
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.3.3 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.3.3
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.3.2...v0.3.3
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.3.2
Installation
bash
pip install openllm==0.3.2
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.3.2
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.3.2 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.3.2
Find more information about this release in the CHANGELOG.md
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.3.1...v0.3.2
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.3.1
Installation
bash
pip install openllm==0.3.1
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.3.1
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.3.1 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.3.1
Find more information about this release in the CHANGELOG.md
What's Changed
- docs: Update the readme by @Sherlock113 in https://github.com/bentoml/OpenLLM/pull/302
- revert: disable compiled wheels for now by @aarnphm in https://github.com/bentoml/OpenLLM/pull/304
New Contributors
- @Sherlock113 made their first contribution in https://github.com/bentoml/OpenLLM/pull/302
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.3.0...v0.3.1
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.3.0
Installation
bash
pip install openllm==0.3.0
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.3.0
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.3.0 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.3.0
Find more information about this release in the CHANGELOG.md
What's Changed
- cron(style): run formatter [generated] [skip ci] by @aarnphm in https://github.com/bentoml/OpenLLM/pull/257
- feat(vllm): streaming by @aarnphm in https://github.com/bentoml/OpenLLM/pull/260
- chore(build): use latest vllm pre-built kernel by @aarnphm in https://github.com/bentoml/OpenLLM/pull/261
- chore(deps): bump bentoml/setup-bentoml-action from 59beefe94e2e8f8ebbedf555fc86bd5d1ae0a708 to d4010d8303684b183f45b33c6a44644f81337bdb by @dependabot in https://github.com/bentoml/OpenLLM/pull/267
- chore(deps): bump actions/checkout from 3.5.3 to 3.6.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/269
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/273
- chore(deps): bump docker/setup-buildx-action from 2.9.1 to 2.10.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/271
- chore(deps): bump aws-actions/configure-aws-credentials from 2.2.0 to 3.0.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/270
- chore(deps): bump taiki-e/install-action from 2.16.0 to 2.17.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/268
- chore(deps): bump @mui/icons-material from 5.14.3 to 5.14.6 by @dependabot in https://github.com/bentoml/OpenLLM/pull/265
- chore(deps-dev): bump shadow-cljs from 2.25.2 to 2.25.3 by @dependabot in https://github.com/bentoml/OpenLLM/pull/262
- chore(deps): bump @mui/base from 5.0.0-beta.11 to 5.0.0-beta.12 by @dependabot in https://github.com/bentoml/OpenLLM/pull/264
- chore(deps): bump @mui/material from 5.14.5 to 5.14.6 by @dependabot in https://github.com/bentoml/OpenLLM/pull/266
- chore(deps): bump @mui/x-date-pickers from 6.11.2 to 6.12.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/263
- infra: update build dependencies [clojure-ui build] by @aarnphm in https://github.com/bentoml/OpenLLM/pull/278
- fix: persistent styling between ruff and yapf by @aarnphm in https://github.com/bentoml/OpenLLM/pull/279
- Fixed weird characters over model dropdowns by @GutZuFusss in https://github.com/bentoml/OpenLLM/pull/280
- refactor(breaking): unify LLM API by @aarnphm in https://github.com/bentoml/OpenLLM/pull/283
- fix(yapf): align weird new lines break [generated] [skip ci] by @aarnphm in https://github.com/bentoml/OpenLLM/pull/284
- fix(serving): vllm distributed size by @aarnphm in https://github.com/bentoml/OpenLLM/pull/285
- chore(deps): bump sigstore/cosign-installer from 3.1.1 to 3.1.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/296
- chore(deps): bump taiki-e/install-action from 2.17.2 to 2.17.8 by @dependabot in https://github.com/bentoml/OpenLLM/pull/295
- chore(deps): bump aquasecurity/trivy-action from 559eb1224e654a86c844a795e6702a0742c60c72 to fbd16365eb88e12433951383f5e99bd901fc618f by @dependabot in https://github.com/bentoml/OpenLLM/pull/294
- chore(deps): bump github/codeql-action from 2.21.4 to 2.21.5 by @dependabot in https://github.com/bentoml/OpenLLM/pull/292
- chore(deps): bump crazy-max/ghaction-import-gpg from 5.3.0 to 5.4.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/293
- fix(gptq): use upstream integration by @aarnphm in https://github.com/bentoml/OpenLLM/pull/297
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/298
- chore(deps): bump @mui/x-date-pickers from 6.12.0 to 6.12.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/291
- chore(deps): bump @mui/icons-material from 5.14.6 to 5.14.7 by @dependabot in https://github.com/bentoml/OpenLLM/pull/290
- chore(deps): bump @mui/material from 5.14.6 to 5.14.7 by @dependabot in https://github.com/bentoml/OpenLLM/pull/289
- chore(deps): bump @mui/base from 5.0.0-beta.12 to 5.0.0-beta.13 by @dependabot in https://github.com/bentoml/OpenLLM/pull/287
- chore(deps): bump @mui/x-data-grid from 6.11.2 to 6.12.1 by @dependabot in https://github.com/bentoml/OpenLLM/pull/288
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.2.27...v0.3.0
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.2.27
Installation
bash
pip install openllm==0.2.27
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.2.27
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.2.27 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.2.27
Find more information about this release in the CHANGELOG.md
What's Changed
- feat: token streaming and SSE support by @aarnphm in https://github.com/bentoml/OpenLLM/pull/240
- chore(deps): bump @mui/x-data-grid from 6.11.1 to 6.11.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/242
- chore(deps): bump peter-evans/create-pull-request from 4.2.4 to 5.0.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/244
- chore(deps): bump taiki-e/install-action from 2.15.4 to 2.16.0 by @dependabot in https://github.com/bentoml/OpenLLM/pull/245
- chore(deps): bump @mui/x-date-pickers from 6.0.0 to 6.11.2 by @dependabot in https://github.com/bentoml/OpenLLM/pull/243
- refactor: packages by @aarnphm in https://github.com/bentoml/OpenLLM/pull/249
- ci: pre-commit autoupdate [pre-commit.ci] by @pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/246
- feat(embeddings): Using self-hosted CPU EC2 runner by @aarnphm in https://github.com/bentoml/OpenLLM/pull/250
- refactor(contrib): similar namespace [clojure-ui build] by @aarnphm in https://github.com/bentoml/OpenLLM/pull/251
- chore: ignore peft and fix adapter loading issue by @aarnphm in https://github.com/bentoml/OpenLLM/pull/255
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.2.26...v0.2.27
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.2.26
Installation
bash
pip install openllm==0.2.26
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.2.26
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.2.26 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.2.26
Find more information about this release in the CHANGELOG.md
What's Changed
- chore(gh): use setup-bentoml-action by @aarnphm in https://github.com/bentoml/OpenLLM/pull/230
- fix(binary): move to correct folders when building standalone installer by @aarnphm in https://github.com/bentoml/OpenLLM/pull/228
- chore: conditional commit for running jobs by @aarnphm in https://github.com/bentoml/OpenLLM/pull/232
- feat(embedding): Adding generic endpoint by @aarnphm in https://github.com/bentoml/OpenLLM/pull/227
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.2.25...v0.2.26
- Python
Published by github-actions[bot] over 2 years ago
openllm - v0.2.25
What changed?
The long-awaited ClojureScript UI is now GA. Try it out with docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.2.25. Thanks @GutZuFusss
Added vLLM support for Falcon, and general CQA.
Installation
bash
pip install openllm==0.2.25
To upgrade from a previous version, use the following command:
bash
pip install --upgrade openllm==0.2.25
Usage
All available models: openllm models
To start a LLM: python -m openllm start opt
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P ghcr.io/bentoml/openllm:0.2.25 start opt
To run OpenLLM Clojure UI (community-maintained): docker run -p 8420:80 ghcr.io/bentoml/openllm-ui-clojure:0.2.25
Find more information about this release in the CHANGELOG.md
What's Changed
- chore: upload nightly wheels to test.pypi.org by @aarnphm in https://github.com/bentoml/OpenLLM/pull/215
- feat(contrib): ClojureScript UI by @GutZuFusss in https://github.com/bentoml/OpenLLM/pull/89
- fix(ci): remove broken build hooks by @aarnphm in https://github.com/bentoml/OpenLLM/pull/216
- chore(ci): add dependabot and fix vllm release container by @aarnphm in https://github.com/bentoml/OpenLLM/pull/217
- feat(models): add vLLM support for Falcon by @aarnphm in https://github.com/bentoml/OpenLLM/pull/223
- chore(readme): update nightly badge [skip ci] by @aarnphm in https://github.com/bentoml/OpenLLM/pull/224
New Contributors
- @GutZuFusss made their first contribution in https://github.com/bentoml/OpenLLM/pull/89
Full Changelog: https://github.com/bentoml/OpenLLM/compare/v0.2.24...v0.2.25
- Python
Published by aarnphm over 2 years ago