Releases: Lightning-AI/LitServe
Releases · Lightning-AI/LitServe
v0.2.6
What's Changed
- feat: info route by @lorenzomassimiani in #368
- Fix CI: async tests with ASGITransport by @aniketmaurya in #376
- Fix: Replace Deprecated
max_tokens
withmax_completion_tokens
in OpenAI Spec by @rittik9 in #375 - feat: Customizable Loops 1/n by @aniketmaurya in #374
- customizable loop - wire up Loops to LitServer 2/n by @aniketmaurya in #378
- Improve CI: retry flaky tests by @aniketmaurya in #379
- [pre-commit.ci] pre-commit suggestions by @pre-commit-ci in #377
- check device format while initialising litserver by @ali-alshaar7 in #380
- Release 0.2.6.dev0 by @aniketmaurya in #383
- Update PR template by @rittik9 in #381
- Include user field and
base64
literal for encoding_format by @aniketmaurya in #388 - Improve error handling and debugging experience by @aniketmaurya in #389
- improved logging with sensible defaults by @aniketmaurya in #391
- add continuous batching loop 1/n by @aniketmaurya in #387
- Add
loop.pre_setup
to allow fine-grained LitAPI validation based on inference loop by @aniketmaurya in #393 - Make
LitAPI.predict
optional and validate API implementation by @aniketmaurya in #394 - Fix OpenAISpec with continuous batching loop by @aniketmaurya in #395
- add tests for continuous batching and Default loops by @aniketmaurya in #396
- Set LitServer.stream using LitSpec.stream by @aniketmaurya in #398
- fix openai usage info for non-streaming response by @aniketmaurya in #399
- [pre-commit.ci] pre-commit suggestions by @pre-commit-ci in #400
- Async continuous batching loop by @aniketmaurya in #401
- add validation for
stream=False
withyield
usage by @aniketmaurya in #402 - fix callback runner to execute after predict by @aniketmaurya in #406
- integrate zmq by @aniketmaurya in #403
- warn users when predict/unbatch output length is not same as #requests by @aniketmaurya in #408
- move built in loops inside classes by @aniketmaurya in #409
- add justus and thomas as codeowners by @aniketmaurya in #410
- enable multiple workers for ZMQ by @aniketmaurya in #411
- Fix: Add Callback Events and Align Hooks in Streaming Loop by @bhimrazy in #407
- bump:
Lightning-AI/utilities
usedmain
by @Borda in #415 - Release v0.2.6 by @aniketmaurya in #413
- fix: don't start zmq when fast_queue=false by @aniketmaurya in #417
- fix release ci by @aniketmaurya in #416
New Contributors
- @rittik9 made their first contribution in #375
- @ali-alshaar7 made their first contribution in #380
Full Changelog: v0.2.5...v0.2.6
v0.2.6.dev3
What's Changed
- Set LitServer.stream using LitSpec.stream by @aniketmaurya in #398
- fix openai usage info for non-streaming response by @aniketmaurya in #399
- [pre-commit.ci] pre-commit suggestions by @pre-commit-ci in #400
- Async continuous batching loop by @aniketmaurya in #401
- add validation for
stream=False
withyield
usage by @aniketmaurya in #402 - fix callback runner to execute after predict by @aniketmaurya in #406
- integrate zmq by @aniketmaurya in #403
- warn users when predict/unbatch output length is not same as #requests by @aniketmaurya in #408
- move built in loops inside classes by @aniketmaurya in #409
- add justus and thomas as codeowners by @aniketmaurya in #410
- enable multiple workers for ZMQ by @aniketmaurya in #411
Full Changelog: v0.2.6.dev2...v0.2.6.dev3
v0.2.6.dev2
What's Changed
- Make
LitAPI.predict
optional and validate API implementation by @aniketmaurya in #394 - Fix OpenAISpec with continuous batching loop by @aniketmaurya in #395
- add tests for continuous batching and Default loops by @aniketmaurya in #396
Full Changelog: v0.2.6.dev1...v0.2.6.dev2
v0.2.6.dev1
What's Changed
- Update PR template by @rittik9 in #381
- Include user field and
base64
literal for encoding_format by @aniketmaurya in #388 - Improve error handling and debugging experience by @aniketmaurya in #389
- improved logging with sensible defaults by @aniketmaurya in #391
- add continuous batching loop 1/n by @aniketmaurya in #387
- Add
loop.pre_setup
to allow fine-grained LitAPI validation based on inference loop by @aniketmaurya in #393
Full Changelog: v0.2.6.dev0...v0.2.6.dev1
v0.2.6.dev0
What's Changed
- feat: info route by @lorenzomassimiani in #368
- Fix CI: async tests with ASGITransport by @aniketmaurya in #376
- Fix: Replace Deprecated
max_tokens
withmax_completion_tokens
in OpenAI Spec by @rittik9 in #375 - feat: Customizable Loops 1/n by @aniketmaurya in #374
- customizable loop - wire up Loops to LitServer 2/n by @aniketmaurya in #378
- Improve CI: retry flaky tests by @aniketmaurya in #379
- [pre-commit.ci] pre-commit suggestions by @pre-commit-ci in #377
- check device format while initialising litserver by @ali-alshaar7 in #380
- Release 0.2.6.dev0 by @aniketmaurya in #383
New Contributors
- @rittik9 made their first contribution in #375
- @ali-alshaar7 made their first contribution in #380
Full Changelog: v0.2.5...v0.2.6.dev0
v0.2.5
What's Changed
- [pre-commit.ci] pre-commit suggestions by @pre-commit-ci in #353
- Fail fast when LitAPI.setup has error by @aniketmaurya in #356
- Development release 0.2.5.dev0 by @aniketmaurya in #357
- make @Borda global code owner by @aniketmaurya in #358
- Update media type and JSON handling in OpenAISpec by @fabigr8 in #360
- Add custom healthcheck path by @akuma12 in #342
- Automatic Serialization/deserialization of images by @aniketmaurya in #361
- Update docker base image to avoid issue with builds by @AlexShukhman in #363
- Enable multiple image input by @aniketmaurya in #364
- Test multiple Python version compatibility by @aniketmaurya in #336
- ci: prune configuration by @Borda in #365
- Feat support OpenAI embedding by @bhimrazy in #367
- Fix: Make OpenAIEmbedding work when token usage info is not set by @aniketmaurya in #369
- Assert batch size to strictly test
OpenAIEmbeddingSpec
for dynamic batching by @aniketmaurya in #370 - Release v0.2.5 by @aniketmaurya in #371
New Contributors
- @fabigr8 made their first contribution in #360
- @akuma12 made their first contribution in #342
- @AlexShukhman made their first contribution in #363
Full Changelog: v0.2.4...v0.2.5
v0.2.5.dev0
What's Changed
- [pre-commit.ci] pre-commit suggestions by @pre-commit-ci in #353
- Fail fast when LitAPI.setup has error by @aniketmaurya in #356
- Development release 0.2.5.dev0 by @aniketmaurya in #357
- make @Borda global code owner by @aniketmaurya in #358
Full Changelog: v0.2.4...v0.2.5.dev0
v0.2.4
What's Changed
- Dynamically grab port number by @grantgasser in #307
- test generate client file by @aniketmaurya in #322
- avoid client generation if already exist by @aniketmaurya in #324
- move middlewares to a module by @aniketmaurya in #326
- enable isort by @aniketmaurya in #327
- Feat: track number of active requests by @aniketmaurya in #325
- aggregate requests across uvicorn processes by @aniketmaurya in #328
- lint: emergency
docformatter
bump by @aniketmaurya in #331 - update(lint): prune config by @Borda in #332
- Avoiding counting active requests in /health and /metrics by @vrdn-23 in #334
- Host param by @grumpyp in #335
- improve error handling from inference worker processes by @aniketmaurya in #337
- Fix Docker Argument Handling by @aniketmaurya in #338
- Add dockerization to feature list by @aniketmaurya in #346
- stream tool calling by @whisper-bye in #349
- Release 0.2.4 by @aniketmaurya in #351
New Contributors
- @grantgasser made their first contribution in #307
- @vrdn-23 made their first contribution in #334
- @whisper-bye made their first contribution in #349
Full Changelog: v0.2.3...v0.2.4
v0.2.3
What's Changed
- test LitAPI decode-encode methods by @aniketmaurya in #244
- fix readme code sanity check by @aniketmaurya in #246
- add e2e test for simple streaming server by @AdolfoVillalobos in #247
- separate loops to new module
loops.py
by @aniketmaurya in #249 - feat: middlewares in LitServer. by @lorenzomassimiani in #241
- test LitServer.run by @aniketmaurya in #248
- fix flaky test_collate_requests by @aniketmaurya in #250
- remove dead code by @aniketmaurya in #251
- add litserve.api tests by @aniketmaurya in #252
- add Starlette middleware support by @aniketmaurya in #253
- Add test for
litserve.examples
by @aniketmaurya in #254 - rename examples to test_examples by @aniketmaurya in #255
- add
litserve.connector
tests by @aniketmaurya in #256 - chore: adds Qwen2-VL Lightning Studio example by @bhimrazy in #259
- Bump mypy from 1.11.1 to 1.11.2 by @dependabot in #261
- Bump Lightning-AI/utilities from 0.11.6 to 0.11.7 by @dependabot in #264
- [pre-commit.ci] pre-commit suggestions by @pre-commit-ci in #266
- add
litserve.loops
tests by @aniketmaurya in #257 - ci: drop dependabot by @Borda in #268
- add perf test CI 1/3 by @aniketmaurya in #269
- install uvloop with uvcorn by @aniketmaurya in #274
- chore: adds Parler TTS Studio example by @bhimrazy in #277
- minor LitServer code clean up by @aniketmaurya in #279
- [Chore] rename setup_server to register_endpoints by @aniketmaurya in #281
- Callback system v0 by @aniketmaurya in #278
- add streaming no-batch speed test by @aniketmaurya in #283
- fix: callbacks EventType typo by @aniketmaurya in #285
- [Feat]: add Logger API by @aniketmaurya in #284
- fix flaky Logger test by @aniketmaurya in #288
- handle exceptions for logger processing by @aniketmaurya in #287
- docs: update feat template / readability by @Borda in #295
- update PR template with hiding instructions by @Borda in #297
- add links to forum and reduce opening issues for docs by @Borda in #296
- update ask a question link to reddit by @aniketmaurya in #299
- Add warning for unexpected model output in batched prediction by @grumpyp in #300
- fix custom HTTPException with argument names by @aniketmaurya in #302
- fix vLLM capitalization by @aniketmaurya in #303
- generate Dockerfile with CLI by @aniketmaurya in #298
- add jsonargparse in requirements by @aniketmaurya in #306
- add codeowners for requirements by @aniketmaurya in #308
- Enable gpu docker image generation by @aniketmaurya in #311
- Pyright linting fixes by @grumpyp in #314
- add Ethan as codeowner by @aniketmaurya in #313
- update
dockerize
CLI success msg copy by @aniketmaurya in #312 - switch to
argparse
for CLI by @aniketmaurya in #309 - fix dockerize CLI test by @aniketmaurya in #318
- fix(Logger): handle unpickleable Logger by @aniketmaurya in #319
- add example usage for
dockerize
CLI by @aniketmaurya in #315 - Release 0.2.3 by @aniketmaurya in #317
New Contributors
- @AdolfoVillalobos made their first contribution in #247
- @lorenzomassimiani made their first contribution in #241
- @grumpyp made their first contribution in #300
Full Changelog: v0.2.2...v0.2.3
v0.2.2
What's Changed
- Fix spelling of IdentityAPI class in test by @eltociear in #222
- improve code readability by @aniketmaurya in #224
- remove dead code by @aniketmaurya in #226
- set default batch timeout to 1 ms by @aniketmaurya in #229
- Revert "set default batch timeout to 1 ms (#229)" by @aniketmaurya in #232
- chore(litserve): added RAG deployment example in readme by @patchy631 in #231
- print swagger ui url by @aniketmaurya in #235
- chore: adds bg removal api example by @bhimrazy in #234
- test
collate_request
w batch_timeout and batch_size by @aniketmaurya in #238 - Release 0.2.2 by @aniketmaurya in #240
- Fix: Spelling Error in PR Template:
becuase
==>because
by @bhimrazy in #242
New Contributors
- @eltociear made their first contribution in #222
- @patchy631 made their first contribution in #231
Full Changelog: v0.2.1...v0.2.2