Releases: Lightning-AI/LitServe
Releases · Lightning-AI/LitServe
v0.2.6.dev2
What's Changed
- Make
LitAPI.predict
optional and validate API implementation by @aniketmaurya in #394 - Fix OpenAISpec with continuous batching loop by @aniketmaurya in #395
- add tests for continuous batching and Default loops by @aniketmaurya in #396
Full Changelog: v0.2.6.dev1...v0.2.6.dev2
v0.2.6.dev1
What's Changed
- Update PR template by @rittik9 in #381
- Include user field and
base64
literal for encoding_format by @aniketmaurya in #388 - Improve error handling and debugging experience by @aniketmaurya in #389
- improved logging with sensible defaults by @aniketmaurya in #391
- add continuous batching loop 1/n by @aniketmaurya in #387
- Add
loop.pre_setup
to allow fine-grained LitAPI validation based on inference loop by @aniketmaurya in #393
Full Changelog: v0.2.6.dev0...v0.2.6.dev1
v0.2.6.dev0
What's Changed
- feat: info route by @lorenzomassimiani in #368
- Fix CI: async tests with ASGITransport by @aniketmaurya in #376
- Fix: Replace Deprecated
max_tokens
withmax_completion_tokens
in OpenAI Spec by @rittik9 in #375 - feat: Customizable Loops 1/n by @aniketmaurya in #374
- customizable loop - wire up Loops to LitServer 2/n by @aniketmaurya in #378
- Improve CI: retry flaky tests by @aniketmaurya in #379
- [pre-commit.ci] pre-commit suggestions by @pre-commit-ci in #377
- check device format while initialising litserver by @ali-alshaar7 in #380
- Release 0.2.6.dev0 by @aniketmaurya in #383
New Contributors
- @rittik9 made their first contribution in #375
- @ali-alshaar7 made their first contribution in #380
Full Changelog: v0.2.5...v0.2.6.dev0
v0.2.5
What's Changed
- [pre-commit.ci] pre-commit suggestions by @pre-commit-ci in #353
- Fail fast when LitAPI.setup has error by @aniketmaurya in #356
- Development release 0.2.5.dev0 by @aniketmaurya in #357
- make @Borda global code owner by @aniketmaurya in #358
- Update media type and JSON handling in OpenAISpec by @fabigr8 in #360
- Add custom healthcheck path by @akuma12 in #342
- Automatic Serialization/deserialization of images by @aniketmaurya in #361
- Update docker base image to avoid issue with builds by @AlexShukhman in #363
- Enable multiple image input by @aniketmaurya in #364
- Test multiple Python version compatibility by @aniketmaurya in #336
- ci: prune configuration by @Borda in #365
- Feat support OpenAI embedding by @bhimrazy in #367
- Fix: Make OpenAIEmbedding work when token usage info is not set by @aniketmaurya in #369
- Assert batch size to strictly test
OpenAIEmbeddingSpec
for dynamic batching by @aniketmaurya in #370 - Release v0.2.5 by @aniketmaurya in #371
New Contributors
- @fabigr8 made their first contribution in #360
- @akuma12 made their first contribution in #342
- @AlexShukhman made their first contribution in #363
Full Changelog: v0.2.4...v0.2.5
v0.2.5.dev0
What's Changed
- [pre-commit.ci] pre-commit suggestions by @pre-commit-ci in #353
- Fail fast when LitAPI.setup has error by @aniketmaurya in #356
- Development release 0.2.5.dev0 by @aniketmaurya in #357
- make @Borda global code owner by @aniketmaurya in #358
Full Changelog: v0.2.4...v0.2.5.dev0
v0.2.4
What's Changed
- Dynamically grab port number by @grantgasser in #307
- test generate client file by @aniketmaurya in #322
- avoid client generation if already exist by @aniketmaurya in #324
- move middlewares to a module by @aniketmaurya in #326
- enable isort by @aniketmaurya in #327
- Feat: track number of active requests by @aniketmaurya in #325
- aggregate requests across uvicorn processes by @aniketmaurya in #328
- lint: emergency
docformatter
bump by @aniketmaurya in #331 - update(lint): prune config by @Borda in #332
- Avoiding counting active requests in /health and /metrics by @vrdn-23 in #334
- Host param by @grumpyp in #335
- improve error handling from inference worker processes by @aniketmaurya in #337
- Fix Docker Argument Handling by @aniketmaurya in #338
- Add dockerization to feature list by @aniketmaurya in #346
- stream tool calling by @whisper-bye in #349
- Release 0.2.4 by @aniketmaurya in #351
New Contributors
- @grantgasser made their first contribution in #307
- @vrdn-23 made their first contribution in #334
- @whisper-bye made their first contribution in #349
Full Changelog: v0.2.3...v0.2.4
v0.2.3
What's Changed
- test LitAPI decode-encode methods by @aniketmaurya in #244
- fix readme code sanity check by @aniketmaurya in #246
- add e2e test for simple streaming server by @AdolfoVillalobos in #247
- separate loops to new module
loops.py
by @aniketmaurya in #249 - feat: middlewares in LitServer. by @lorenzomassimiani in #241
- test LitServer.run by @aniketmaurya in #248
- fix flaky test_collate_requests by @aniketmaurya in #250
- remove dead code by @aniketmaurya in #251
- add litserve.api tests by @aniketmaurya in #252
- add Starlette middleware support by @aniketmaurya in #253
- Add test for
litserve.examples
by @aniketmaurya in #254 - rename examples to test_examples by @aniketmaurya in #255
- add
litserve.connector
tests by @aniketmaurya in #256 - chore: adds Qwen2-VL Lightning Studio example by @bhimrazy in #259
- Bump mypy from 1.11.1 to 1.11.2 by @dependabot in #261
- Bump Lightning-AI/utilities from 0.11.6 to 0.11.7 by @dependabot in #264
- [pre-commit.ci] pre-commit suggestions by @pre-commit-ci in #266
- add
litserve.loops
tests by @aniketmaurya in #257 - ci: drop dependabot by @Borda in #268
- add perf test CI 1/3 by @aniketmaurya in #269
- install uvloop with uvcorn by @aniketmaurya in #274
- chore: adds Parler TTS Studio example by @bhimrazy in #277
- minor LitServer code clean up by @aniketmaurya in #279
- [Chore] rename setup_server to register_endpoints by @aniketmaurya in #281
- Callback system v0 by @aniketmaurya in #278
- add streaming no-batch speed test by @aniketmaurya in #283
- fix: callbacks EventType typo by @aniketmaurya in #285
- [Feat]: add Logger API by @aniketmaurya in #284
- fix flaky Logger test by @aniketmaurya in #288
- handle exceptions for logger processing by @aniketmaurya in #287
- docs: update feat template / readability by @Borda in #295
- update PR template with hiding instructions by @Borda in #297
- add links to forum and reduce opening issues for docs by @Borda in #296
- update ask a question link to reddit by @aniketmaurya in #299
- Add warning for unexpected model output in batched prediction by @grumpyp in #300
- fix custom HTTPException with argument names by @aniketmaurya in #302
- fix vLLM capitalization by @aniketmaurya in #303
- generate Dockerfile with CLI by @aniketmaurya in #298
- add jsonargparse in requirements by @aniketmaurya in #306
- add codeowners for requirements by @aniketmaurya in #308
- Enable gpu docker image generation by @aniketmaurya in #311
- Pyright linting fixes by @grumpyp in #314
- add Ethan as codeowner by @aniketmaurya in #313
- update
dockerize
CLI success msg copy by @aniketmaurya in #312 - switch to
argparse
for CLI by @aniketmaurya in #309 - fix dockerize CLI test by @aniketmaurya in #318
- fix(Logger): handle unpickleable Logger by @aniketmaurya in #319
- add example usage for
dockerize
CLI by @aniketmaurya in #315 - Release 0.2.3 by @aniketmaurya in #317
New Contributors
- @AdolfoVillalobos made their first contribution in #247
- @lorenzomassimiani made their first contribution in #241
- @grumpyp made their first contribution in #300
Full Changelog: v0.2.2...v0.2.3
v0.2.2
What's Changed
- Fix spelling of IdentityAPI class in test by @eltociear in #222
- improve code readability by @aniketmaurya in #224
- remove dead code by @aniketmaurya in #226
- set default batch timeout to 1 ms by @aniketmaurya in #229
- Revert "set default batch timeout to 1 ms (#229)" by @aniketmaurya in #232
- chore(litserve): added RAG deployment example in readme by @patchy631 in #231
- print swagger ui url by @aniketmaurya in #235
- chore: adds bg removal api example by @bhimrazy in #234
- test
collate_request
w batch_timeout and batch_size by @aniketmaurya in #238 - Release 0.2.2 by @aniketmaurya in #240
- Fix: Spelling Error in PR Template:
becuase
==>because
by @bhimrazy in #242
New Contributors
- @eltociear made their first contribution in #222
- @patchy631 made their first contribution in #231
Full Changelog: v0.2.1...v0.2.2
v0.2.1
What's Changed
- add phi3 multimodal template by @aniketmaurya in #211
- Update README.md by @ckark in #213
- Chore: update README.md by @ckark in #214
- Added updated litserve vid to README.md by @likethecognac in #215
- Correct spelling of AuraFlow by @isidentical in #216
- Fix flaky test by @aniketmaurya in #219
- Enable batch-unbatch by default by @aniketmaurya in #220
- bump version by @aniketmaurya in #221
New Contributors
- @ckark made their first contribution in #213
- @likethecognac made their first contribution in #215
- @isidentical made their first contribution in #216
Full Changelog: v0.2.0...v0.2.1
v0.2.0
What's Changed
- Add warning message if
batch
andunbatch
is implemented but max_batch_size is unset. by @bhimrazy in #185 - cleanup: move middleware to utils by @aniketmaurya in #189
- Add meaningful error message if response queues are not initialized by @rasbt in #191
- [pre-commit.ci] pre-commit suggestions by @pre-commit-ci in #193
- add codeowners by @aniketmaurya in #194
- cleanup: fix test naming convention by @aniketmaurya in #190
- properly shutdown litserve workers by @aniketmaurya in #192
- provide uvicorn configs with kwargs by @aniketmaurya in #198
- remove uvicorn from argument names by @aniketmaurya in #199
- fix flaky batch timeout test by @aniketmaurya in #200
- moved wrap_litserve_start to utils by @ankitsharma07 in #201
- bump version by @aniketmaurya in #202
- remove dead code by @aniketmaurya in #204
- bump version v0.2.0 by @aniketmaurya in #203
- add fastapi parity test by @aniketmaurya in #183
- Feat: Add support for response_format field in OpenAI Spec by @bhimrazy in #197
- Enable parity test for GPU by @aniketmaurya in #205
- add classical ML examples by @aniketmaurya in #206
- fix Studio links by @aniketmaurya in #207
- minor cleanup by @aniketmaurya in #209
New Contributors
- @ankitsharma07 made their first contribution in #201
Full Changelog: v0.1.5...v0.2.0