Skip to content

Releases: xorbitsai/inference

v0.14.1.post1

13 Aug 06:56
9afee76

Choose a tag to compare

What's new in 0.14.1.post1 (2024-08-13)

These are the changes in inference v0.14.1.post1.

Bug fixes

  • BUG: limit AutoAWQ version to fix docker issue by @qinxuye in #2067

Documentation

  • DOC: update readme & add tips for large image models by @qinxuye in #2056

Full Changelog: v0.14.1...v0.14.1.post1

v0.14.1

09 Aug 10:14
3e7ed86

Choose a tag to compare

What's new in 0.14.1 (2024-08-09)

These are the changes in inference v0.14.1.

New features

Enhancements

Bug fixes

Tests

Documentation

Others

New Contributors

Full Changelog: v0.14.0...v0.14.1

v0.14.0.post1

05 Aug 06:06
1112993

Choose a tag to compare

What's new in 0.14.0.post1 (2024-08-05)

These are the changes in inference v0.14.0.post1.

Enhancements

Bug fixes

Full Changelog: v0.14.0...v0.14.0.post1

v0.14.0

02 Aug 08:42
dd85cfe

Choose a tag to compare

What's new in 0.14.0 (2024-08-02)

These are the changes in inference v0.14.0.

New features

  • FEAT: Supports model_path input when launching models by @Valdanitooooo in #1918
  • FEAT: Support gte-Qwen2-7B-instruct and multi gpu deploy by @amumu96 in #1994

Enhancements

Bug fixes

Documentation

New Contributors

Full Changelog: v0.13.3...v0.14.0

v0.13.3

26 Jul 10:59
aa51ff2

Choose a tag to compare

What's new in 0.13.3 (2024-07-26)

These are the changes in inference v0.13.3.

New features

Enhancements

  • ENH: added MLX for llama-3-instruct, codestral, Yi-1.5-chat, internlm2.5-chat by @qinxuye in #1908
  • ENH: add gptq for llama-3-instruct by @Phoenix500526 in #1934

Bug fixes

  • BUG: fix inpainting and flexible infer due to inner API change by @qinxuye in #1907

Documentation

New Contributors

Full Changelog: v0.13.2...v0.13.3

v0.13.2

19 Jul 11:26
880929c

Choose a tag to compare

What's new in 0.13.2 (2024-07-19)

These are the changes in inference v0.13.2.

New features

Enhancements

Bug fixes

  • BUG: Fix stream unicode issue for chinese characters when using vllm backend by @ChengjieLi28 in #1865
  • BUG: sglang stream error while stream_option not set by @wxiwnd in #1901
  • BUG: fix client import by @amumu96 in #1905

Full Changelog: v0.13.1...v0.13.2

v0.13.1

12 Jul 11:10
5e3f254

Choose a tag to compare

What's new in 0.13.1 (2024-07-12)

These are the changes in inference v0.13.1.

New features

Enhancements

Bug fixes

  • BUG: cache status missing for model id with quantization placeholder by @Zihann73 in #1849

Documentation

Others

New Contributors

Full Changelog: v0.13.0...v0.13.1

v0.13.0

05 Jul 10:33
007408c

Choose a tag to compare

What's new in 0.13.0 (2024-07-05)

These are the changes in inference v0.13.0.

New features

Enhancements

Bug fixes

Tests

Documentation

Full Changelog: v0.12.3...v0.13.0

v0.12.3

28 Jun 07:36
3d9c261

Choose a tag to compare

What's new in 0.12.3 (2024-06-28)

These are the changes in inference v0.12.3.

New features

Enhancements

Bug fixes

Others

  • CHORE: upgrade version fix security vulnerability by @rickywu in #1674

New Contributors

Full Changelog: v0.12.2...v0.12.3

v0.12.2.post1

22 Jun 17:37
7705d4a

Choose a tag to compare

What's new in 0.12.2.post1 (2024-06-22)

These are the changes in inference v0.12.2.post1.

Enhancements

Full Changelog: v0.12.2...v0.12.2.post1