Skip to content

Releases: SciSharp/LLamaSharp

v0.8.1 - Major BUG fix and better feature detection

28 Nov 17:16
1f97ad8
Compare
Choose a tag to compare

Break changes

  • Change NativeLibraryConfig.Default to NativeLibraryConfig.Instance.

Major features and fix

Other changes

New Contributors

Thank you so much for all the contributions!

Full Changelog: v0.8.0...v0.8.1

v0.8.0: performance improvement, cuda feature detection and kernel-memory integration

12 Nov 19:53
597d329
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.7.0...v0.8.0

v0.7.0 - improve performance

01 Nov 18:28
5a9e13c
Compare
Choose a tag to compare

This release fixes the performance problem in v0.6.0, so that it's strongly recommended to upgraded to this version. Many thanks for the catch of this problem by @lexxsoft and the fix from @martindevans !

What's Changed

Full Changelog: v0.6.0...v0.7.0

v0.6.0 - follow major llama.cpp changes

24 Oct 17:39
1299de3
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.5.1...v0.6.0

v0.5.1 - GGUF, grammar and semantic-kernel integration

05 Sep 18:49
f3e23d6
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.4.0...v0.5.0

v0.4.2-preview: new backends

06 Aug 13:34
bfe9cc8
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.4.1-preview...v0.4.2-preview

v0.4.1-preview - follow up llama.cpp latest commit

21 Jun 21:08
f836a2c
Compare
Choose a tag to compare

This is a preview version which followed up the latest modifications of llama.cpp.

For some reasons the cuda backend hasn't been okay, we'll release v0.4.1 after dealing with that.

v0.4.0 - Executor and ChatSession

19 Jun 19:13
43dcce8
Compare
Choose a tag to compare

Version 0.4.0 introduces many break changes. However we strongly recommend to upgrade to 0.4.0 because it provides better abstractions and stability by refactoring the framework. The backend v0.3.0 and v0.3.1 still works for LLamaSharp v0.4.0.

The main changes:

  1. Add three-level abstractions: LLamaModel, LLamaExecutor and ChatSession.
  2. Fix the BUG of saving and loading state.
  3. Support saving/loading chat session directly.
  4. Add more flexible APIs in the chat session.
  5. Add detailed documentations: https://scisharp.github.io/LLamaSharp/0.4/

Acknowledge

During the development, thanks a lot for the help from @TheTerrasque ! His/Her fork gives us many inspirations. Besides, many thanks for the following contributors!

New Contributors

v0.3.0 - Load and save state

22 May 21:49
9a4bf8e
Compare
Choose a tag to compare
  1. Support loading and saving state.
  2. Support tokenization and detokenization.
  3. Fix BUGs of instruct mode.
  4. break change: n_parts param is removed.
  5. break change: LLamaModelV1 is dropped.
  6. Remove dependencies for third-party loggers.
  7. Verified model repo is added on huggingface.
  8. Optimize the examples.

v0.2.3 - Inference BUG Fix

17 May 21:53
3e53ed4
Compare
Choose a tag to compare

Fix some strange behaviors of model inference.