Releases: SciSharp/LLamaSharp
Releases · SciSharp/LLamaSharp
v0.24.0
Major Changes
- May Binary Update by @martindevans, @AmSmart and @nipeone in #1179 / #1163
- Android Support by @AmSmart in #1130 / #1179
- Linux-ARM64 support by @nipeone in #1138 / #1179
- Feat/tensor override by @dpmm99 in #1180 and @martindevans in #1185
- Add LLamaReranker by @nipeone in #1150
Bug Fixes
- Adding extra values to the decode_result enum by @phil-scott-78 in #1140
- adding missing using statement by @NicklausBrain in #1154
- Fixed LLama.Web error starting session by @tonib in #1158
- Fixed CI (Ubuntu-24.04) by @martindevans in #1162
- Update LLamaEmbedder, Examples packages, and KernelMemory examples by @zsogitbe in #1170
Other Changes
- docs: Added link to new package MaIN.NET by @wisedev-pstach in #1155
- Create stale_issues.yml by @sangyuxiaowu in #1171
- docs: add deep-wiki link to readme. by @AsakusaRinne in #1172
- Update to M.E.AI 9.4.0-preview.1.25207.5 by @jozkee in #1156
- Update to M.E.AI 9.4.3-preview.1.25230.7 by @jozkee in #1182
- Update to M.E.AI 9.5.0-preview by @jozkee in #1186
New Contributors
- @NicklausBrain made their first contribution in #1154
- @wisedev-pstach made their first contribution in #1155
- @jozkee made their first contribution in #1156
- @tonib made their first contribution in #1158
- @nipeone made their first contribution in #1150
Full Changelog: v0.22.0...v0.24.0
v0.22.0
Major Changes
- Grammar Resampling by @martindevans, @m0nsky & @phil-scott-78 in #1109
- Added 'strict' parameter for template retrieval by @Lyrcaxis in #1124
- March 2025 version update by @martindevans in #1126
Bug Fixes
- Various fixes regarding special tokens (embeddings & anti-prompt) by @Lyrcaxis in #1112
- Fix Kernel Memory by @martindevans in #1081
- Updates json gbnf to match llama.cpp example by @phil-scott-78 in #1098
- InteractiveExecutor now stops at EOS tokens again by @dpmm99 in #1092
Other Changes
- Update README.md by @sangyuxiaowu in #1086
- Smaller Unit Test Model by @martindevans in #1105
- Update Microsoft.Extensions.AI to 9.3.0-preview.1.25114.11 by @stephentoub in #1102
- Cleaned up non-important warnings from the error list by @Lyrcaxis in #1111
- Update to M.E.AI 9.3.0-preview.1.25161.3 by @stephentoub in #1127
- Improved Memory Rental by @martindevans in #1113
Full Changelog: v0.21.0...v0.22.0
v0.21.0
v0.20.0
Major Changes
- January 2025 Update by @m0nsky, @Lyrcaxis, @SignalRT & @martindevans in #1036
- Removed .NET6 by @martindevans in #982
- Implemented
LogitBias
forDefaultSamplingPipeline
by @martindevans in #986
Bug Fixes
- Fixed
LLamaTokenDataArrayNative.Create
's size initialization by @Lyrcaxis in #981 - Fixed
ChatSession.LoadSession
by @martindevans in #976 - Non-deterministic default seed by @Lyrcaxis in #993
- Fixed log levels (issue #995) by @LoicDagnas in #997
- Fix Batching shared token issues by @martindevans in #1045
Other Changes
- Adding simple batch example by @phil-scott-78 in #1038
- Add an option to apply the chat template to prompts when using StateslessExecutor by @phil-scott-78 in #1025
- Adds an exception for missing templates. by @phil-scott-78 in #1034
- Updated tests model to "LLama 3.2 1B Instruct" by @Lyrcaxis in #983
- Update M.E.AI.Abstractions by @stephentoub in #979, #992 and #1044
- Added sampling tests by @Lyrcaxis in #984
- Updated Nuget Dependencies by @martindevans in #987
- Helper method for sampling
Conversation
withSafeLLamaSamplerChainHandle
by @martindevans in #994
New Contributors
- @LoicDagnas made their first contribution in #997
- @AcoranGonzalezMoray made their first contribution in #1015
- @phil-scott-78 made their first contribution in #1025
Full Changelog: v0.19.0...v0.20.0
v0.19.0
Major Changes
- November Binary Update by @martindevans in #962
- Custom Sampler Stages by @martindevans in #961
- Add Microsoft.Extensions.AI support for IChatClient / IEmbeddingGenerator by @stephentoub in #964
Bug Fixes
- Non-Null Default
SamplingPipeline
by @martindevans in #973
New Contributors
- @stephentoub made their first contribution in #964
Full Changelog: v0.18.0...v0.19.0
v0.18.0
Major Changes
Other Changes
- Updates to ContributingGuide for latest llama.cpp repo by @scritch1sm in #953
- Fix README chat session example by @easis in #956
New Contributors
Full Changelog: v0.17.0...v0.18.0
v0.17.0
Important: The CUDA packages for 0.17.0 exceeded the maximum size for a nuget package. This means some of the 0.17.0 packages are not available until a new way is to deploy those packages. If you need one of the missing packages, use 0.16.0 instead.
Major Changes
- October binary update by @martindevans in #940
Bug Fixes
- Fix: Increase Vulkan detection timeout to 12 seconds by @LSXPrime in #931
- Specify file open permissions needed as read only. by @scritch1sm in #943
- Fix
llama_get_logits_ith
Null Handling by @martindevans in #920 - Fix CUDA compile error by @martindevans in #924
Other Changes
- Add LLama2 Chat Session example with a custom templator by @asmirnov82 in #938
New Contributors
- @LSXPrime made their first contribution in #931
- @scritch1sm made their first contribution in #943
Full Changelog: v0.16.0...v0.17.0
v0.16.0
Major Changes
- August 2024 Binary Update by @martindevans in #905
- LLamaEmbedder 2.0 by @martindevans in #902
Bug Fixes
- Fixed constrained output (Grammar) support in BatchedExecutorGuidance by @webitube in #904
- Fixed Issue where Model with Grammar constrains only generates one output in Example by @L-Wehmschulte in #908
Other Changes
- Remove OpenCL & add Vulkan to readme by @m0nsky in #907
- Bump Swashbuckle.AspNetCore from 6.6.2 to 6.7.3 by @dependabot in #912
- Bump BenchmarkDotNet from 0.13.12 to 0.14.0 by @dependabot in #913
- Bump Microsoft.NET.Test.Sdk from 17.10.0 to 17.11.0 by @dependabot in #914
- Bump Microsoft.AspNetCore.Mvc.Razor.RuntimeCompilation from 8.0.7 to 8.0.8 by @dependabot in #916
- Bump Microsoft.KernelMemory.Abstractions and Microsoft.SemanticKernel.Abstractions by @dependabot in #915
New Contributors
- @webitube made their first contribution in #904
- @L-Wehmschulte made their first contribution in #908
Full Changelog: v0.15.0...v0.16.0
v0.15.0 - Llama 3.1
Major Changes
- Llama 3.1 update binaries by @martindevans in #874
- Microsoft.KernelMemory version 0.68+ compatibility fix by @SpaceAntelope in #862
Bug Fixes
- Fixed missing
[DllImport]
forllama_encode
by @martindevans in #864 - Safer Vulkan System Info by @martindevans in #885
- Fixed LLamaTimings by @martindevans in #881
Other Changes
- Fixed Incorrect Exception Docs Comments by @martindevans in #883
- Deprecate
InferenceParams
sampler configuration properties by @martindevans in #871 - Updated
Semantic Kernel
andKernel Memory
by @martindevans in #880 - Update README.md by @sangyuxiaowu in #872
New Contributors
- @SpaceAntelope made their first contribution in #862
Full Changelog: v0.14.0...v0.15.0
0.14.0 - Gemma2
Major Changes
- July 2024 binaries by @martindevans, @SignalRT & @m0nsky in #833, #843
- Vulkan support by @m0nsky in #794, #797, #799
- Generic Prompt Formatter by @patrick-hovsepian in #787
- Async implementation of LLamaExecutors by @asmirnov82 in #834
- Include debug symbols with NuGet packages by @dpmm99 in #851
- Allow continuation in Instruct and Interact executors by @dpmm99 in #852
Bug Fixes
- Fixes Typo in IContextParamsExtensions by @SerialKicked in #806
- Fix possible file locks due to race conditions during build & test by @m0nsky in #836
- Fixed CI Out Of Disk Space by @martindevans in #846
- Fix CUDA detection on WSL by @m0nsky in #847
Other Changes
- Remove Beam Search by @martindevans in #783
- Batched Beam Search by @martindevans in #796
- Remove OpenCL Backend. by @martindevans in #782
- BoolQ Benchmark by @martindevans in #802
New Contributors
- @patrick-hovsepian made their first contribution in #787
- @SerialKicked made their first contribution in #806
- @dpmm99 made their first contribution in #851
Full Changelog: v0.13.0...v0.14.0