LLamaSharp/LLama
Scott W Harden a6394001a1
NativeLibraryConfig: WithLogs(LLamaLogLevel) (#529)
Adds a NativeLibraryConfig.WithLogs() overload to let the user indicate the log level (with "info" as the default)
2024-02-21 23:51:09 +00:00
..
Abstractions Assorted small changes to clean up some code warnings 2024-02-17 23:07:10 +00:00
Batched Assorted small changes to clean up some code warnings 2024-02-17 23:07:10 +00:00
Common Updated everything to work with llama.cpp ce32060198b7e2d6a13a9b8e1e1369e3c295ae2a 2024-02-01 16:35:05 +00:00
Exceptions Swapped `StatelessExecutor` to use `llama_decode`! 2024-01-20 21:18:35 +00:00
Extensions Updated everything to work with llama.cpp ce32060198b7e2d6a13a9b8e1e1369e3c295ae2a 2024-02-01 16:35:05 +00:00
Grammars Code cleanup driven by R# suggestions: 2024-01-02 03:20:21 +00:00
Native NativeLibraryConfig: WithLogs(LLamaLogLevel) (#529) 2024-02-21 23:51:09 +00:00
Sampling Created a new `BatchedExecutor` which processes multiple "Conversations" in one single inference batch. This is faster, even when the conversations are unrelated, and is much faster if the conversations share some overlap (e.g. a common system prompt prefix). 2024-02-09 23:57:03 +00:00
runtimes Merge pull request #479 from martindevans/update_binaries_feb_2024 2024-02-06 01:08:09 +00:00
AntipromptProcessor.cs Make AntipromptProcessor public 2023-11-14 21:41:12 +13:00
AssemblyAttributes.cs - Created a higher level `Grammar` class which is immutable and contains a list of grammar rules. This is the main "entry point" to the grammar system. 2023-08-31 00:02:50 +01:00
ChatSession.cs ChatSession: improve exception message 2024-02-19 17:50:37 -05:00
GlobalSuppressions.cs - Created a higher level `Grammar` class which is immutable and contains a list of grammar rules. This is the main "entry point" to the grammar system. 2023-08-31 00:02:50 +01:00
LLamaContext.cs Created a new `BatchedExecutor` which processes multiple "Conversations" in one single inference batch. This is faster, even when the conversations are unrelated, and is much faster if the conversations share some overlap (e.g. a common system prompt prefix). 2024-02-09 23:57:03 +00:00
LLamaEmbedder.cs Normalizing embeddings in `LLamaEmbedder`. As is done in llama.cpp: 2891c8aa9a/examples/embedding/embedding.cpp (L92) 2024-02-13 02:09:35 +00:00
LLamaExecutorBase.cs Updated everything to work with llama.cpp ce32060198b7e2d6a13a9b8e1e1369e3c295ae2a 2024-02-01 16:35:05 +00:00
LLamaInstructExecutor.cs Created a new `BatchedExecutor` which processes multiple "Conversations" in one single inference batch. This is faster, even when the conversations are unrelated, and is much faster if the conversations share some overlap (e.g. a common system prompt prefix). 2024-02-09 23:57:03 +00:00
LLamaInteractExecutor.cs Assorted small changes to clean up some code warnings 2024-02-17 23:07:10 +00:00
LLamaQuantizer.cs Added new file types to quantisation 2024-02-06 18:06:10 +00:00
LLamaSharp.Runtime.targets Use llama instead of libllama in `[DllImport]` 2024-01-30 02:40:13 -05:00
LLamaSharp.csproj 0.10.0 2024-02-15 14:40:56 +00:00
LLamaStatelessExecutor.cs Fixed build error caused by extra unnecessary parameter 2024-01-23 15:09:20 +00:00
LLamaTransforms.cs - Improved coverage of `GBNFGrammarParser` up to 96% 2023-10-13 23:54:01 +01:00
LLamaWeights.cs Improved the BatchedDecoding demo: 2024-01-20 17:39:50 +00:00
StreamingTokenDecoder.cs Assorted small changes to clean up some code warnings 2024-02-17 23:07:10 +00:00