.. |
Abstractions
|
Assorted small changes to clean up some code warnings
|
2024-02-17 23:07:10 +00:00 |
Batched
|
Assorted small changes to clean up some code warnings
|
2024-02-17 23:07:10 +00:00 |
Common
|
Updated everything to work with llama.cpp ce32060198b7e2d6a13a9b8e1e1369e3c295ae2a
|
2024-02-01 16:35:05 +00:00 |
Exceptions
|
Swapped `StatelessExecutor` to use `llama_decode`!
|
2024-01-20 21:18:35 +00:00 |
Extensions
|
Updated everything to work with llama.cpp ce32060198b7e2d6a13a9b8e1e1369e3c295ae2a
|
2024-02-01 16:35:05 +00:00 |
Grammars
|
Code cleanup driven by R# suggestions:
|
2024-01-02 03:20:21 +00:00 |
Native
|
NativeLibraryConfig: WithLogs(LLamaLogLevel) (#529)
|
2024-02-21 23:51:09 +00:00 |
Sampling
|
Created a new `BatchedExecutor` which processes multiple "Conversations" in one single inference batch. This is faster, even when the conversations are unrelated, and is much faster if the conversations share some overlap (e.g. a common system prompt prefix).
|
2024-02-09 23:57:03 +00:00 |
runtimes
|
Merge pull request #479 from martindevans/update_binaries_feb_2024
|
2024-02-06 01:08:09 +00:00 |
AntipromptProcessor.cs
|
Make AntipromptProcessor public
|
2023-11-14 21:41:12 +13:00 |
AssemblyAttributes.cs
|
- Created a higher level `Grammar` class which is immutable and contains a list of grammar rules. This is the main "entry point" to the grammar system.
|
2023-08-31 00:02:50 +01:00 |
ChatSession.cs
|
ChatSession: improve exception message
|
2024-02-19 17:50:37 -05:00 |
GlobalSuppressions.cs
|
- Created a higher level `Grammar` class which is immutable and contains a list of grammar rules. This is the main "entry point" to the grammar system.
|
2023-08-31 00:02:50 +01:00 |
LLamaContext.cs
|
Created a new `BatchedExecutor` which processes multiple "Conversations" in one single inference batch. This is faster, even when the conversations are unrelated, and is much faster if the conversations share some overlap (e.g. a common system prompt prefix).
|
2024-02-09 23:57:03 +00:00 |
LLamaEmbedder.cs
|
Normalizing embeddings in `LLamaEmbedder`. As is done in llama.cpp: 2891c8aa9a/examples/embedding/embedding.cpp (L92)
|
2024-02-13 02:09:35 +00:00 |
LLamaExecutorBase.cs
|
Updated everything to work with llama.cpp ce32060198b7e2d6a13a9b8e1e1369e3c295ae2a
|
2024-02-01 16:35:05 +00:00 |
LLamaInstructExecutor.cs
|
Created a new `BatchedExecutor` which processes multiple "Conversations" in one single inference batch. This is faster, even when the conversations are unrelated, and is much faster if the conversations share some overlap (e.g. a common system prompt prefix).
|
2024-02-09 23:57:03 +00:00 |
LLamaInteractExecutor.cs
|
Assorted small changes to clean up some code warnings
|
2024-02-17 23:07:10 +00:00 |
LLamaQuantizer.cs
|
Added new file types to quantisation
|
2024-02-06 18:06:10 +00:00 |
LLamaSharp.Runtime.targets
|
Use llama instead of libllama in `[DllImport]`
|
2024-01-30 02:40:13 -05:00 |
LLamaSharp.csproj
|
0.10.0
|
2024-02-15 14:40:56 +00:00 |
LLamaStatelessExecutor.cs
|
Fixed build error caused by extra unnecessary parameter
|
2024-01-23 15:09:20 +00:00 |
LLamaTransforms.cs
|
- Improved coverage of `GBNFGrammarParser` up to 96%
|
2023-10-13 23:54:01 +01:00 |
LLamaWeights.cs
|
Improved the BatchedDecoding demo:
|
2024-01-20 17:39:50 +00:00 |
StreamingTokenDecoder.cs
|
Assorted small changes to clean up some code warnings
|
2024-02-17 23:07:10 +00:00 |