LLamaSharp/LLama.Web/appsettings.json

41 lines
985 B
JSON

{
"Logging": {
"LogLevel": {
"Default": "Information",
"Microsoft.AspNetCore": "Warning"
}
},
"AllowedHosts": "*",
"LLamaOptions": {
"ModelLoadType": 0,
"Models": [
{
"Name": "LLama2-7b-Chat",
"MaxInstances": 20,
"ModelPath": "..\\LLama.Unittest\\Models\\llama-2-7b-chat.Q4_0.gguf",
"ContextSize": 2048,
"BatchSize": 2048,
"Threads": 4,
"GpuLayerCount": 6,
"UseMemorymap": true,
"UseMemoryLock": false,
"MainGpu": 0,
"LowVram": false,
"Seed": 1686349486,
"UseFp16Memory": true,
"Perplexity": false,
"LoraAdapter": "",
"LoraBase": "",
"EmbeddingMode": false,
"TensorSplits": null,
"GroupedQueryAttention": 1,
"RmsNormEpsilon": 0.000005,
"RopeFrequencyBase": 10000.0,
"RopeFrequencyScale": 1.0,
"MulMatQ": false,
"Encoding": "UTF-8"
}
]
}
}