41 lines
985 B
JSON
41 lines
985 B
JSON
{
|
|
"Logging": {
|
|
"LogLevel": {
|
|
"Default": "Information",
|
|
"Microsoft.AspNetCore": "Warning"
|
|
}
|
|
},
|
|
"AllowedHosts": "*",
|
|
"LLamaOptions": {
|
|
"ModelLoadType": 0,
|
|
"Models": [
|
|
{
|
|
"Name": "LLama2-7b-Chat",
|
|
"MaxInstances": 20,
|
|
"ModelPath": "..\\LLama.Unittest\\Models\\llama-2-7b-chat.Q4_0.gguf",
|
|
"ContextSize": 2048,
|
|
"BatchSize": 2048,
|
|
"Threads": 4,
|
|
"GpuLayerCount": 6,
|
|
"UseMemorymap": true,
|
|
"UseMemoryLock": false,
|
|
"MainGpu": 0,
|
|
"LowVram": false,
|
|
"Seed": 1686349486,
|
|
"UseFp16Memory": true,
|
|
"Perplexity": false,
|
|
"LoraAdapter": "",
|
|
"LoraBase": "",
|
|
"EmbeddingMode": false,
|
|
"TensorSplits": null,
|
|
"GroupedQueryAttention": 1,
|
|
"RmsNormEpsilon": 0.000005,
|
|
"RopeFrequencyBase": 10000.0,
|
|
"RopeFrequencyScale": 1.0,
|
|
"MulMatQ": false,
|
|
"Encoding": "UTF-8"
|
|
}
|
|
]
|
|
}
|
|
}
|