docs: update the readme.

This commit is contained in:
Yaohui Liu 2023-05-18 20:34:08 +08:00
parent d6bd1b7107
commit 1573ae049f
No known key found for this signature in database
GPG Key ID: E86D01E1809BD23E
1 changed files with 5 additions and 0 deletions

View File

@ -42,6 +42,11 @@ Note that version v0.2.1 has a package named `LLamaSharp.Cpu`. After v0.2.2 it w
We publish the backend with cpu, cuda11 and cuda12 because they are the most popular ones. If none of them matches, please compile the [llama.cpp](https://github.com/ggerganov/llama.cpp)
from source and put the `libllama` under your project's output path. When building from source, please add `-DBUILD_SHARED_LIBS=ON` to enable the library generation.
## FAQ
1. GPU out of memory: v0.2.3 put all layers into GPU by default. If the momory use is out of the capacity of your GPU, please set `n_gpu_layers` to a smaller number.
2. Unsupported model: `llama.cpp` is under quick development and often has break changes. Please check the release date of the model and find a suitable version of LLamaSharp to install.
## Simple Benchmark
Currently it's only a simple benchmark to indicate that the performance of `LLamaSharp` is close to `llama.cpp`. Experiments run on a computer