39 lines
1.6 KiB
C#
39 lines
1.6 KiB
C#
using LLama.Native;
|
|
using Spectre.Console;
|
|
|
|
AnsiConsole.MarkupLineInterpolated(
|
|
$"""
|
|
[purple]======================================================================================================[/]
|
|
__ __ ____ __
|
|
/\ \ /\ \ /\ _`\ /\ \
|
|
\ \ \ \ \ \ __ ___ ___ __ \ \,\L\_\\ \ \___ __ _ __ _____
|
|
\ \ \ __\ \ \ __ /'__`\ /' __` __`\ /'__`\ \/_\__ \ \ \ _ `\ /'__`\ /\` __\/\ __`\
|
|
\ \ \L\ \\ \ \L\ \/\ \L\.\_ /\ \/\ \/\ \ /\ \L\.\_ /\ \L\ \\ \ \ \ \ /\ \L\.\_\ \ \/ \ \ \L\ \
|
|
\ \____/ \ \____/\ \__/.\_\\ \_\ \_\ \_\\ \__/.\_\\ `\____\\ \_\ \_\\ \__/.\_\\ \_\ \ \ ,__/
|
|
\/___/ \/___/ \/__/\/_/ \/_/\/_/\/_/ \/__/\/_/ \/_____/ \/_/\/_/ \/__/\/_/ \/_/ \ \ \/
|
|
[purple]=========================================================================================[/] \ \_\ [purple]======[/]
|
|
\/_/
|
|
|
|
""");
|
|
|
|
// Configure native library to use. This must be done before any other llama.cpp methods are called!
|
|
NativeLibraryConfig
|
|
.Instance
|
|
.WithCuda();
|
|
|
|
// Configure logging. Change this to `true` to see log messages from llama.cpp
|
|
var showLLamaCppLogs = false;
|
|
NativeLibraryConfig
|
|
.Instance
|
|
.WithLogCallback((level, message) =>
|
|
{
|
|
if (showLLamaCppLogs)
|
|
Console.WriteLine($"[llama {level}]: {message.TrimEnd('\n')}");
|
|
});
|
|
|
|
// Calling this method forces loading to occur now.
|
|
NativeApi.llama_empty_call();
|
|
|
|
await ExampleRunner.Run();
|
|
|