feiyun0112.SemanticKernel.Connectors.OnnxRuntimeGenAI.CUDA
1.0.0
dotnet add package feiyun0112.SemanticKernel.Connectors.OnnxRuntimeGenAI.CUDA --version 1.0.0
NuGet\Install-Package feiyun0112.SemanticKernel.Connectors.OnnxRuntimeGenAI.CUDA -Version 1.0.0
This command is intended to be used within the Package Manager Console in Visual Studio, as it uses the NuGet module's version of Install-Package.
<PackageReference Include="feiyun0112.SemanticKernel.Connectors.OnnxRuntimeGenAI.CUDA" Version="1.0.0" />
For projects that support PackageReference, copy this XML node into the project file to reference the package.
paket add feiyun0112.SemanticKernel.Connectors.OnnxRuntimeGenAI.CUDA --version 1.0.0
The NuGet Team does not provide support for this client. Please contact its maintainers for support.
#r "nuget: feiyun0112.SemanticKernel.Connectors.OnnxRuntimeGenAI.CUDA, 1.0.0"
#r directive can be used in F# Interactive and Polyglot Notebooks. Copy this into the interactive tool or source code of the script to reference the package.
// Install feiyun0112.SemanticKernel.Connectors.OnnxRuntimeGenAI.CUDA as a Cake Addin #addin nuget:?package=feiyun0112.SemanticKernel.Connectors.OnnxRuntimeGenAI.CUDA&version=1.0.0 // Install feiyun0112.SemanticKernel.Connectors.OnnxRuntimeGenAI.CUDA as a Cake Tool #tool nuget:?package=feiyun0112.SemanticKernel.Connectors.OnnxRuntimeGenAI.CUDA&version=1.0.0
The NuGet Team does not provide support for this client. Please contact its maintainers for support.
SemanticKernel.Connectors.OnnxRuntimeGenAI
Semantic Kernel connector for ONNX models.
How to use
Prerequisites
ONNX models, for example Phi-3 Mini-4K-Instruct
git lfs install
git clone https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-onnx
Code
Create a new console app and add Nuget Package:
-- for CPU
feiyun0112.SemanticKernel.Connectors.OnnxRuntimeGenAI.CPU
-- for CUDA
feiyun0112.SemanticKernel.Connectors.OnnxRuntimeGenAI.CUDA
Then change Program.cs to:
Kernel kernel = Kernel.CreateBuilder()
.AddOnnxRuntimeGenAIChatCompletion(
modelPath: @"d:\Phi-3-mini-4k-instruct-onnx\cpu_and_mobile\cpu-int4-rtn-block-32-acc-level-4")
.Build();
string prompt = @"Write a joke";
await foreach (string text in kernel.InvokePromptStreamingAsync<string>(prompt,
new KernelArguments(new OnnxRuntimeGenAIPromptExecutionSettings() { MaxLength = 2048 })))
{
Console.Write(text);
}
使用说明
先决条件
ONNX models, for example Phi-3 Mini-4K-Instruct
git lfs install
git clone https://hf-mirror.com/microsoft/Phi-3-mini-4k-instruct-onnx
示例代码
创建新的控制台应用并添加 Nuget 包:
-- for CPU
feiyun0112.SemanticKernel.Connectors.OnnxRuntimeGenAI.CPU
-- for CUDA
feiyun0112.SemanticKernel.Connectors.OnnxRuntimeGenAI.CUDA
然后将Program.cs更改为:
Kernel kernel = Kernel.CreateBuilder()
.AddOnnxRuntimeGenAIChatCompletion(
modelPath: @"d:\Phi-3-mini-4k-instruct-onnx\cpu_and_mobile\cpu-int4-rtn-block-32-acc-level-4")
.Build();
string prompt = @"Write a joke";
await foreach (string text in kernel.InvokePromptStreamingAsync<string>(prompt,
new KernelArguments(new OnnxRuntimeGenAIPromptExecutionSettings() { MaxLength = 2048 })))
{
Console.Write(text);
}
Product | Versions Compatible and additional computed target framework versions. |
---|---|
.NET | net5.0 was computed. net5.0-windows was computed. net6.0 was computed. net6.0-android was computed. net6.0-ios was computed. net6.0-maccatalyst was computed. net6.0-macos was computed. net6.0-tvos was computed. net6.0-windows was computed. net7.0 was computed. net7.0-android was computed. net7.0-ios was computed. net7.0-maccatalyst was computed. net7.0-macos was computed. net7.0-tvos was computed. net7.0-windows was computed. net8.0 was computed. net8.0-android was computed. net8.0-browser was computed. net8.0-ios was computed. net8.0-maccatalyst was computed. net8.0-macos was computed. net8.0-tvos was computed. net8.0-windows was computed. |
.NET Core | netcoreapp3.0 was computed. netcoreapp3.1 was computed. |
.NET Standard | netstandard2.1 is compatible. |
MonoAndroid | monoandroid was computed. |
MonoMac | monomac was computed. |
MonoTouch | monotouch was computed. |
Tizen | tizen60 was computed. |
Xamarin.iOS | xamarinios was computed. |
Xamarin.Mac | xamarinmac was computed. |
Xamarin.TVOS | xamarintvos was computed. |
Xamarin.WatchOS | xamarinwatchos was computed. |
Compatible target framework(s)
Included target framework(s) (in package)
Learn more about Target Frameworks and .NET Standard.
-
.NETStandard 2.1
- Microsoft.ML.OnnxRuntimeGenAI.Cuda (>= 0.2.0-rc4)
- Microsoft.SemanticKernel.Core (>= 1.10.0)
NuGet packages
This package is not used by any NuGet packages.
GitHub repositories
This package is not used by any popular GitHub repositories.
Version | Downloads | Last updated |
---|---|---|
1.0.0 | 200 | 5/3/2024 |