RLMatrix 0.1.2

There is a newer version of this package available.
See the version list below for details.
dotnet add package RLMatrix --version 0.1.2                
NuGet\Install-Package RLMatrix -Version 0.1.2                
This command is intended to be used within the Package Manager Console in Visual Studio, as it uses the NuGet module's version of Install-Package.
<PackageReference Include="RLMatrix" Version="0.1.2" />                
For projects that support PackageReference, copy this XML node into the project file to reference the package.
paket add RLMatrix --version 0.1.2                
#r "nuget: RLMatrix, 0.1.2"                
#r directive can be used in F# Interactive and Polyglot Notebooks. Copy this into the interactive tool or source code of the script to reference the package.
// Install RLMatrix as a Cake Addin
#addin nuget:?package=RLMatrix&version=0.1.2

// Install RLMatrix as a Cake Tool
#tool nuget:?package=RLMatrix&version=0.1.2                

RLMatrixBanner

🚀 RL Matrix - Elevate Your .NET Reinforcement Learning Experience with TorchSharp!

Dive into the future of type-safe Reinforcement Learning with RL Matrix, powered by the might of TorchSharp. RL Matrix stands out as a user-friendly toolkit offering a collection of RL algorithms—primed for plug, play, and prosper! NuGet NuGet

  • PPO
  • DQN

🎯 What Sparks RL Matrix?

While embarking on my RL journey, I sensed a gap in the reinforcement learning world even with TorchSharp's solid foundation. It struck me—C# is the ideal choice for RL outside research circles, thanks to its pristine and intuitive coding experience. No more guessing games in environment or agent building!

With RL Matrix, our vision is to offer a seamless experience. By simply incorporating the IEnvironment interface, you're equipped to rapidly craft and unleash Reinforcement Learning Agents (RL Agents). Switching between algorithms? A breeze! It’s our nod to the elegance of Matlab's toolkit methodology.

enter image description here

🚀 Getting Started:

Peek into the /examples/ directory for illustrative code. But to give you a quick start:

  1. Craft an IEnvironment class: Comply with reinforcement learning guidelines, defining your observation shapes and action count:
public class CartPole : IEnvironment<float[]>
{
  public int stepCounter { get; set; }
  public int maxSteps { get; set; }
  public bool isDone { get; set; }
  public OneOf<int, (int, int)> stateSize { get; set; }
  public int actionSize { get; set; }

  CartPoleEnv myEnv;

  private float[] myState;

  public CartPole()
  {
      Initialise();
  }

  public float[] GetCurrentState()
  {
      if (myState == null)
          myState = new float[4] {0,0,0,0};
      return myState;
  }

  public void Initialise()
  {
      myEnv = new CartPoleEnv(WinFormEnvViewer.Factory);
      stepCounter = 0;
      maxSteps = 100000;
      stateSize = myEnv.ObservationSpace.Shape.Size;
      actionSize = myEnv.ActionSpace.Shape.Size;
      myEnv.Reset();
      isDone = false; 
  }

  public void Reset()
  {
  //For instance:
      myEnv.Reset();
      isDone = false;
      stepCounter = 0;
  }

  public float Step(int actionId)
  {
  //Whatever step logic, returns reward
      return reward;
  }
}
  1. Agent Instance & Training: Spawn an agent for your environment and ignite the TrainEpisode method:
var opts = new DQNAgentOptions(batchSize: 64, memorySize: 10000, gamma: 0.99f, epsStart: 1f, epsEnd: 0.05f, epsDecay: 50f, tau: 0.005f, lr: 1e-4f, displayPlot: myChart);
var env = new CartPole();
var myAgent = new D2QNAgent<float[,]>(opts, env);

for (int i = 0; i < 400; i++)
{
    myAgent.TrainEpisode();
}

📌 Current Roadmap:

Adapting data to 1D (feed-forward networks) and 2D (convolutional networks) without source modifications. Support for discrete actions only. As we innovate, anticipate breaking changes. We'll keep you in the loop!

💌 Get in Touch:

Questions? Ideas? Collaborations? Drop a line at: 📧 contact@exmachinasoft.com

🤝 Join the Journey:

We believe in collective brilliance! If you're inspired to enhance RL Matrix, please open an issue or PR. Let’s shape the future of RL together!

Product Compatible and additional computed target framework versions.
.NET net6.0 is compatible.  net6.0-android was computed.  net6.0-ios was computed.  net6.0-maccatalyst was computed.  net6.0-macos was computed.  net6.0-tvos was computed.  net6.0-windows was computed.  net7.0 was computed.  net7.0-android was computed.  net7.0-ios was computed.  net7.0-maccatalyst was computed.  net7.0-macos was computed.  net7.0-tvos was computed.  net7.0-windows was computed.  net8.0 was computed.  net8.0-android was computed.  net8.0-browser was computed.  net8.0-ios was computed.  net8.0-maccatalyst was computed.  net8.0-macos was computed.  net8.0-tvos was computed.  net8.0-windows was computed. 
Compatible target framework(s)
Included target framework(s) (in package)
Learn more about Target Frameworks and .NET Standard.

NuGet packages (3)

Showing the top 3 NuGet packages that depend on RLMatrix:

Package Downloads
RLMatrix.Godot

Deep Reinfocement Learning Library wiring RLMatrix with Godot objects for easy of use with Godot Engine

RLMatrix.WinformsChart

Quick Winforms Chart for RL Matrix

RLMatrix.Server

Server components for Deep Reinforcement Learning in .NET

GitHub repositories (1)

Showing the top 1 popular GitHub repositories that depend on RLMatrix:

Repository Stars
asieradzk/RL_Matrix
Deep Reinforcement Learning in C#
Version Downloads Last updated
0.4.3.3 576 10/10/2024
0.4.3.2 356 10/10/2024
0.4.3 935 8/12/2024
0.4.2 587 8/12/2024
0.4.1 630 8/9/2024
0.4.0 636 8/8/2024
0.3.0 712 7/22/2024
0.2.1 610 3/1/2024
0.2.0 1,068 1/19/2024
0.1.2 425 10/16/2023
0.1.1 360 10/4/2023
0.1.0 703 8/25/2023