MASES.EntityFrameworkCore.KNet.Serialization.Protobuf 2.3.3

There is a newer version of this package available.
See the version list below for details.
dotnet add package MASES.EntityFrameworkCore.KNet.Serialization.Protobuf --version 2.3.3                
NuGet\Install-Package MASES.EntityFrameworkCore.KNet.Serialization.Protobuf -Version 2.3.3                
This command is intended to be used within the Package Manager Console in Visual Studio, as it uses the NuGet module's version of Install-Package.
<PackageReference Include="MASES.EntityFrameworkCore.KNet.Serialization.Protobuf" Version="2.3.3" />                
For projects that support PackageReference, copy this XML node into the project file to reference the package.
paket add MASES.EntityFrameworkCore.KNet.Serialization.Protobuf --version 2.3.3                
#r "nuget: MASES.EntityFrameworkCore.KNet.Serialization.Protobuf, 2.3.3"                
#r directive can be used in F# Interactive and Polyglot Notebooks. Copy this into the interactive tool or source code of the script to reference the package.
// Install MASES.EntityFrameworkCore.KNet.Serialization.Protobuf as a Cake Addin
#addin nuget:?package=MASES.EntityFrameworkCore.KNet.Serialization.Protobuf&version=2.3.3

// Install MASES.EntityFrameworkCore.KNet.Serialization.Protobuf as a Cake Tool
#tool nuget:?package=MASES.EntityFrameworkCore.KNet.Serialization.Protobuf&version=2.3.3                

title: Serialization in KEFCore _description: Describes how works the serialization in Entity Framework Core provider for Apache Kafka

KEFCore: serialization

Entity Framework Core provider for Apache Kafka shall convert the entities used within the model in something viable from the backend. Each backend has its own schema to convert entities into something else; database providers converts entities into database schema or blob in Cosmos.

IMPORTANT NOTE: till the first major version, all releases shall be considered not stable: this means the API public, or internal, can change without notice.

Basic concepts

Entity Framework Core provider for Apache Kafka shall convert the entities into record will be stored in the topics of Apache Kafka cluster. The way the entities are converted shall follows a schema. The current schema follows a JSON pattern and reports the information of each entity as:

  • Primary Key:

    • Simple: if the Primary Key is a native type (e.g. int, long, double, and so on) the serialization is made using the Apache Kafka default serializer for that type
    • Complex: if the Primary Key is a complex type (e.g. int-int, int-long, int-string, and so on), Entity Framework reports it as an array of objects and the serialization is made using a JSON serializer
  • Entity data: the Entity is managed, from Entity Framework Core provider for Apache Kafka, as an array of objects associated to properties of the Entity. The schema of the Apache Kafka record value follows the code definition in DefaultValueContainer<T>. Below two examples:

    {
      "EntityName": "MASES.EntityFrameworkCore.KNet.Test.Blog",
      "ClrType": "MASES.EntityFrameworkCore.KNet.Test.Blog",
      "Data": {
        "0": {
          "PropertyName": "BlogId",
          "ClrType": "System.Int32",
          "Value": 8
        },
        "1": {
          "PropertyName": "Rating",
          "ClrType": "System.Int32",
          "Value": 7
        },
        "2": {
          "PropertyName": "Url",
          "ClrType": "System.String",
          "Value": "http://blogs.msdn.com/adonet7"
        }
      }
    }
    
    {
      "EntityName": "MASES.EntityFrameworkCore.KNet.Test.Post",
      "ClrType": "MASES.EntityFrameworkCore.KNet.Test.Post",
      "Data": {
        "0": {
          "PropertyName": "PostId",
          "ClrType": "System.Int32",
          "Value": 44
        },
        "1": {
          "PropertyName": "BlogId",
          "ClrType": "System.Int32",
          "Value": 44
        },
        "2": {
          "PropertyName": "Content",
          "ClrType": "System.String",
          "Value": "43"
        },
        "3": {
          "PropertyName": "Title",
          "ClrType": "System.String",
          "Value": "title"
        }
      }
    }
    

The equivalent JSON schema is not available till now.

Code and user override

The code is based on three elements shall be available to Entity Framework Core provider for Apache Kafka in order to work:

  • ValueContainer type: a type which encapsulate the Entity and stores needed information
  • Key SerDes: the serializer of the Primary Key
  • ValueContainer SerDes: the serializer of the ValueContainer

Default types

Entity Framework Core provider for Apache Kafka comes with some default values:

  • ValueContainer class: KEFCore uses DefaultValueContainer<T> (i.e. DefaultKEFCoreSerDes.DefaultValueContainer) which stores the CLR type of Entity, the properties ordered by their index with associated CLT type, name and JSON serializaed value; the class is marked for JSON serialization and it is used from the ValueContainer SerDes;
  • Key SerDes class: KEFCore uses DefaultKEFCoreSerDes.Key.JsonRaw<T> (i.e. DefaultKEFCoreSerDes.DefaultKeySerialization), the type automatically manages simple or complex Primary Key
  • ValueContainer SerDes class: KEFCore uses DefaultKEFCoreSerDes.ValueContainer.JsonRaw<> (i.e. DefaultKEFCoreSerDes.DefaultValueContainerSerialization)

Both Key and ValueContainer SerDes come with two kind of data transfer mechanisms:

  • Raw: it uses byte arrays for data transfer
  • Buffered: they use ByteBuffer for data transfer

User override

The default serialization can be overridden with user defined ValueContainer class, Key SerDes or ValueContainer SerDes.

ValueContainer class

A custom ValueContainer class must contains enough information and shall follow the following rules:

  • must implements the IValueContainer<T> interface
  • must be a generic type
  • must have at least a default constructor and a constructor which accept two parameters: a first parameter which is IEntityType and a second paramater of object[]

An example snippet is the follow:

public class CustomValueContainer<TKey> : IValueContainer<TKey> where TKey : notnull
{
    /// <summary>
    /// Initialize a new instance of <see cref="CustomValueContainer{TKey}"/>
    /// </summary>
    /// <param name="tName">The <see cref="IEntityType"/> requesting the ValueContainer for <paramref name="rData"/></param>
    /// <param name="rData">The data, built from EFCore, to be stored in the ValueContainer</param>
    /// <remarks>This constructor is mandatory and it is used from KEFCore to request a ValueContainer</remarks>
    public CustomValueContainer(IEntityType tName, object[] rData)
    {

    }

    /// <inheritdoc/>
    public string EntityName { get; set; }
    /// <inheritdoc/>
    public string ClrType { get; set; }
    /// <inheritdoc/>
    public void GetData(IEntityType tName, ref object[] array)
    {

    }
    /// <inheritdoc/>
    public IReadOnlyDictionary<int, string> GetProperties()
    {
        // build properties
    }
}
Key SerDes and ValueContainer SerDes class

A custom Key SerDes class shall follow the following rules:

  • must implements the ISerDes<T> interface or extend SerDes<T>
  • must be a generic type
  • must have a parameterless constructor
  • can store serialization information using Headers of Apache Kafka record (this information will be used from EntityExtractor)

An example snippet is the follow based on JSON serializer:

public class CustomKeySerDes<T> : SerDesRaw<T>
{
    readonly byte[] keyTypeName = Encoding.UTF8.GetBytes(typeof(T).FullName!);
    readonly byte[] customSerDesName = Encoding.UTF8.GetBytes(typeof(CustomKeySerDes<>).FullName!);

    /// <inheritdoc cref="SerDes{T, TJVM}.Serialize(string, T)"/>
    public override byte[] Serialize(string topic, T data)
    {
        return SerializeWithHeaders(topic, null, data);
    }
    /// <inheritdoc cref="SerDes{T, TJVM}.SerializeWithHeaders(string, Headers, T)"/>
    public override byte[] SerializeWithHeaders(string topic, Headers headers, T data)
    {
        headers?.Add(KEFCoreSerDesNames.KeyTypeIdentifier, keyTypeName);
        headers?.Add(KEFCoreSerDesNames.KeySerializerIdentifier, customSerDesName);

        var jsonStr = System.Text.Json.JsonSerializer.Serialize<T>(data);
        return Encoding.UTF8.GetBytes(jsonStr);
    }
    /// <inheritdoc cref="SerDes{T, TJVM}.Deserialize(string, TJVM)"/>
    public override T Deserialize(string topic, byte[] data)
    {
        return DeserializeWithHeaders(topic, null, data);
    }
    /// <inheritdoc cref="SerDes{T, TJVM}.DeserializeWithHeaders(string, Headers, TJVM)"/>
    public override T DeserializeWithHeaders(string topic, Headers headers, byte[] data)
    {
        if (data == null || data.Length == 0) return default;
        return System.Text.Json.JsonSerializer.Deserialize<T>(data)!;
    }
}
public class CustomValueContainerSerDes<T> : SerDesRaw<T>
{
    readonly byte[] valueContainerSerDesName = Encoding.UTF8.GetBytes(typeof(CustomValueContainerSerDes<>).FullName!);
    readonly byte[] valueContainerName = null!;
    /// <summary>
    /// Default initializer
    /// </summary>
    public CustomValueContainerSerDes()
    {
        var tt = typeof(T);
        if (tt.IsGenericType)
        {
            var keyT = tt.GetGenericArguments();
            if (keyT.Length != 1) { throw new ArgumentException($"{typeof(T).Name} does not contains a single generic argument and cannot be used because it is not a valid ValueContainer type"); }
            var t = tt.GetGenericTypeDefinition();
            if (t.GetInterface(typeof(IValueContainer<>).Name) != null)
            {
                valueContainerName = Encoding.UTF8.GetBytes(t.FullName!);
                return;
            }
            else throw new ArgumentException($"{typeof(T).Name} does not implement IValueContainer<> and cannot be used because it is not a valid ValueContainer type");
        }
        throw new ArgumentException($"{typeof(T).Name} is not a generic type and cannot be used as a valid ValueContainer type");
    }

    /// <inheritdoc cref="SerDes{T, TJVM}.Serialize(string, T)"/>
    public override byte[] Serialize(string topic, T data)
    {
        return SerializeWithHeaders(topic, null, data);
    }
    /// <inheritdoc cref="SerDes{T, TJVM}.SerializeWithHeaders(string, Headers, T)"/>
    public override byte[] SerializeWithHeaders(string topic, Headers headers, T data)
    {
        headers?.Add(KEFCoreSerDesNames.ValueContainerSerializerIdentifier, valueContainerSerDesName);
        headers?.Add(KEFCoreSerDesNames.ValueContainerIdentifier, valueContainerName);

        var jsonStr = System.Text.Json.JsonSerializer.Serialize<T>(data);
        return Encoding.UTF8.GetBytes(jsonStr);
    }
    /// <inheritdoc cref="SerDes{T, TJVM}.Deserialize(string, TJVM)"/>
    public override T Deserialize(string topic, byte[] data)
    {
        return DeserializeWithHeaders(topic, null, data);
    }
    /// <inheritdoc cref="SerDes{T, TJVM}.DeserializeWithHeaders(string, Headers, TJVM)"/>
    public override T DeserializeWithHeaders(string topic, Headers headers, byte[] data)
    {
        if (data == null || data.Length == 0) return default;
        return System.Text.Json.JsonSerializer.Deserialize<T>(data)!;
    }
}

How to use custom types

KafkaDbContext contains three properties can be used to override the default types:

  • KeySerializationType: set the value of the Key SerDes type in the form CustomSerDes<>
  • ValueSerializationType: set the value of the ValueContainer SerDes type in the form CustomSerDes<>
  • ValueContainerType: set the value of the ValueContainer type in the form CustomValueContainer<>

IMPORTANT NOTE: the type applied in the previous properties of KafkaDbContext shall be a generic type definition, Entity Framework Core provider for Apache Kafka will apply the right generic type when needed.

Avro serialization

With package MASES.EntityFrameworkCore.KNet.Serialization.Avro an user can choose two different Avro serializers: The engine comes with two different encoders

  • Binary
  • Json

Both Key and ValueContainer SerDes, Binary or Json, come with two kind of data transfer mechanisms:

  • Raw: it uses byte arrays for data transfer
  • Buffered: they use ByteBuffer for data transfer

Avro schema

The following schema is the default used from the engine and can be registered in Apache Schema registry so other applications can use it to extract the data stored in the topics:

  • Complex Primary Key schema:

    {
      "namespace": "MASES.EntityFrameworkCore.KNet.Serialization.Avro.Storage",
      "type": "record",
      "name": "AvroKeyContainer",
      "doc": "Represents the storage container type to be used from KEFCore for keys",
      "fields": [
      	{
      		"name": "PrimaryKey",
      		"type": {
      			"type": "array",
      			"items": [
      				"null",
      				"boolean",
      				"int",
      				"long",
      				"float",
      				"double",
      				"string"
      			]
      		}
      	}
      ]
    }
    
  • ValueContainer schema:

    {
      "namespace": "MASES.EntityFrameworkCore.KNet.Serialization.Avro.Storage",
      "type": "record",
      "name": "AvroValueContainer",
      "doc": "Represents the storage container type to be used from KEFCore",
      "fields": [
      	{
      		"name": "EntityName",
      		"type": "string"
      	},
      	{
      		"name": "ClrType",
      		"type": "string"
      	},
      	{
      		"name": "Data",
      		"type": {
      			"type": "array",
      			"items": {
      				"namespace": "MASES.EntityFrameworkCore.KNet.Serialization.Avro.Storage",
      				"type": "record",
      				"name": "PropertyDataRecord",
      				"doc": "Represents the single container for Entity properties stored in AvroValueContainer and used from KEFCore",
      				"fields": [
      					{
      						"name": "PropertyIndex",
      						"type": "int"
      					},
      					{
      						"name": "PropertyName",
      						"type": "string"
      					},
      					{
      						"name": "ClrType",
      						"type": "string"
      					},
      					{
      						"name": "Value",
      						"type": [
      							"null",
      							"boolean",
      							"int",
      							"long",
      							"float",
      							"double",
      							"string"
      						]
      					}
      				]
      			}
      		}
      	}
      ]
    }
    

The extension converted this schema into code to speedup the exection of serialization/deserialization operations.

How to use Avro

KafkaDbContext contains three properties can be used to override the default types:

  • KeySerializationType: set this value to AvroKEFCoreSerDes.Key.BinaryRaw<> or AvroKEFCoreSerDes.Key.JsonRaw<> or use AvroKEFCoreSerDes.DefaultKeySerialization (defaults to AvroKEFCoreSerDes.Key.BinaryRaw<>), both types automatically manages simple or complex Primary Key
  • ValueSerializationType: set this value to AvroKEFCoreSerDes.ValueContainer.BinaryRaw<> or AvroKEFCoreSerDes.ValueContainer.JsonRaw<> or use AvroKEFCoreSerDes.DefaultValueContainerSerialization (defaults to AvroKEFCoreSerDes.ValueContainer.BinaryRaw<>)
  • ValueContainerType: set this value to AvroValueContainerRaw<> or use AvroKEFCoreSerDes.DefaultValueContainer

An example is:

using (context = new BloggingContext()
{
    BootstrapServers = "KAFKA-SERVER:9092",
    ApplicationId = "MyAppid",
    DbName = "MyDBName",
    KeySerializationType = UseAvroBinary ? typeof(AvroKEFCoreSerDes.Key.BinaryRaw<>) : typeof(AvroKEFCoreSerDes.Key.JsonRaw<>),
    ValueContainerType = typeof(AvroValueContainer<>),
    ValueSerializationType = UseAvroBinary ? typeof(AvroKEFCoreSerDes.ValueContainer.BinaryRaw<>) : typeof(AvroKEFCoreSerDes.ValueContainer.JsonRaw<>),
})
{
	// execute stuff here
}

Protobuf serialization

With package MASES.EntityFrameworkCore.KNet.Serialization.Protobuf an user can choose the Protobuf serializer.

Both Key and ValueContainer SerDes come with two kind of data transfer mechanisms:

  • Raw: it uses byte arrays for data transfer
  • Buffered: they use ByteBuffer for data transfer

Protobuf schema

The following schema is the default used from the engine and can be registered in Apache Schema registry so other applications can use it to extract the data stored in the topics:

  • Common multitype value:

    // [START declaration]
    syntax = "proto3";
    package storage;
    
    import "google/protobuf/struct.proto";
    import "google/protobuf/timestamp.proto";
    // [END declaration]
    
    // [START java_declaration]
    option java_multiple_files = true;
    option java_package = "mases.entityframeworkcore.knet.serialization.protobuf";
    option java_outer_classname = "GenericValue";
    // [END java_declaration]
    
    // [START csharp_declaration]
    option csharp_namespace = "MASES.EntityFrameworkCore.KNet.Serialization.Protobuf.Storage";
    // [END csharp_declaration]
    
    // [START messages]
    // Our address book file is just one of these.
    message GenericValue {
      // The kind of value.
      oneof kind {
        // Represents a null value.
        google.protobuf.NullValue null_value = 1;
        // Represents a boolean value.
        bool bool_value = 2;
        // Represents a int value.
        int32 byte_value = 3;
        // Represents a int value.
        int32 short_value = 4;
        // Represents a int value.
        int32 int_value = 5;
        // Represents a long value.
        int64 long_value = 6;
        // Represents a float value.
        float float_value = 7;
        // Represents a double value.
        double double_value = 8;
        // Represents a string value.
        string string_value = 9;
        // Represents a Guid value.
        bytes guid_value = 10;
        // Represents a Timestamp value.
        google.protobuf.Timestamp datetime_value = 11;
        // Represents a Timestamp value.
        google.protobuf.Timestamp datetimeoffset_value = 12;
      }
    }
    // [END messages]
    
  • Complex Primary Key schema:

    // [START declaration]
    syntax = "proto3";
    package storage;
    
    import "GenericValue.proto";
    // [END declaration]
    
    // [START java_declaration]
    option java_multiple_files = true;
    option java_package = "mases.entityframeworkcore.knet.serialization.protobuf";
    option java_outer_classname = "KeyContainer";
    // [END java_declaration]
    
    // [START csharp_declaration]
    option csharp_namespace = "MASES.EntityFrameworkCore.KNet.Serialization.Protobuf.Storage";
    // [END csharp_declaration]
    
    // [START messages]
    message PrimaryKeyType {
      repeated GenericValue values = 1; 
    }
    
    // Our address book file is just one of these.
    message KeyContainer {
      PrimaryKeyType PrimaryKey = 1;
    }
    // [END messages]
    
  • ValueContainer schema:

    // [START declaration]
    syntax = "proto3";
    package storage;
    
    import "GenericValue.proto";
    // [END declaration]
    
    // [START java_declaration]
    option java_multiple_files = true;
    option java_package = "mases.entityframeworkcore.knet.serialization.protobuf";
    option java_outer_classname = "ValueContainer";
    // [END java_declaration]
    
    // [START csharp_declaration]
    option csharp_namespace = "MASES.EntityFrameworkCore.KNet.Serialization.Protobuf.Storage";
    // [END csharp_declaration]
    
    // [START messages]
    message PropertyDataRecord {
      int32 PropertyIndex = 1;
      string PropertyName = 2; 
      string ClrType = 3;
      GenericValue Value = 4;
    }
    
    // Our address book file is just one of these.
    message ValueContainer {
      string EntityName = 1;
      string ClrType = 2;
      repeated PropertyDataRecord Data = 3;
    }
    // [END messages]
    

The extension converted this schema into code to speedup the exection of serialization/deserialization operations.

How to use Protobuf

KafkaDbContext contains three properties can be used to override the default types:

  • KeySerializationType: set this value to ProtobufKEFCoreSerDes.Key.BinaryRaw<> or use ProtobufKEFCoreSerDes.DefaultKeySerialization, the type automatically manages simple or complex Primary Key
  • ValueSerializationType: set this value to ProtobufKEFCoreSerDes.ValueContainer.BinaryRaw<> or use ProtobufKEFCoreSerDes.DefaultValueContainerSerialization
  • ValueContainerType: set this value to ProtobufValueContainerRaw<> or use ProtobufKEFCoreSerDes.DefaultValueContainer

An example is:

using (context = new BloggingContext()
{
    BootstrapServers = "KAFKA-SERVER:9092",
    ApplicationId = "MyAppid",
    DbName = "MyDBName",
    KeySerializationType = typeof(ProtobufKEFCoreSerDes.Key<>),
    ValueContainerType = typeof(ProtobufValueContainer<>),
    ValueSerializationType = typeof(ProtobufKEFCoreSerDes.ValueContainer<>),
})
{
	// execute stuff here
}
Product Compatible and additional computed target framework versions.
.NET net6.0 is compatible.  net6.0-android was computed.  net6.0-ios was computed.  net6.0-maccatalyst was computed.  net6.0-macos was computed.  net6.0-tvos was computed.  net6.0-windows was computed.  net7.0 is compatible.  net7.0-android was computed.  net7.0-ios was computed.  net7.0-maccatalyst was computed.  net7.0-macos was computed.  net7.0-tvos was computed.  net7.0-windows was computed.  net8.0 is compatible.  net8.0-android was computed.  net8.0-browser was computed.  net8.0-ios was computed.  net8.0-maccatalyst was computed.  net8.0-macos was computed.  net8.0-tvos was computed.  net8.0-windows was computed. 
Compatible target framework(s)
Included target framework(s) (in package)
Learn more about Target Frameworks and .NET Standard.

NuGet packages

This package is not used by any NuGet packages.

GitHub repositories

This package is not used by any popular GitHub repositories.

Version Downloads Last updated
2.4.2 73 11/5/2024
2.4.1 84 9/20/2024
2.4.0 126 8/17/2024
2.3.5 66 8/1/2024
2.3.4 54 7/30/2024
2.3.3 113 7/2/2024
2.3.2 95 6/27/2024
2.3.1 100 6/24/2024
2.3.0 100 5/27/2024
2.2.0 82 5/18/2024
2.1.1 112 5/4/2024
2.1.0 117 3/1/2024
2.0.2 129 2/12/2024
2.0.1 112 1/27/2024
2.0.0 110 1/21/2024
1.1.0 205 11/25/2023