Skip to main content

Efficient custom mapping from data model to ViewModel/Dto or vice versa with LINQ support

How many lines of codes we waste on mapping from one source to a destination like DataModel to ViewModel.  There are well-known Mappers available to do the same but nothing can come close in terms of performance by using manual mapping just that we need to write at too many places or takes a huge chunk of code lines in our main codes, may be under controllers, services or where ever you prefer projection in code.

The idea is to develop a proper way to deal with mappings. Some time back I had written code for manual mapping under LINQ queries. Manual model mapping - LINQ projection technique, this still works great. The idea is to create a cleaner solution as a whole new package for handling Mapping which can support general mappings between models and support of LINQ projection to select the limited number of rows as required.

Primary objectives of the implementation.
- A Data model to view model/DTO/or any other.
- view model/DTO/or any other to a data model.
- Support of LINQ projection without selecting all columns from a table.
- Support of IEnumerable.

The above items are compelling enough to create mapping structure with performance benefits.

Let's start with creations of interfaces to support One-way mapper, Two-way mapper, and support of LINQ queries in both case.

Basically, all we need is creation of interfaces and implementation of mapping classes based on models.

A simple IMapper implementation which simply converts from one model to other. It is simple enough not to have any kind of explanation.

 public interface IMapper<TSource, TResult>  
 {  
   /// <summary>  
   /// Maps the specified model.  
   /// </summary>  
   /// <param name="model">The model.</param>  
   /// <returns>Projection/Mapping from <paramref name="model"/> to a new type.</returns>  
   TResult Map(TSource model);  
 }  


IQueryMapper is for supporting LINQ projection, it is a kind of extension of above.

 public interface IQueryMapper<TSource, TResult>  
 {  
   /// <summary>  
   /// Gets the expression for projection/mapping.  
   /// </summary>  
   /// <value>  
   /// The expression for projection/mapping.  
   /// </value>  
   Expression<Func<TSource, TResult>> Expression { get; }  
 }  


ITwoWayMapper, It is opposite of  IMapper by inheriting the feature. Like if IMapper is for User to UserDto, this would allow having UserDto to User.

 /// <summary>  
 /// Two way mapping for models.  
 /// </summary>  
 /// <typeparam name="TFirst">The type of the first.</typeparam>  
 /// <typeparam name="TSecond">The type of the second.</typeparam>  
 /// <seealso cref="IMapper{TFirst, TSecond}" />  
 public interface ITwoWayMapper<TFirst, TSecond>  
     : IMapper<TFirst, TSecond>  
 {  
   /// <summary>  
   /// Maps the specified model.  
   /// </summary>  
   /// <param name="model">The model.</param>  
   /// <returns>Projection/Mapping from <paramref name="model"/> to a new type.</returns>  
   TFirst Map(TSecond model);  
 }  


IQueryTwoWayMapper, is similar to ITwoWayMapper but with LINQ support.

 public interface IQueryTwoWayMapper<TFirst, TSecond>  
   : IQueryMapper<TFirst, TSecond>  
 {  
   /// <summary>  
   /// Gets the reverse projection/mapping expression.  
   /// </summary>  
   /// <value>  
   /// The reverse projection/mapping expression.  
   /// </value>  
   Expression<Func<TSecond, TFirst>> ReverseExpression { get; }  
 }  


All these above implementations use Interface Segregation Principle principle, we would see advantages for them soon.

Based on above interfaces, we can create extension methods to map it to expressions or System.Func implementation based on need. I have not entirely tested extension methods based on above interfaces but it is really simple to extend based on a need to do something else. This is one of the major parts of Mapping Structure.

 /// <summary>  
 /// Extension methods for mapper  
 /// </summary>  
 public static class ExtensionMapper  
 {  
   /// <summary>  
   /// Maps the model from <typeparamref name="TSource"/> to <typeparamref name="TResult"/>.  
   /// </summary>  
   /// <typeparam name="TSource">The type of the source.</typeparam>  
   /// <typeparam name="TResult">The type of the result.</typeparam>  
   /// <param name="mapper">The mapper.</param>  
   /// <param name="source">The source.</param>  
   /// <returns>Projection from <typeparamref name="TSource"/> to <typeparamref name="TResult"/>.</returns>  
   public static IEnumerable<TResult> Map<TSource, TResult>(  
     this ITwoWayMapper<TSource, TResult> mapper,  
     IEnumerable<TSource> source)  
   {  
     return source.Select(val => mapper.Map(val));  
   }  
   /// <summary>  
   /// Maps the model from <typeparamref name="TSource"/> to <typeparamref name="TResult"/>.  
   /// </summary>  
   /// <typeparam name="TSource">The type of the source.</typeparam>  
   /// <typeparam name="TResult">The type of the result.</typeparam>  
   /// <param name="source">The source.</param>  
   /// <param name="mapper">The mapper.</param>  
   /// <returns>Projection from <typeparamref name="TSource"/> to <typeparamref name="TResult"/>.</returns>  
   public static IEnumerable<TResult> Map<TSource, TResult>(  
     this IQueryable<TSource> source,  
     IQueryTwoWayMapper<TSource, TResult> mapper)  
   {  
     return source.Select(mapper.Expression);  
   }  
   /// <summary>  
   /// Maps the model from <typeparamref name="TSource"/> to <typeparamref name="TResult"/>.  
   /// </summary>  
   /// <typeparam name="TSource">The type of the source.</typeparam>  
   /// <typeparam name="TResult">The type of the result.</typeparam>  
   /// <param name="mapper">The mapper.</param>  
   /// <param name="source">The source.</param>  
   /// <returns>Projection from <typeparamref name="TSource"/> to <typeparamref name="TResult"/>.</returns>  
   public static IEnumerable<TResult> Map<TSource, TResult>(  
     this ITwoWayMapper<TResult, TSource> mapper,  
     IEnumerable<TSource> source)  
   {  
     return source.Select(val => mapper.Map(val));  
   }  
   /// <summary>  
   /// Maps the model from <typeparamref name="TSource"/> to <typeparamref name="TResult"/>.  
   /// </summary>  
   /// <typeparam name="TSource">The type of the source.</typeparam>  
   /// <typeparam name="TResult">The type of the result.</typeparam>  
   /// <param name="source">The source.</param>  
   /// <param name="mapper">The mapper.</param>  
   /// <returns>Projection from <typeparamref name="TSource"/> to <typeparamref name="TResult"/>.</returns>  
   public static IEnumerable<TResult> Map<TSource, TResult>(  
     this IQueryable<TSource> source,  
     IQueryTwoWayMapper<TResult, TSource> mapper)  
   {  
     return source.Select(mapper.ReverseExpression);  
   }  
 }  


Implementation

Sample DTO

 public class Book  
 {  
   public string Name {get;set;}  
   public string Publisher {get;set;}  
 }  
 public class BookDto  
 {  
   public string BookName {get;set;}  
   public string BookPublisher {get;set;}  
 }  

A simple one-way mapper implementation of IMapper based on above models.

 public class BookMapper  
      : IMapper<BookDto, Book>  
 {  
      public Book Map(BookDto model)  
         => new Book  
         {  
           Name = model.BookName,  
           Publisher = model.BookPublisher  
         };  
 }  


Usage

 public ActionResult AddBook(BookDto model)  
 {  
      var bookMapper = new BookMapper();  
      var bookDataModel = bookMapper.Map(model);  
      // Code to insert book.  
 }  


Two-way with IQueryTwoWayMapper sample mapper implementation.

 public class BookMapper  
   : ITwoWayMapper<Book, BookDto>  
     , IQueryTwoWayMapper<Book, BookDto>  
 {  
   public Book Map(BookDto model)  
         => new Book  
         {  
           Name = model.BookName,  
           Publisher = model.BookPublisher  
         };  
   public BookDto Map(Book model)  
       => new BookDto  
       {  
         BookName = model.Name,  
         BookPublisher = model.Publisher  
       };  
   public Expression<Func<BookDto, Book>> ReverseExpression  
       => model  
       => new Book  
       {  
         Name = model.BookName,  
         Publisher = model.BookPublisher  
       };  
   public Expression<Func<Book, BookDto>> Expression  
       => model  
       => new BookDto  
       {  
         BookName = model.Name,  
         BookPublisher = model.Publisher  
       };  
 }  


With above codes, it is a bit confusing why similar type of codes are repeated for function and expression implementation. The thing is System.Func and general function cannot be truly converted to expression easily. If we directly try to pass Func than it would execute on CLR rather than SQL level which would result in the selection of all columns and then the projection. With Expressions, only required columns can be selected.

Also, Expression cannot be simply converted to Func since those can be really complex with multi-level of Expression Tree. If you still need a cleaner way and ready to compromise performance than this line can be used with expression compilation but I won't recommend going in that way.

 public BookDto Map(Book model) => Expression.Compile().Invoke(model);  


Implementation of above

The separation of interface allows us to use in any combination. Use, IMapper with IQueryMapper in a case of one-way mapping with LINQ queries. IMapper could be individually used like in the first example and in above we used all interface but just by inheriting two of them which depends upon their root inheritance.

 public ActionResult AddBook(BookDto model)  
 {  
      var bookMapper = new BookMapper();  
      var bookDataModel = bookMapper.Map(model);  
      // Projection from Context Book data to BookDto through Expression by using extension method.  
      var bookDtos = BookRepository.All().Select(bookMapper.Map());  
 }  


I have used the implementation as a concept but use as per your requirement, altering codes and further separation/removal of interface and extra implementation of Extension methods.

Comments

Popular posts from this blog

Elegantly dealing with TimeZones in MVC Core / WebApi

In any new application handling TimeZone/DateTime is mostly least priority and generally, if someone is concerned then it would be handled by using DateTime.UtcNow on codes while creating current dates and converting incoming Date to UTC to save on servers. Basically, the process is followed by saving DateTime to UTC format in a database and keep converting data to native format based on user region or single region in the application's presentation layer. The above is tedious work and have to be followed religiously. If any developer misses out the manual conversion, then that area of code/view would not work. With newer frameworks, there are flexible ways to deal/intercept incoming or outgoing calls to simplify conversion of TimeZones. These are steps/process to achieve it. 1. Central code for storing user's state about TimeZone. Also, central code for conversion logic based on TimeZones. 2. Dependency injection for the above class to be able to use global

Using Redis distributed cache in dotnet core with helper extension methods

Redis cache is out process cache provider for a distributed environment. It is popular in Azure Cloud solution, but it also has a standalone application to operate upon in case of small enterprises application. How to install Redis Cache on a local machine? Redis can be used as a local cache server too on our local machines. At first install, Chocolatey https://chocolatey.org/ , to make installation of Redis easy. Also, the version under Chocolatey supports more commands and compatible with Official Cache package from Microsoft. After Chocolatey installation hit choco install redis-64 . Once the installation is done, we can start the server by running redis-server . Distributed Cache package and registration dotnet core provides IDistributedCache interface which can be overrided with our own implementation. That is one of the beauties of dotnet core, having DI implementation at heart of framework. There is already nuget package available to override IDistributedCache i

Making FluentValidation compatible with Swagger including Enum or fixed List support

FluentValidation is not directly compatible with Swagger API to validate models. But they do provide an interface through which we can compose Swagger validation manually. That means we look under FluentValidation validators and compose Swagger validator properties to make it compatible. More of all mapping by reading information from FluentValidation and setting it to Swagger Model Schema. These can be done on any custom validation from FluentValidation too just that proper schema property has to be available from Swagger. Custom validation from Enum/List values on FluentValidation using FluentValidation.Validators; using System.Collections.Generic; using System.Linq; using static System.String; /// <summary> /// Validator as per list of items. /// </summary> /// <seealso cref="PropertyValidator" /> public class FixedListValidator : PropertyValidator { /// <summary> /// Gets the valid items /// <

Handling JSON DateTime format on Asp.Net Core

This is a very simple trick to handle JSON date format on AspNet Core by global settings. This can be applicable for the older version as well. In a newer version by default, .Net depends upon Newtonsoft to process any JSON data. Newtonsoft depends upon Newtonsoft.Json.Converters.IsoDateTimeConverter class for processing date which in turns adds timezone for JSON data format. There is a global setting available for same that can be adjusted according to requirement. So, for example, we want to set default formatting to US format, we just need this code. services.AddMvc() .AddJsonOptions(options => { options.SerializerSettings.DateTimeZoneHandling = "MM/dd/yyyy HH:mm:ss"; });

Kendo MVC Grid DataSourceRequest with AutoMapper

Kendo Grid does not work directly with AutoMapper but could be managed by simple trick using mapping through ToDataSourceResult. The solution works fine until different filters are applied. The problems occurs because passed filters refer to view model properties where as database model properties are required after AutoMapper is implemented. So, the plan is to intercept DataSourceRequest  and modify names based on database model. To do that we are going to create implementation of  CustomModelBinderAttribute to catch calls and have our own implementation of DataSourceRequestAttribute from Kendo MVC. I will be using same source code from Kendo but will replace column names for different criteria for sort, filters, group etc. Let's first look into how that will be implemented. public ActionResult GetRoles([MyDataSourceRequest(GridId.RolesUserGrid)] DataSourceRequest request) { if (request == null) { throw new ArgumentNullExce

Trim text in MVC Core through Model Binder

Trimming text can be done on client side codes, but I believe it is most suitable on MVC Model Binder since it would be at one place on infrastructure level which would be free from any manual intervention of developer. This would allow every post request to be processed and converted to a trimmed string. Let us start by creating Model binder using Microsoft.AspNetCore.Mvc.ModelBinding; using System; using System.Threading.Tasks; public class TrimmingModelBinder : IModelBinder { private readonly IModelBinder FallbackBinder; public TrimmingModelBinder(IModelBinder fallbackBinder) { FallbackBinder = fallbackBinder ?? throw new ArgumentNullException(nameof(fallbackBinder)); } public Task BindModelAsync(ModelBindingContext bindingContext) { if (bindingContext == null) { throw new ArgumentNullException(nameof(bindingContext)); } var valueProviderResult = bindingContext.ValueProvider.GetValue(bin

Kendo MVC Grid DataSourceRequest with AutoMapper - Advance

The actual process to make DataSourceRequest compatible with AutoMapper was explained in my previous post  Kendo MVC Grid DataSourceRequest with AutoMapper , where we had created custom model binder attribute and in that property names were changed as data models. In this post we will be looking into using AutoMapper's Queryable extension to retrieve the results based on selected columns. When  Mapper.Map<RoleViewModel>(data)  is called it retrieves all column values from table. The Queryable extension provides a way to retrieve only selected columns from table. In this particular case based on properties of  RoleViewModel . The previous approach that we implemented is perfect as far as this article ( 3 Tips for Using Telerik Data Access and AutoMapper ) is concern about performance where it states: While this functionality allows you avoid writing explicit projection in to your LINQ query it has the same fatal flaw as doing so - it prevents the query result from

OpenId Authentication with AspNet Identity Core

This is a very simple trick to make AspNet Identity work with OpenId Authentication. More of all both approach is completely separate to each other, there is no any connecting point. I am using  Microsoft.AspNetCore.Authentication.OpenIdConnect  package to configure but it should work with any other. Configuring under Startup.cs with IAppBuilder app.UseCookieAuthentication(new CookieAuthenticationOptions { AuthenticationScheme = CookieAuthenticationDefaults.AuthenticationScheme, LoginPath = new PathString("/Account/Login"), CookieName = "MyProjectName", }) .UseIdentity() .UseOpenIdConnectAuthentication(new OpenIdConnectOptions { ClientId = "<AzureAdClientId>", Authority = String.Format("https://login.microsoftonline.com/{0}", "<AzureAdTenant>"), ResponseType = OpenIdConnectResponseType.IdToken, PostLogoutRedirectUri = "<my website url>",

Data seed for the application with EF, MongoDB or any other ORM.

Most of ORMs has moved to Code first approach where everything is derived/initialized from codes rather than DB side. In this situation, it is better to set data through codes only. We would be looking through simple technique where we would be Seeding data through Codes. I would be using UnitOfWork and Repository pattern for implementing Data Seeding technique. This can be applied to any data source MongoDB, EF, or any other ORM or DB. Things we would be doing. - Creating a base class for easy usage. - Interface for Seed function for any future enhancements. - Individual seed classes. - Configuration to call all seeds. - AspNet core configuration to Seed data through Seed configuration. Creating a base class for easy usage public abstract class BaseSeed<TModel> where TModel : class { protected readonly IMyProjectUnitOfWork MyProjectUnitOfWork; public BaseSeed(IMyProjectUnitOfWork MyProjectUnitOfWork) { MyProject

MongoDB navigation property or making it behave as ORM in .Net

This is an implementation to make models to have  navigation properties work like ORM does for us. What actually happens in ORM to make navigation properties work? Entity Framework has proxy classes implementation to allow lazy loading and eager loading implementation. While creating proxy classes it also changes definitions for actual classes to make navigation properties work to get values based on Model's for navigation properties. Most of ORMs work in same fashion like Telerik DataAccess has enhancer tool which changes class definition at compile time to enable navigation properties. In this implementation, we would retain the original class but we would have extension methods to allow initializing properties to have navigation proprieties work. Let's first create desire model on which we need to implement. I am picking up simple one-to-many relationship example from Person to Address. public class Person { public int PersonId { get; set; }