Skip to main content

Data seed for the application with EF, MongoDB or any other ORM.

Most of ORMs has moved to Code first approach where everything is derived/initialized from codes rather than DB side. In this situation, it is better to set data through codes only. We would be looking through simple technique where we would be Seeding data through Codes.

I would be using UnitOfWork and Repository pattern for implementing Data Seeding technique. This can be applied to any data source MongoDB, EF, or any other ORM or DB.

Things we would be doing.
- Creating a base class for easy usage.
- Interface for Seed function for any future enhancements.
- Individual seed classes.
- Configuration to call all seeds.
- AspNet core configuration to Seed data through Seed configuration.

Creating a base class for easy usage

  
 public abstract class BaseSeed<TModel>  
      where TModel : class  
 {  
      protected readonly IMyProjectUnitOfWork MyProjectUnitOfWork;  
 
      public BaseSeed(IMyProjectUnitOfWork MyProjectUnitOfWork)  
      {  
           MyProjectUnitOfWork = MyProjectUnitOfWork;  
      }  
     
      protected void Save(TModel model, IRepository<TModel> repository)  
      {  
           repository.Create(model);  
           MyProjectUnitOfWork.Save();  
      }  
 }  

This would be just helpful for individual data seeds.

Interface for Seed function for any future enhancements


  
 public interface ISeed  
 {  
  
   void Seed();  
 }  


Individual seed classes


  
 public class LookupSeeder  
      : BaseSeed<Lookup>, ISeed  
 {  
     
      public LookupSeeder(IMyProjectUnitOfWork MyProjectUnitOfWork)  
           : base(MyProjectUnitOfWork)  
      {  
      }  
     
      public void Seed()  
      {  
           InsertState();  
           InsertLanguage();  
      }    
     
   private readonly Lazy<List<Lookup>> States = new Lazy<List<Lookup>>(() => new List<Lookup>  
     {  
       new Lookup {  
         LookupType = LookupType.State,  
         LookupValue= "ALABAMA",  
         LookupIdString= "AL"  
       },  
       new Lookup {  
         LookupType = LookupType.State,  
         LookupValue= "MyProject",  
         LookupIdString= "AK"  
       },  
       new Lookup {  
         LookupType = LookupType.State,  
         LookupValue= "MINNESOTA",  
         LookupIdString= "MN"  
       },  
       new Lookup {  
         LookupType = LookupType.State,  
         LookupValue= "MISSISSIPPI",  
         LookupIdString= "MS"  
       },  
       new Lookup {  
         LookupType = LookupType.State,  
         LookupValue= "NEVADA",  
         LookupIdString= "NV"  
       },  
       new Lookup {  
         LookupType = LookupType.State,  
         LookupValue= "NEW HAMPSHIRE",  
         LookupIdString= "NH"  
       },  
     }  
     );  
   /// <summary>  
   /// Initial inserts for <see cref="LookupType.State"/>  
   /// </summary>  
   private void InsertState()  
   {  
     var lookupRepo = MyProjectUnitOfWork.LookupRepository;  
     if (!lookupRepo.Any(lookup => lookup.LookupType == LookupType.State))  
     {  
       lookupRepo.Create(States.Value);  
       MyProjectUnitOfWork.Save();  
     }  
   }  
   #endregion " Initial records for State"  
   /// <summary>  
   /// Inserts the <see cref="LookupType.Language"/> lookups.  
   /// </summary>  
   private void InsertLanguage()  
   {  
     var lookupRepo = MyProjectUnitOfWork.LookupRepository;  
     if (!lookupRepo.Any(lookup => lookup.LookupType == LookupType.Language))  
     {  
       lookupRepo.Create(new List<Lookup> {  
         new Lookup {  
           LookupType = LookupType.Language,  
           LookupIdString = "ENGLISH",  
           LookupValue = "ENGLISH"  
         },  
         new Lookup {  
           LookupType = LookupType.Language,  
           LookupIdString = "SPANISH",  
           LookupValue = "SPANISH"  
         },  
       });  
       MyProjectUnitOfWork.Save();  
     }  
   }   
 }       

If you see, these are pretty simple queries statements. There are two major things, checking if any data exists and then inserting based on condition. In a similar pattern, we can have multiple conditions and insert statements with multiple classes based on Domain/Data models.

Configuration to call all seeds

This is just a simple class that would call all individual Seeder classes based on domain models.

 /// <summary>  
 /// Database seeding configuration.  
 /// </summary>  
 public static class DataSeederConfig  
 {  
      /// <summary>  
      /// Seeds the database.  
      /// </summary>  
      /// <param name="serviceCollection">The service collection.</param>  
      public static void SeedDatabase(this IServiceProvider serviceCollection)  
      {  
           var uow = serviceCollection.GetService<IMyProjectUnitOfWork>();  
           Seed(uow);  
      }  
      /// <summary>  
      /// Seeds the specified MyProject unit of work.  
      /// </summary>  
      /// <param name="myProjectUnitOfWork">The MyProject unit of work.</param>  
      private static void Seed(IMyProjectUnitOfWork myProjectUnitOfWork)  
      {  
           // Order is important as id generation are dependent on each other.  
           // Parent to child item is followed.  
           new LookupSeederSql(myProjectUnitOfWork).Seed();  
           myProjectUnitOfWork.Save();  
      }  
 }  

AspNet core configuration to Seed data through Seed configuration

Now, we are all set for the final piece. There is only one endpoint to configure entire Data seeds through DataSeederConfig class having extension method SeedDatabase.

This has to be configured on Startup.cs

 public void Configure(IApplicationBuilder app, IHostingEnvironment env, ILoggerFactory loggerFactory)  
 {  
      if (env.IsDevelopment())  
      {  
           using (var serviceScope = app.ApplicationServices.GetRequiredService<IServiceScopeFactory>()  
                          .CreateScope())  
           {  
                serviceScope.ServiceProvider.GetService<MyProjectContext>()  
                      .Database.Migrate();  
                serviceScope.ServiceProvider.SeedDatabase();  
           }  
           app.UseBrowserLink();  
           app.UseDeveloperExceptionPage();  
      }  
 }  

The above code would do DB migration and seeding of data in DB.


Popular posts from this blog

LDAP with ASP.Net Identity Core in MVC with project.json

Lightweight Directory Access Protocol (LDAP), the name itself explain it. An application protocol used over an IP network to access the distributed directory information service.

The first and foremost thing is to add references for consuming LDAP. This has to be done by adding reference from Global Assembly Cache (GAC) into project.json

"frameworks": { "net461": { "frameworkAssemblies": { "System.DirectoryServices": "4.0.0.0", "System.DirectoryServices.AccountManagement": "4.0.0.0" } } },
These System.DirectoryServices and System.DirectoryServices.AccountManagement references are used to consume LDAP functionality.

It is always better to have an abstraction for irrelevant items in consuming part. For an example, the application does not need to know about PrincipalContext or any other dependent items from those two references to make it extensible. So, we can begin with some bas…

Handling JSON DateTime format on Asp.Net Core

This is a very simple trick to handle JSON date format on AspNet Core by global settings. This can be applicable for the older version as well.

In a newer version by default, .Net depends upon Newtonsoft to process any JSON data. Newtonsoft depends upon Newtonsoft.Json.Converters.IsoDateTimeConverter class for processing date which in turns adds timezone for JSON data format.

There is a global setting available for same that can be adjusted according to requirement. So, for example, we want to set default formatting to US format, we just need this code.


services.AddMvc() .AddJsonOptions(options => { options.SerializerSettings.DateTimeZoneHandling = "MM/dd/yyyy HH:mm:ss"; });



Architecture solution composting Repository Pattern, Unit Of Work, Dependency Injection, Factory Pattern and others

Project architecture is like garden, we plant the things in certain order and eventually they grow in similar manner. If things are planted well then they will all look(work) great and easier to manage. If they grow as cumbersome it would difficult to maintain and with time more problems would be happening in maintenance.

There is no any fixed or known approach to decide project architecture and specially with Agile Methodology. In Agile Methodology, we cannot predict how our end products will look like similarly we cannot say a certain architecture will fit well for entire development lifespan for project. So, the best thing is to modify the architecture as per our application growth. I understand that it sounds good but will be far more problematic with actual development. If it is left as it is then more problems will arise with time. Just think about moving plant vs a full grown tree.

Coming to technical side, In this article, I will be explaining about the various techniques tha…

Unit Of Work injection through Asp.Net Core Dependency Injection

This article is not directly related to UnitOfWork but leveraging Asp.Net Core Dependency Injection to consume Unit Of Work.

In one of the previous article about project architecture, I was not very satisfied with the approach for Unit Of Work implementation for initialization of repository even if with some advantage.

Here is old code for UnitOfWork.

public sealed partial class MyProjectUnitOfWork : UnitOfWork<DbContext>, IMyProjectUnitOfWork { public MyProjectUnitOfWork(IContextFactory<DbContext> contextFactory) : base(contextFactory) { } /// <summary> /// BookRepository holder /// </summary> private MyProject.DB.Repository.BookRepository _bookRepository; /// <summary> /// Gets the BookRepository repository. /// </summary> /// <value> /// The BookRepository repository. /// </value> MyProject.Interface.Repository.IBoo…

Global exception handling and custom logging in AspNet Core with MongoDB

In this, we would be looking into logging and global exception handling in the AspNet Core application with proper registration of logger and global exception handling.

Custom logging
The first step is to create a data model that we want to save into DB.

Error log Data model
These are few properties to do logging which could be extended or reduced based on need.

public class ErrorLog { /// <summary> /// Gets or sets the Error log identifier. /// </summary> /// <value> /// The Error log identifier. /// </value> [BsonRepresentation(BsonType.ObjectId)] public ObjectId Id { get; set; /// <summary> /// Gets or sets the date. /// </summary> /// <value> /// The date. /// </value> public DateTime Date { get; set; } /// <summary> /// Gets or sets the thread. /// </summary> /// <v…

Configuring Ninject, Asp.Net Identity UserManager, DataProtectorTokenProvider with Owin

It can be bit tricky to configure both Ninject and Asp.Net Identity UserManager if some value is expected from DI to configure UserManager. We will look into configuring both and also use OwinContext to get UserManager.

As usual, all configuration need to be done on Startup.cs. It is just a convention but can be used with different name, the important thing is to decorate class with following attribute to make it Owin start-up:

[assembly: OwinStartup(typeof(MyProject.Web.Startup))]
Ninject configuration

Configuring Ninject kernel through method which would be used to register under Owin.

Startup.cs
public IKernel CreateKernel() { var kernel = new StandardKernel(); try { //kernel.Bind<IHttpModule>().To<HttpApplicationInitializationHttpModule>(); // TODO: Put any other injection which are required. return kernel; } catch { kernel.Dispose(); throw; }…

OpenId Authentication with AspNet Identity Core

This is a very simple trick to make AspNet Identity work with OpenId Authentication. More of all both approach is completely separate to each other, there is no any connecting point.

I am using Microsoft.AspNetCore.Authentication.OpenIdConnect package to configure but it should work with any other.

Configuring under Startup.cs with IAppBuilder
app.UseCookieAuthentication(new CookieAuthenticationOptions { AuthenticationScheme = CookieAuthenticationDefaults.AuthenticationScheme, LoginPath = new PathString("/Account/Login"), CookieName = "MyProjectName", }) .UseIdentity() .UseOpenIdConnectAuthentication(new OpenIdConnectOptions { ClientId = "<AzureAdClientId>", Authority = String.Format("https://login.microsoftonline.com/{0}", "<AzureAdTenant>"), ResponseType = OpenIdConnectResponseType.IdToken, PostLogoutRedirectUri = "<my website url>", Au…

Kendo MVC Grid DataSourceRequest with AutoMapper

Kendo Grid does not work directly with AutoMapper but could be managed by simple trick using mapping through ToDataSourceResult. The solution works fine until different filters are applied.
The problems occurs because passed filters refer to view model properties where as database model properties are required after AutoMapper is implemented.
So, the plan is to intercept DataSourceRequest  and modify names based on database model. To do that we are going to create implementation of CustomModelBinderAttribute to catch calls and have our own implementation of DataSourceRequestAttribute from Kendo MVC. I will be using same source code from Kendo but will replace column names for different criteria for sort, filters, group etc.
Let's first look into how that will be implemented.
public ActionResult GetRoles([MyDataSourceRequest(GridId.RolesUserGrid)] DataSourceRequest request) { if (request == null) { throw new ArgumentNullException("reque…

Centralized model validation both for MVC/WebApi and SPA client-side validation using FluentValidation

Validation is one of the crucial parts of any application. It has to validate on both client side and server side requests.
What are target features or implementation from this article?Model validation for any given model.Centralized/One code for validation on both server-side and client-side.Automatic validation of model without writing any extra codes on/under actions for validation. NO EXTRA/ANY codes on client-side to validate any form.Compatible with SPA.Can be compatible with any client-side validation framework/library. Like Angular Reactive form validation or any jquery validation libraries. Tools used in the implementation?FluentValidation: I feel DataAnnotation validation are excellent and simple to use, but in case of complex validation or writing any custom validations are always tricker and need to write a lot of codes to achieve whereas FluentValidations are simple even in case of complex validation. Generally, we need to validate incoming input against database values,…

Strongly typed SingalR on server and client end through TypeScript.

SignalR is a very flexible framework to create bidirectional communication between client and server. The general process that is followed won't allow to have strongly typed calls on the server and also in client code. We will look into how to make strongly typed calls on the server through interface and client end through TypeScript.

SignalR - making strong type on server side code
Since client-side methods are very dynamic in nature, so calling those from server side behaves similarly to allow any calls.

SignalR Hubs are derived from Microsoft.AspNet.SignalR.Hub class, there is also a generic version available to follow typed items.

Ex:

The interface is the replication of possible calls that would be received on the client end and calling of client-side methods on server code.

/// <summary> /// Client(JS) side chatting interface callbacks. /// </summary> public interface IChatHub { /// <summary> /// Gets the online users. ///…