Skip to main content

A wrapper implementation for Kendo Grid usage

A wrapper implementation for any heavily used item is always a good practice. Whatever is not written by us and used at a lot of places should be wrapped within specific functionality to keep it future proof and easily changeable. This also encourages DRY principle to keep our common setting at a central place.

Kendo UI items are enormous in configuration, one of an issue I find people keep repeating codes for Kendo Grid configuration. They have built very flexible system to have any configuration, but in most of the cases, we do not need all of those complicated configuration. We would try to see a simpler configuration of same. The actual core implementation is bit complex, but we do not have to bother about it once done since the focus is just on usage only.

I recommend doing this practice for as simple as jQuery events, form handling or as simple as any notification system. This just won't make things simple but makes codes much more manageable, easy understand, read or open for future enhancements.

The core implementation:GridWrapper.ts

 import * as _ from "lodash";  
 class GridWrapper {  
   constructor() {  
   }  
   InitGrid(gridInitOption: {  
     GridSelector: string;  
     GridOption?: (opt: kendo.ui.GridOptions) => void;  
     DataSource?: (dsOption: kendo.data.DataSourceOptions) => void;  
     GridColumns: IAppKendoColumn[];  
     DataSourceUrl?: IDataSourceUrlOption;  
     Width?: number;   
   }): kendo.ui.Grid {  
     let baseUrl = '';  // TODO: Base URL ideally should be done on creation of instance
    // Default setting for Kendo Grid. 
    let gridOptions: kendo.ui.GridOptions = {  
       columns: gridInitOption.GridColumns,  
       sortable: true,  
       noRecords: {  
         template: 'No records found'  
       },  
       pageable: {  
         refresh: true,  
         buttonCount: 5  
       },  
     };  
     // Callback for grid option, in case further customization by callee 
     typeof gridInitOption.GridOption == 'function' && gridInitOption.GridOption(gridOptions);  
     // DataSource configuration  
     let dataSource = this.GetDefaultDataSource();  
     this.ConfigureColumnDataSource(gridInitOption.GridColumns, dataSource);  
     this.ConfigureTransport(gridInitOption.DataSourceUrl, dataSource);  
     // Callback for data source, in case if further change required.
     typeof gridInitOption.DataSource == 'function' && gridInitOption.DataSource(dataSource);  
     gridOptions.dataSource = new kendo.data.DataSource(dataSource);  
     let grid = $(gridInitOption.GridSelector).kendoGrid(gridOptions);  
     return grid.data('kendoGrid');  
   }  
   private GetDefaultDataSource(): kendo.data.DataSourceOptions {  
     // Default setting for Kendo DataSource
     return {  
       // This is for Kendo MVC but can be removed in case of Kendo UI
       type: (() => {  
         if (kendo.data.transports['aspnetmvc-ajax']) {  
           return 'aspnetmvc-ajax';  
         } else {  
           throw new Error('The kendo.aspnetmvc.min.js script is not included.');  
         }  
       })(),  
       transport: {},  
       serverPaging: true,  
       serverSorting: true,  
       pageSize: 10,  
       schema: {  
         data: 'Data',  
         total: 'Total',  
         errors: 'Errors',  
         model: {  
           fields: {}  
         }  
       },  
       error: (e) => {  
         // Handling error as requied.
         this.ParseModelStateError(e.xhr,  
           (errMessage, propName) => alert(errMessage)); // TODO: Hook with notification system  
       }  
     };  
   }  
   GetDataItem(gridSelector: string, evt) {  
     var dataItem = $(gridSelector).data('kendoGrid').dataItem($(evt.currentTarget).closest("tr"));  
     return dataItem;  
   }  
   private ConfigureTransport(urlOption: IDataSourceUrlOption, datasource: kendo.data.DataSourceOptions) {  
     if (urlOption == undefined) {  
       return;  
     }  
     // Datasource setting for all URLs required.
     datasource.transport.read = typeof urlOption.read == 'string' ?  
       {  
         url: baseUrl + urlOption.read,  
         type: 'POST'  
       } : urlOption.read;  
     datasource.transport.create = typeof urlOption.create == 'string' ?  
       {  
         url: baseUrl + urlOption.create,  
         type: 'POST'  
       } : urlOption.create;  
     datasource.transport.update = typeof urlOption.update == 'string' ?  
       {  
         url: baseUrl + urlOption.update,  
         type: 'POST'  
       } : urlOption.update;  
     datasource.transport.destroy = typeof urlOption.destroy == 'string' ?  
       {  
         url: baseUrl + urlOption.destroy,  
         type: 'DELETE'  
       } : urlOption.destroy;  
   }  
   private ConfigureColumnDataSource(columns: IAppKendoColumn[], dataSource: kendo.data.DataSourceOptions) {  
     _.forEach(columns, col => {  
       if (_.has(col, 'field') && _.has(col, 'dataSet')) {  
         let fieldName = col.field;  
         let dsFieldOption = col.dataSet;  
         dataSource.schema.model.fields[fieldName] = dsFieldOption;  
         // Primary key  
         let primaryKetSet = false;  
         // Customization of as per our need to make our life easier.
         // this is direct setting based on IAppKendoColumn for data source setting. 
         if (!primaryKetSet && _.has(dsFieldOption, 'isPrimaryKey')) {  
           primaryKetSet = true;  
           dataSource.schema.model.id = fieldName;  
           delete dsFieldOption.isPrimaryKey;  
         }  
         // Default sort  
         let defaultSortDone = false;  
         // Default sort settings
         if (!defaultSortDone && _.has(dsFieldOption, 'defaultSortOrder')) {  
           defaultSortDone = true;  
           dataSource.sort = { field: fieldName, dir: dsFieldOption.defaultSortOrder };  
           delete dsFieldOption.defaultSortOrder;  
         }  
         delete col.dataSet;  
       }  
     });  
   }  
   ParseModelStateError(data: JQuery.jqXHR<any>, eachErrorCallback: (message: string, propName: string) => void) {  
     if (data == undefined || data.responseJSON == undefined) {  
       return;  
     }  
     var message = '';  
     // Error setting initialization.
     var propStrings = Object.keys(data.responseJSON);  
     $.each(propStrings, (errIndex, propString) => {  
       var propErrors = data.responseJSON[propString];  
       $.each(propErrors, (errMsgIndex, propError) => {  
         message += propError;  
       });  
       message += '\n';  
       eachErrorCallback(message, propString);  
       message = '';  
     });  
   }  
 }  
 interface IAppKendoColumn  
   extends kendo.ui.GridColumn {  
   dataSet?: IDataSetOption;  
   command?: any;  
 }  
 interface IDataSourceUrlOption {  
   read: kendo.data.DataSourceTransportRead | string;  
   create?: kendo.data.DataSourceTransportCreate | string;  
   destroy?: kendo.data.DataSourceTransportDestroy | string;  
   update?: kendo.data.DataSourceTransportUpdate | string;  
 }  
 interface IDataSetOption {  
   isPrimaryKey?: boolean;  
   defaultSortOrder?: SortOrder;  
   type?: DataType;  
   editable?: boolean;  
   validation?: any;  
   defaultValue?: any;  
 }  
 enum SortOrder {  
   asc = 'asc',  
   desc = 'desc'  
 }  
 enum DataType {  
   Number = 'number',  
   String = 'string',  
   Boolean = 'boolean',  
   Date = 'date'  
 }  
 export { KendoGrid, IAppKendoColumn, SortOrder, DataType, IDataSetOption }  

Usage

Once a common code is completed, usage is much more straightforward. In one go everything can be set. Along with such a small setting, we can utilize complex one through callbacks or include more properties.

 var grid = new Grid(); // In my case I had used DI, but you can use direct method call  
 grid.InitGrid({  
      GridSelector: '#UserGrid',  
      GridColumns: [  
           {  
                field: 'Id', title: 'User Id', hidden: true,  
                dataSet: { isPrimaryKey: true, type: DataType.Number }  
           },  
           {  
                field: 'UserName', title: 'User', dataSet: { type: DataType.String }  
           },  
           {  
                field: 'EmailId', title: 'Email', dataSet: { type: DataType.String }  
           },  
           {  
                field: 'IsActive', title: 'Is User Active',  
                dataSet: { type: DataType.Boolean, defaultSortOrder: SortOrder.desc }  
           },  
      ],  
      Url: 'users/list',  
 });  

This is just an example of coding practice, but this kind of approach shall be used at any places where ever we are relying heavily on single type of usage.
Happy coding!


Comments

Popular posts from this blog

Elegantly dealing with TimeZones in MVC Core / WebApi

In any new application handling TimeZone/DateTime is mostly least priority and generally, if someone is concerned then it would be handled by using DateTime.UtcNow on codes while creating current dates and converting incoming Date to UTC to save on servers. Basically, the process is followed by saving DateTime to UTC format in a database and keep converting data to native format based on user region or single region in the application's presentation layer. The above is tedious work and have to be followed religiously. If any developer misses out the manual conversion, then that area of code/view would not work. With newer frameworks, there are flexible ways to deal/intercept incoming or outgoing calls to simplify conversion of TimeZones. These are steps/process to achieve it. 1. Central code for storing user's state about TimeZone. Also, central code for conversion logic based on TimeZones. 2. Dependency injection for the above class to ...

Trim text in MVC Core through Model Binder

Trimming text can be done on client side codes, but I believe it is most suitable on MVC Model Binder since it would be at one place on infrastructure level which would be free from any manual intervention of developer. This would allow every post request to be processed and converted to a trimmed string. Let us start by creating Model binder using Microsoft.AspNetCore.Mvc.ModelBinding; using System; using System.Threading.Tasks; public class TrimmingModelBinder : IModelBinder { private readonly IModelBinder FallbackBinder; public TrimmingModelBinder(IModelBinder fallbackBinder) { FallbackBinder = fallbackBinder ?? throw new ArgumentNullException(nameof(fallbackBinder)); } public Task BindModelAsync(ModelBindingContext bindingContext) { if (bindingContext == null) { throw new ArgumentNullException(nameof(bindingContext)); } var valueProviderResult = bindingContext.ValueProvider.GetValue(bin...

Handling JSON DateTime format on Asp.Net Core

This is a very simple trick to handle JSON date format on AspNet Core by global settings. This can be applicable for the older version as well. In a newer version by default, .Net depends upon Newtonsoft to process any JSON data. Newtonsoft depends upon Newtonsoft.Json.Converters.IsoDateTimeConverter class for processing date which in turns adds timezone for JSON data format. There is a global setting available for same that can be adjusted according to requirement. So, for example, we want to set default formatting to US format, we just need this code. services.AddMvc() .AddJsonOptions(options => { options.SerializerSettings.DateTimeZoneHandling = "MM/dd/yyyy HH:mm:ss"; });

Using Redis distributed cache in dotnet core with helper extension methods

Redis cache is out process cache provider for a distributed environment. It is popular in Azure Cloud solution, but it also has a standalone application to operate upon in case of small enterprises application. How to install Redis Cache on a local machine? Redis can be used as a local cache server too on our local machines. At first install, Chocolatey https://chocolatey.org/ , to make installation of Redis easy. Also, the version under Chocolatey supports more commands and compatible with Official Cache package from Microsoft. After Chocolatey installation hit choco install redis-64 . Once the installation is done, we can start the server by running redis-server . Distributed Cache package and registration dotnet core provides IDistributedCache interface which can be overrided with our own implementation. That is one of the beauties of dotnet core, having DI implementation at heart of framework. There is already nuget package available to override IDistributedCache i...

Architecture solution composting Repository Pattern, Unit Of Work, Dependency Injection, Factory Pattern and others

Project architecture is like garden, we plant the things in certain order and eventually they grow in similar manner. If things are planted well then they will all look(work) great and easier to manage. If they grow as cumbersome it would difficult to maintain and with time more problems would be happening in maintenance. There is no any fixed or known approach to decide project architecture and specially with Agile Methodology. In Agile Methodology, we cannot predict how our end products will look like similarly we cannot say a certain architecture will fit well for entire development lifespan for project. So, the best thing is to modify the architecture as per our application growth. I understand that it sounds good but will be far more problematic with actual development. If it is left as it is then more problems will arise with time. Just think about moving plant vs a full grown tree. Coming to technical side, In this article, I will be explaining about the various techniques ...

LDAP with ASP.Net Identity Core in MVC with project.json

Lightweight Directory Access Protocol (LDAP), the name itself explain it. An application protocol used over an IP network to access the distributed directory information service. The first and foremost thing is to add references for consuming LDAP. This has to be done by adding reference from Global Assembly Cache (GAC) into project.json "frameworks": { "net461": { "frameworkAssemblies": { "System.DirectoryServices": "4.0.0.0", "System.DirectoryServices.AccountManagement": "4.0.0.0" } } }, These  System.DirectoryServices  and  System.DirectoryServices.AccountManagement  references are used to consume LDAP functionality. It is always better to have an abstraction for irrelevant items in consuming part. For an example, the application does not need to know about PrincipalContext or any other dependent items from those two references to make it extensible. So, we can begin wi...

Blazor 0.5.0 Interop JavaScript from C# and C# to JavaScript call

Blazor provides two-way communication from JS to C# and C# to JS which is called Interop. The version upgrade to Blazor 0.5.0 changed the approach in interactions between cshtml and JS files. The newer version simplifies JS calling by avoiding pre-registration of JS function. In this article, we would see how to call a JS method by passing multiple parameters from cshtml and parameterized call from JS to cshtml. The scenario that is used in this article has a Kendo AutoComplete to search book and populate information based on the selection. An example of a call to JS from cshtml await JSRuntime.Current.InvokeAsync<string>( "searchBook.Init", "#SearchTitle", new DotNetObjectRef(this)); Takeaways from the above example: - JSRuntime.Current give environment to execute out-process JS. - InvokeAsync is a function that would allow executing the JS function. - The searchBook.Init is a JS function call, we would see it in details in the second sec...

Channel, ChannelReader and ChannelWriter to manage data streams in multi-threading environment

I came across Channel class while working with SignalR which looks really interesting. By looking into NuGet packages ( https://www.nuget.org/packages/System.Threading.Channels ), it seems just 4 months old. The Channel class provides infrastructure to have multiple reads and write simuletensely through it's Reader and Writer properties. This is where it is handy in case of SignalR where data streaming needs to be done but is not just limited to that but wherever something needs to be read/write/combination of both in a multi-threading environment. In my case with SignalR, I had to stream stock data at a regular interval of time. public ChannelReader<StockData> StreamStock() { var channel = Channel.CreateUnbounded<StockData>(); _stockManager.OnStockData = stockData => { channel.Writer.TryWrite(stockData); }; return channel.Reader; } The SignalR keeps return type of ChannelReader<StockData> open so that whatev...

Making FluentValidation compatible with Swagger including Enum or fixed List support

FluentValidation is not directly compatible with Swagger API to validate models. But they do provide an interface through which we can compose Swagger validation manually. That means we look under FluentValidation validators and compose Swagger validator properties to make it compatible. More of all mapping by reading information from FluentValidation and setting it to Swagger Model Schema. These can be done on any custom validation from FluentValidation too just that proper schema property has to be available from Swagger. Custom validation from Enum/List values on FluentValidation using FluentValidation.Validators; using System.Collections.Generic; using System.Linq; using static System.String; /// <summary> /// Validator as per list of items. /// </summary> /// <seealso cref="PropertyValidator" /> public class FixedListValidator : PropertyValidator { /// <summary> /// Gets the valid items /// <...

Centralized model validation both for MVC/WebApi and SPA client-side validation using FluentValidation

Validation is one of the crucial parts of any application. It has to validate on both client side and server side requests. What are target features or implementation from this article? Model validation for any given model. Centralized/One code for validation on both server-side and client-side. Automatic validation of model without writing any extra codes on/under actions for validation.  NO EXTRA/ANY codes on client-side to validate any form. Compatible with SPA. Can be compatible with any client-side validation framework/library. Like Angular Reactive form validation or any jquery validation libraries. Tools used in the implementation? FluentValidation : I feel DataAnnotation validation are excellent and simple to use, but in case of complex validation or writing any custom validations are always tricker and need to write a lot of codes to achieve whereas FluentValidations are simple even in case of complex validation. Generally, we need to validate inc...