Skip to main content

Displaying progress status in asynchronous programming

UX is most important in any application and responsive application is an add-on to make UI free by running long running task in background. Sometimes we need to show more useful information about long running task to keep user informed about progress. In this article, we will be looking into showing up progress status for long running process with Task Parallel Library (TPL).

In this particular example, we will be creating a Console application to re-size any amount of jpg files under directory with new image dimension but our main focus would be on showing progress status.

Prior to .Net 4.5 Framework version, there was no inbuilt mechanism to show status of task. If it was needed, then we need to create event handler on class that is processing the long running task with custom defined event argument for progress status class and then subscription need to be done on consuming class for getting progress information. This was better and tidy approach for getting progress status but lot of things were needed to make it work.

In 4.5, an interface named IProgress<T> is introduced having just single function Report<T value>. Due to generic by nature we could use custom defined classes or any on other built in types to show status. It is included in mscorlib.dll. http://msdn.microsoft.com/en-us/library/hh138298(v=vs.110).aspx

Let's roll and explore on the new approach.

We will be having our own custom class to show status of total files, skipped, processed and percentage of completion.

   public class ImageResizeProgress  
   {  
     public int TotalFiles { get; set; }  
     public int SkippedFiles { get; set; }  
     public int ProcessedFiles { get; set; }  
     public decimal Percentage  
     {  
       get  
       {  
         if ((TotalFiles - SkippedFiles) != 0)  
         {  
           return ProcessedFiles / (decimal)(TotalFiles - SkippedFiles);  
         }  
         return 1;  
       }  
     }  
   }  

The above class looks pretty simple without any kind of noise in it. In case of percentage there is no multiplication of 100 by we will be using ToString("p") to show up percentage value.

After completion of progress class, we need to populate classes member according to need on processing function for image re-sizing. If you look into below function, ImageResizeProgress class is wrapped into IProgress interface under parameter. After changing values of curProgress object, whenever showing status is required we just need to call progress.Report(curProgress), in which it will trigger function that we are going to create next.

 public Task ProcessFilesForImageDimension(string fileOrDirectory,  
       IProgress<ImageResizeProgress> progress)  
     {  
       var curProgress = new ImageResizeProgress();  
       // Delegate to resize image  
       Action<string> resize = (file) =>  
       {  
         Thread.Sleep(100);  
         if (ResizeImage(file))  
         {  
           curProgress.ProcessedFiles++;  
         }  
         else  
         {  
           curProgress.SkippedFiles++;  
         }  
         progress.Report(curProgress);  
       };  
       return Task.Factory.StartNew(() =>  
       {  
         if (Directory.Exists(fileOrDirectory))  
         {  
           var files = Directory.GetFiles(fileOrDirectory, "*.jpg");  
           curProgress.TotalFiles = files.Length;  
           progress.Report(curProgress);  
           foreach (var file in files)  
           {  
             resize(file);  
           }  
         }  
         else if (File.Exists(fileOrDirectory))  
         {  
           resize(fileOrDirectory);  
         }  
         else  
         {  
           progress.Report(curProgress);  
         }  
       });  
     }  

Let's create function which will display status on console. This static method will only take ImageResizeProgress as a argument. You can consume object to show status in your own way.

     private static void ShowStatus(ImageResizeProgress progress)  
     {  
       Console.Clear();  
       Console.SetCursorPosition(0, 0);  
       Console.WriteLine("Completion percentage: " + progress.Percentage.ToString("p"));  
       Console.WriteLine("Total Files: " + progress.TotalFiles);  
       Console.WriteLine("Processed Files: " + progress.ProcessedFiles);  
       Console.WriteLine("Skipped Files: " + progress.SkippedFiles);  
     }  

Now, everything is setup we just need to call ProcessFilesForImageDimension function to initiate operation. For IProgress interface we need to pass concrete class implementation which again is defined under mscorlib http://msdn.microsoft.com/en-us/library/hh194158(v=vs.110).aspx. The Progress class constructor takes action as parameter. In our case we have defined method named ShowStatus to pass in constructor.

 private static void Main(string[] args)  
     {  
       ImageDimensionManager dimensionManager = new ImageDimensionManager();  
       // TODO: Change location  
       Task tskImageResize = dimensionManager.ProcessFilesForImageDimension  
         (@"C:\Users\Public\Pictures\Sample Pictures",  
         new Progress<ImageResizeProgress>(ShowStatus));  
       tskImageResize.Wait();  
     }  

Now, everything is set we can expect result something like this.

Source code: https://www.dropbox.com/s/tf9vughitrj426s/AsyncProgessbar.zip

Comments

Popular posts from this blog

Elegantly dealing with TimeZones in MVC Core / WebApi

In any new application handling TimeZone/DateTime is mostly least priority and generally, if someone is concerned then it would be handled by using DateTime.UtcNow on codes while creating current dates and converting incoming Date to UTC to save on servers. Basically, the process is followed by saving DateTime to UTC format in a database and keep converting data to native format based on user region or single region in the application's presentation layer. The above is tedious work and have to be followed religiously. If any developer misses out the manual conversion, then that area of code/view would not work. With newer frameworks, there are flexible ways to deal/intercept incoming or outgoing calls to simplify conversion of TimeZones. These are steps/process to achieve it. 1. Central code for storing user's state about TimeZone. Also, central code for conversion logic based on TimeZones. 2. Dependency injection for the above class to ...

Handling JSON DateTime format on Asp.Net Core

This is a very simple trick to handle JSON date format on AspNet Core by global settings. This can be applicable for the older version as well. In a newer version by default, .Net depends upon Newtonsoft to process any JSON data. Newtonsoft depends upon Newtonsoft.Json.Converters.IsoDateTimeConverter class for processing date which in turns adds timezone for JSON data format. There is a global setting available for same that can be adjusted according to requirement. So, for example, we want to set default formatting to US format, we just need this code. services.AddMvc() .AddJsonOptions(options => { options.SerializerSettings.DateTimeZoneHandling = "MM/dd/yyyy HH:mm:ss"; });

Enum generation for lookup table through T4

Sometime there is need of mapping some values from database to code level, basically in look up tables. I am going to generate enum based on database values. In this example, I am going for T4 template to generate up enum by using SqlDataReader. We can have any SQL query to generate enum. In my case, Privilege code, name and description from table would generate up enum. <#@ template debug="true" hostSpecific="true" #> <#@ output extension=".cs" #> <#@ Assembly Name="System.Data" #> <#@ include file="EF.Utility.CS.ttinclude"#> <#@ import namespace="System.Data.SqlClient" #> <# var code = new CodeGenerationTools(this); var connectString = "Set connection string"; var queryString = "select PrivilegeCode, PrivilegeName, PrivilegeDescription from Privilege"; #> namespace <#= code.VsNamespaceSuggestion()#> { ...

Making FluentValidation compatible with Swagger including Enum or fixed List support

FluentValidation is not directly compatible with Swagger API to validate models. But they do provide an interface through which we can compose Swagger validation manually. That means we look under FluentValidation validators and compose Swagger validator properties to make it compatible. More of all mapping by reading information from FluentValidation and setting it to Swagger Model Schema. These can be done on any custom validation from FluentValidation too just that proper schema property has to be available from Swagger. Custom validation from Enum/List values on FluentValidation using FluentValidation.Validators; using System.Collections.Generic; using System.Linq; using static System.String; /// <summary> /// Validator as per list of items. /// </summary> /// <seealso cref="PropertyValidator" /> public class FixedListValidator : PropertyValidator { /// <summary> /// Gets the valid items /// <...

Kendo MVC Grid DataSourceRequest with AutoMapper

Kendo Grid does not work directly with AutoMapper but could be managed by simple trick using mapping through ToDataSourceResult. The solution works fine until different filters are applied. The problems occurs because passed filters refer to view model properties where as database model properties are required after AutoMapper is implemented. So, the plan is to intercept DataSourceRequest  and modify names based on database model. To do that we are going to create implementation of  CustomModelBinderAttribute to catch calls and have our own implementation of DataSourceRequestAttribute from Kendo MVC. I will be using same source code from Kendo but will replace column names for different criteria for sort, filters, group etc. Let's first look into how that will be implemented. public ActionResult GetRoles([MyDataSourceRequest(GridId.RolesUserGrid)] DataSourceRequest request) { if (request == null) { throw new Argume...

Centralized model validation both for MVC/WebApi and SPA client-side validation using FluentValidation

Validation is one of the crucial parts of any application. It has to validate on both client side and server side requests. What are target features or implementation from this article? Model validation for any given model. Centralized/One code for validation on both server-side and client-side. Automatic validation of model without writing any extra codes on/under actions for validation.  NO EXTRA/ANY codes on client-side to validate any form. Compatible with SPA. Can be compatible with any client-side validation framework/library. Like Angular Reactive form validation or any jquery validation libraries. Tools used in the implementation? FluentValidation : I feel DataAnnotation validation are excellent and simple to use, but in case of complex validation or writing any custom validations are always tricker and need to write a lot of codes to achieve whereas FluentValidations are simple even in case of complex validation. Generally, we need to validate inc...

Using Redis distributed cache in dotnet core with helper extension methods

Redis cache is out process cache provider for a distributed environment. It is popular in Azure Cloud solution, but it also has a standalone application to operate upon in case of small enterprises application. How to install Redis Cache on a local machine? Redis can be used as a local cache server too on our local machines. At first install, Chocolatey https://chocolatey.org/ , to make installation of Redis easy. Also, the version under Chocolatey supports more commands and compatible with Official Cache package from Microsoft. After Chocolatey installation hit choco install redis-64 . Once the installation is done, we can start the server by running redis-server . Distributed Cache package and registration dotnet core provides IDistributedCache interface which can be overrided with our own implementation. That is one of the beauties of dotnet core, having DI implementation at heart of framework. There is already nuget package available to override IDistributedCache i...

LDAP with ASP.Net Identity Core in MVC with project.json

Lightweight Directory Access Protocol (LDAP), the name itself explain it. An application protocol used over an IP network to access the distributed directory information service. The first and foremost thing is to add references for consuming LDAP. This has to be done by adding reference from Global Assembly Cache (GAC) into project.json "frameworks": { "net461": { "frameworkAssemblies": { "System.DirectoryServices": "4.0.0.0", "System.DirectoryServices.AccountManagement": "4.0.0.0" } } }, These  System.DirectoryServices  and  System.DirectoryServices.AccountManagement  references are used to consume LDAP functionality. It is always better to have an abstraction for irrelevant items in consuming part. For an example, the application does not need to know about PrincipalContext or any other dependent items from those two references to make it extensible. So, we can begin wi...

Voice control Sony Bravia Television through Alexa

This is my second useful thing done through Alexa after simple implementation of switching on/off light. This is not just applicable to Sony Bravia TVs but any device which can be controlled through HTTP/JSON request or via any other protocol. Hardware prerequisites for making whole thing work are as follows: 1. Sony Bravia Android TV or other devices which can accept input through HTTP or different protocol. 2. Raspberry Pi to keep running program/service. 3. Alexa device. Software prerequisites: 1. Alexa Skill: https://developer.amazon.com/edw/home.html#/skills 2. Lambda: https://console.aws.amazon.com/lambda/home?region=us-east-1#/functions 3. AWS IoT: https://console.aws.amazon.com/iot/home?region=us-east-1 How the whole process would work? Alexa would accept voice commands and converts it to intend to make a request to Lambda function. Lambda function would use converted user-friendly commands to MQTT request on AWS IoT service which would be listened through MQTT ...

Using LINQ to Entity efficiently with First/FirstOrDefault/Last/LastOrDefault/Single/SingleOrDefault

We generally use these extension methods First/FirstOrDefault/Last/LastOrDefault/Single/SingleOrDefault with predicates like ctx=> ctx.Model.FirstOrDefault(item => item.Id == 1 ) Or ctx=> ctx.Model.Where(item => item.Id == 1 ).FirstOrDefault() What is the problem with these? FirstOrDefault or similar methods immediately loads all data at once. So, let's say we have fifty columns on table then all those columns data would be retrieved from DB and saved into memory. This link gives a fair idea of different function behavior.  https://msdn.microsoft.com/en-us/library/bb882641.aspx .  So, even if we require only one value from selected field it retrieves all values. What is the solution? The solution is pretty simple. Whenever we need selected items better to do projection before calling FirstOrDefault or similar methods. Ex: Selecting single item ctx.Model.Where(itm => itm.Id == 1) .Select(itm => itm.Name).FirstO...