Skip to main content


Showing posts from September, 2017

Enabling Application Insights on an existing project

Yesterday I lost some time searching how to Enable Application Insights on an existing project in Visual Studio. I thought it was available on the context menu when you right click on your Visual Studio project, but no option found there: Turns out you need to go one level deeper ; Right click on your project Click on Add and select Application Insights Telemetry… Now you can go through the configuration wizard by clicking on Start Free :

Visual Studio 2017 Offline install - “Unable to download installation files”

After creating an offline installer for Visual Studio 2017 using vs_enterprise.exe --layout c:\vs2017offline we were ready to install Visual Studio on our build servers(which don’t have Internet access). However when we tried to run the installer, it failed after running for a few minutes with the following error message: Unable to download installation files This error message was not that useful as we found out that the problem was not related to missing installation files but due to the fact that we forgot to install the required certificates first. To install the certificates first, you have to Browse to the " certificates " folder inside the layout folder you created(e.g. c:\vs2017offline\certificates) Right-click each one and choose Install PFX . Specify Local machine as target certificate store Leave the password field empty More information:

Team Foundation Server–Upgrade your build agents

If you upgrade your TFS installation to a newer version, a new version of the build agent is available as well. To upgrade your agents, you have 2 options: If a new major version of the agent is released, you’ll have to manually delete the old agent and install a new agent. If a new minor version of the agent is released, the existing agent is upgraded automatically when it runs a task that requires a newer version of the agent. If you want to trigger the update manually, you can go to the Agent Pool hub, right click on a Queue and click on Update All Agents . More information at

NPGSQL–Relation does not exist

PostgreSQL has great .NET support thanks to the open source NPGSQL library . From the home page : Npgsql is an open source ADO.NET Data Provider for PostgreSQL, it allows programs written in C#, Visual Basic, F# to access the PostgreSQL database server. It is implemented in 100% C# code, is free and is open source. In addition, providers have been written for Entity Framework Core and for Entity Framework 6.x. However I immediately had some problems the moment I tried to execute a query. Strange thing was that my code almost the same as what could be found on the Getting Started page; The error I got was the following: Query failed: ERROR: relation "Northwind.Products" does not exist I tried to execute the same query directly in the PGAdmin tool and indeed I got the same error. What am I doing wrong? The problem is that PostgreSQL by default implicitly converts unquoted identifiers in my query to lowercase. So the following query; SELECT Id,

TypeScript error–Property ‘assign’ does not exists on type ‘ObjectConstructor’

A colleague asked me for help when he got into trouble with his TypeScript code. Here is a simplified version: Although this looks like valid code , the TypeScript compiler complained: After some headscratching, we discovered that there was a “rogue” tsconfig.json at a higher level that set “ES5” as the target. Object.Assign was added as part of “ES6” explaining why TypeScript complained. After changing the target to “es6”, the error disappeared.

VSWhere.exe–The Visual Studio Locator

As someone who has built a lot if CI and CD pipelines, one of the struggles I always got when new Visual Studio versions were released was how to make my build server use the correct version of MSBuild when multiple Visual Studio versions were installed. It got a lot better over the years, but even recently I was sitting together with a customer to investigate how we could make the build server understand that the Visual Studio 2017 Build tools should be used. One of the (badly documented) tricks you could use was scanning the registry for specific registry keys. Luckily Microsoft released recently a new tool that makes finding your Visual Studio instances a lot easier: vswhere.exe From the documentation: vswhere is designed to be a redistributable, single-file executable that can be used in build or deployment scripts to find where Visual Studio - or other products in the Visual Studio family - is located. For example, if you know the relative path to MSBuild, you can find

.NET Standard: Using the InternalsVisibleToAttribute

In .NET Standard projects, there is an AssemblyInfo class built-in , so you no longer need a separate AssemblyInfo.cs file in your project Properties. But what if you want to use the InternalsVisibleToAttribute ? This was one of the attributes I used a lot to expose the internals of my Assembly to my test projects. Turns out that it doesn’t matter really where you put this attribute. It is applied at the assembly level, so you can include in any source code file you like. Using the AssemblyInfo file was just a convenience. So what I did, was creating an empty .cs file and add the following code:

.NET Standard: Duplicate 'System.Reflection.AssemblyCompanyAttribute' attribute

In a ‘classic’ .NET project, you have an AssemblyInfo.cs file. This file contains all kind of information about your assembly After upgrading a classic .NET project to .NET Standard, I started to get errors about some of the properties inside the AssemblyInfo.cs file: A .NET Standard project already has the AssemblyInfo information built-in. So after upgrading you end up with 2 AssemblyInfo specifications, leading to the errors above. The solution is to remove the original AssemblyInfo.cs file in the Properties folder. Remark: If you want to change the assembly information, you now have to use the Package tab inside your Project Properties .

RabbitMQ–Configure access to Management portal

As mentioned in a previous post , it is probably a good idea to enable the RabbitMQ Management plugin to help you track what’s going inside your service broker. Now if you try to access the Management plugin using the default guest account(which you should probably remove), outside the server itself, you get a ‘Login failed’ error. Let’s fix this: Logon to the server Browse to the management portal using the localhost address: http://localhost:15762 Logon using the guest account Click on the Admin tab and scroll to the Add a user section Enter a Username and Password Specify one or more Tags as a comma separated list. If you want to give full access, enter ‘administrator’. Click on the Add user button Now click on the newly created user in the user list The set permission section is shown Leave the default settings and click Set permission . That’s it! Remark: You can do the same steps using the com

RabbitMQ–Enable Management plugin

To simplify management and monitoring of your RabbitMQ Service Broker it is a good idea to install the management plugin(don’t expect anything fancy). To install it, logon to the server where you installed RabbitMQ Open a RabbitMQ command prompt Enter the following command rabbitmq-plugins enable rabbitmq_management You’ll get the following log output D:\Program Files\RabbitMQ Server\rabbitmq_server-3.6.12\sbin>rabbitmq-plugins en able rabbitmq_management The following plugins have been enabled:   amqp_client   cowlib   cowboy   rabbitmq_web_dispatch   rabbitmq_management_agent   rabbitmq_management Applying plugin configuration to rabbit@SERVER01 ... started 6 plugins. Now you can access the management portal using http://localhost:15672 If you want to access the portal from outside the server, you have to configure a firewall rule that allows TCP traffic on port 15672. Remark: Notice t

ASP.NET Core–Configuring a WCF service

In an ASP.NET Core application(using the full .NET framework) we had to consume a WCF service. Should be easy right? Unfortunately it turned out that be more work than I expected. In a first post I explained the steps how to generate a Client Proxy, this post is about  setting the configuration. WCF configuration can be a daunting beast with a lot of options and things that can go wrong. The code generated by the proxy hardcodes (some part) of the configuration in the WCF proxy and provides you a partial method to override it but that’s not the approach we want to take. I know we’ll host the WCF service in IIS, so adding a web.config and putting the configuration logic over there sounds nice… Let’s try that: Open the generated proxy reference file  and remove the call to Service1Client.GetDefaultBinding() and Service1Client.GetDefaultEndpointAddress() in the constructor. ( Note: this is only for testing purposes) Right click on your ASP.NET Core project and ad

ASP.NET Core–Connecting to a WCF service

In an ASP.NET Core application(using the full .NET framework) we had to consume a WCF service. Should be easy right? Unfortunately it turned out that be more work than I expected. I right clicked on my project and searched for an Add service reference… option. No luck, instead I saw a Connected Services section. Maybe that will do it? I right clicked on the Connected Services section and choose Add Connected Service . This opened up the Connected Services window but no option was available to connect to an existing WCF service Maybe the Find more services… link at the button will help me? This brought me to the Visual Studio Marketplace. And yes… a search for ‘WCF’ showed up a Visual Studio Connected Services plugin that allows to add a WCF Web service reference to .NET Core projects. Exactly what I needed. I clicked on Download, closed Visual Studio after which the installer appeared and I could install the extension. After the installat


I’m currently working at a client where are (finally) migration from DB2 to SQL Server. One of the things we encountered is that DB2 is using a different precision(6 digits) for their DateTime than SQL Server, so as part of the migration process we change all target dates on SQL Server to DateTime2 to not loose any data. After migrating everything seemed to work until we tried to save an object through NHibernate to the database; we always got a StaleObjectStateException. Problem was that we were using one of these DateTime columns for concurrency checks. As NHibernate by default expects a DateTime instead of a DateTime2 we lost some precision when hydrating the objects from the database. When we later on tries to persist our changes, the concurrency check will see that the DateTimes are different resulting in a StaleObjectStateException. The solution was to change our mapping code to use DateTime2 instead. Here is our (updated) Fluent NHibernate code: And here is a similar

Can I retarget my libraries to .NET Standard 2.0?

With the release of .NET Core 2.0 and the .NET Standard 2.0 specification, it’s time to check if I can retarget some of my old libraries to .NET Standard 2.0. The tool you need is the .NET Portability Analyzer: After downloading and installing the Visual Studio extension, it is time to configure it first: Open the project you want to analyze in Visual Studio Go to Tools –> Options and click on the .NET Portability Analyzer from the left menu Select your Target Platforms and the Output formats of the generated report and click OK . Now you can right click on a specific project or your solution and choose Analyze Assembly/Project Portability . After the analysis has completed you’ll get a report that contains a nice summary, a long list of details and a list of missing assemblies:

JSON.NET–Using a Custom Contract Resolver without loosing CamelCasing

For a project I’m working we created a custom ContractResolver to apply some localization magic before JSON data is send to the client. Here is the code we are using: Let’s try this code: As you can see the ContractResolver does its job, only problem is that we loose the CamelCasing. Here is how you can fix it: Let’s run our code again:

How to start using C# 7.1?

Today I wanted to use C# 7.1 to take advantage of the new async main functionality. However I couldn’t find immediatelly where to activate it. Let’s walk through the steps: Right click on your project and choose Properties Go to the Build tab and click on the Advanced… button at the bottom Now you can either choose C#7.1 from the list or select C# latest minor version (latest) to always use the latest version. After clicking OK, you can start using the new C# 7.1 features:

Marten–Soft Deletes

On one of my projects we are using Marten , which provides a Document Store and Event Store api on top of PostgreSQL.  Behind the scenes it uses the powerfull JSON functionality that is built into the PostgreSQL database engine. The fact that we are still using an ACID compliant database makes it all a lot easier. One of the features that Marten supports are ‘soft-deletes’ . This means that documents are never actually deleted from the database. Marten will automatically filter out documents marked as deleted unless you explicitly state otherwise in the Linq Where clause. However when requesting a specific document I noticed that I still got my deleted document back; var document=await session.LoadAsync(id); If I used the query syntax instead the specific document was filtered out as expected: var documents=await session.Query().Where(s => ..); The following GitHub issue brought some insights: . This functionalit

NPM–Specifying a different registry

I already talked about the NPM registry before . Today I want to share another trick I discovered when looking in the SignalR documentation ; The JavaScript client is being published to our dev npm registry as @aspnet/signalr-client. The module contains a browserfied version of the client. You can install the module as follows: Create an .npmrc file with the following line: @aspnet:registry= Run: npm install @aspnet/signalr-client I wasn’t aware of the existance of an .npmrc file. NPM gets its config settings from the 3 locations: command line environment variables npmrc files. The nice thing is that you can create a .npmrc file at multiple levels that will be picked up when executing NPM commands. You can set a .npmrc file per-project config file (/path/to/my/project/.npmrc) per-user config file (~/.npmrc) global config file ($PREFIX/etc/npmrc) npm builtin config file (/pat

Git: Remote origin already exist

Yesterday when trying to execute the following command in Git git remote add origin I got the following error message fatal: remote origin already exists. Problem was that my Git repository was already linked to a remote named “origin”. “origin” is just a naming convention used by Git to indicate your ‘master’ repo. To solve the error you have 2 options: Create a different remote with another name: git remote add githuborigin Now I can push my code to this remote using git push -u githuborigin master Remove the existing remove first and retry the original command afterwards git remote rm origin

Upgrading to .NET Standard 2.0

After installing Visual Studio 2017 Update 15.3 and the .NET Core 2.0 SDK I thought I was finally ready to ‘upgrade’ some of my projects to .NET Standard 2.0. I right clicked on my project, selected Properties and tried to change the Target Framework on the Application tab. Unfortunately I didn’t see a .NET Standard 2.0 option in the list. OK, let’s try a different approach. I right clicked on the project and choose Edit .csproj file and updated the TargetFramework to netstandard2.0 directly in the project file This worked but resulted in a set of strange error messages when I tried to compile a .NET Core project that was referencing this library. In the end it turned out that I had a global.json file that was the root cause of my problems. The global.json file defines which version of the .NET Core SDK to use: { "sdk": { "version": "1.0.0" } } When you run dotnet new or dotnet build , the dotnet host looks in the current fol

Test Impact Analysis is back!

A loooong time ago in Visual Studio 2010 Microsoft introduced a great new feature Test Impact Analysis. Test Impact Analysis tries to predict based on your code changes which tests should be executed. In more recent versions of Visual Studio this feature disappeared…until now. Version 2 of the Visual Studio Test task reintroduces the “Run only impacted tests” checkbox. Checking this checkbox will automatically configure the Test Impact data collector to identify and run only the impacted tests. Remark: If you don’t see this option, check that you are not still using version 1 of the Visual Studio Test task: More information: