Considering appropriate app service plan for your Azure Functions

As we all(or most of us) know, Azure Functions can be hosted using either a regular app service plan or a consumption plan. We can quickly summarize pros and cons of both:

App service plan:

  • fixed cost(+)
  • easy to scale(+)
  • ability to run 64-bit applications(+)
  • can reuse other app service plans(+)
  • fixed cost(--)
  • some triggers need Always On enabled(-)

Consumption plan

  • pay-as-you-go(++)
  • no need to have Always On for e.g. TimerTrigger(+)
  • somehow more difficult to scale(-)
  • if not designed carefully, the cost may exceed our expectations seriously(-)

All right - but what really should I consider when it comes to choosing the correct plan for my application?


In the current world scalability is something, what should be really considered when designing an application and choosing technologies for it. Let's consider a following example - you're designing an e-commerce application, which has to handle really big traffic spikes from time to time(imagine Black Friday or any other black something). The specifics of traffic on your website could be described as:

  • stable and low traffic for the most of a week
  • an increase during weekend
  • occasional huge spikes during special events 

Now let's relate this to our service plans. What is better for us in such scenario?

Well, the problem with consumption plan is the fact, that it needs a running start. You can't expect, that if your application is hit by a huge traffic spike, it'll scale out immediately. What is more, you don't have a possibility to scale up - you're limited to the resources allocated for you for your instance of the consumption plan. You have to consider one more thing - execution of functions is not throttled in any way by default so you may face a situation, where under a heavy load your function utilize too much CPU/memory at once. You can control throttling by following three properties:

  • maxOutstandingRequests
  • maxConcurrentRequests
  • dynamicThrottlesEnabled

which are described here. The one thing you have to remember when using throttling is the possibility, that your client may get HTTP 429 Too Busy responses. Whether it's a problem, only you can decide.

When using s regular app service plan those traffic spikes are much easier to handle. Since you can have scaling rules, you don't care when scaling out happens - it just does when it hits CPU/memory threshold(or other metric if autoscale is enabled). Additionally you can preprovision extra resources if you know when a traffic spike will happen - this way you're prepared.


When designing cloud solutions, the cost is one of the most important factors. If you choose components poorly or overdesign them, the bill at the end of a month won't make you happy for sure. This is the second thing directly related to service plans, which affects our choice when it comes to select what is the best.

Pay-as-you-go model in consumption plan is something, what really make functions interesting. When designed carefully, you can run them for almost free each month(or pay only a few USD/EUR after the free quota is exceeded). The problem is when you keep your functions "red" - in such scenario, it may be easier and cheaper to use a regular app service plan, which ensures the constant cost of this component and won't surprise you after a busy weekend(how is that I have to pay extra 500$ this month?).

Of course with a regular app service plan you lose flexibility and have to remember to scale your application down(or at least have something to automate this). The compromise here depends on your current needs and how the model of your business looks like. However it's still better to discuss it now and be aware of your possibilities rather than discovering them when functions start to respond with HTTP 503 status.

Handling your local directory in Azure Functions

Since Azure Functions are not a typical(and least in terms how they're executed) application, you may from time to time face a problems, which are not a case in more common solutions. One of those is a problem with storing local files. Consider following snippet:

using System.Net;
using System.IO;

public static async Task<HttpResponseMessage> Run(HttpRequestMessage req, TraceWriter log)
    File.WriteAllText("test.txt", "Foo");

    return req.CreateResponse(HttpStatusCode.OK, "Hello ");

when you run this code in as your function you will get following error:

2017-05-17T06:28:37.775 Exception while executing function: Functions.HttpTriggerCSharp1. mscorlib: Access to the path 'D:\Windows\system32\test.txt' is denied.

This is because if you don't provide a full path to your file, the default working directory of the process, inside which your functions runs, is D:\Windows\system32. Fortunately there's an easy fix to this.

Finding your local directory

The receipt to get your local directory is really simple. There's one thing to remember - your function app is hosted inside a following directory: D:\home\site\wwwroot. When you look at it, you can expect following structure:

	- site
		- wwwroot
			- Function1
			- Function2
			- ...

All we need to do here is to combine main directory path with the name of the function. That way we can modify the snippet from above slightly:

using System.Net;
using System.IO;

public static async Task<HttpResponseMessage> Run(HttpRequestMessage req, TraceWriter log)
    File.WriteAllText("D:\\home\\site\\wwwroot\\HttpTriggerCSharp1\\test.txt", "Foo");

    return req.CreateResponse(HttpStatusCode.OK, "Hello ");

and when we run it, you should see following output:

2017-05-17T06:43:29.408 Function started (Id=1baa9a60-d549-492d-8fef-089cd8e8de06)
2017-05-17T06:43:29.408 Function completed (Success, Id=1baa9a60-d549-492d-8fef-089cd8e8de06, Duration=4ms)