Optimizing Passenger + Nginx

There are two aspects with regard to optimizing Passenger's server performance.

  • The first aspect is settings tuning. Passenger's default settings are not aimed at optimizing, but at safety. The defaults are designed to conserve resources, to prevent server overload and to keep web apps up and running. To optimize for performance, you need to tweak some settings whose values depend on your hardware and your environment.

    Besides Passenger settings, you may also want to tune kernel-level settings.

  • The second aspect is using performance-enhancing features. This requires small application-level changes.

If you are optimizing Passenger for the purpose of benchmarking then you should also follow the benchmarking recommendations.

Table of contents

  1. Loading...

Minimizing process spawning

By default, Passenger spawns and shuts down application processes according to traffic. This allows it to use more resources during busy times, while conserving resources during idle times. This is especially useful if you host more than 1 app on a single server: if not all apps are used at the same time, then you don't have to keep all apps running at the same time.

However, spawning a process takes a lot of time (in the order of 10-20 seconds for a Rails app), and CPU usage will be near 100% during spawning. Therefore, while spawning, your server will be slower at performing other activities, such as handling requests.

For consistent performance, it is thus recommended that you configure a fixed process pool: telling Passenger to use a fixed number of processes, instead of spawning and shutting them down dynamically.

Let N be the number of processes you want. Set the following configuration in your http block:

passenger_max_pool_size N;

Set the following configuration in your server block:

passenger_min_instances N;

You should also configure passenger_pre_start in the http block so that your app is started during web server launch:

passenger_pre_start http://your-website-url.com;

Maximizing throughput

This section provides guidance on maximizing Passenger's throughput. The amount of throughput that Passenger handles is proportional to the number of processes or threads that you've configured. More processes/threads generally means more throughput, but there is an upper limit. Past a certain value, further increasing the number of processes/threads won't help. If you increase the number of processes/threads even further, then performance may even go down.

The optimal value depends on the hardware and the environment. This section will provide you with formulas to calculate that optimal value. The following factors are involved in calculation:

  • Memory. More processes implies a higher memory usage. If too much memory is used then the machine will hit swap, which slows everything down. You should only have as many processes as memory limits comfortably allow. Threads use less memory, so prefer threads when possible. You can create tens of threads in place of one process.
  • Number of CPUs. True (hardware) concurrency cannot be higher than the number of CPUs. In theory, if all processes/threads on your system use the CPUs constantly, then:

    • You can increase throughput up to NUMBER_OF_CPUS processes/threads.
    • Increasing the number of processes/threads after that point will increase virtual (software) concurrency, but will not increase true (hardware) concurrency and will not increase maximum throughput.

    Having more processes than CPUs may decrease total throughput a little thanks to context switching overhead, but the difference is not big because OSes are good at context switching these days.

    On the other hand, if your CPUs are not used constantly, e.g. because they’re often blocked on I/O, then the above does not apply and increasing the number of processes/threads does increase concurrency and throughput, at least until the CPUs are saturated.

  • Blocking I/O. This covers all blocking I/O, including hard disk access latencies, database call latencies, web API calls, etc. Handling input from the client and output to the client does not count as blocking I/O, because Passenger has buffering layers that relief the application from worrying about this.

    The more blocking I/O calls your application process/thread makes, the more time it spends on waiting for external components. While it’s waiting it does not use the CPU, so that’s when another process/thread should get the chance to use the CPU. If no other process/thread needs CPU right now (e.g. all processes/threads are waiting for I/O) then CPU time is essentially wasted. Increasing the number processes or threads decreases the chance of CPU time being wasted. It also increases concurrency, so that clients do not have to wait for a previous I/O call to be completed before being served.

The formulas in this section assume that your machine is dedicated to Passenger. If your machine also hosts other software (e.g. a database) then you'll need to tweak the formulas a little bit.

Tuning the application process and thread count

Step 1: determining the application's memory usage

The amount of memory that your application uses on a per-process basis, is key to our calculation. You should first figure out how much memory your application typically needs. Every application has different memory usage patterns, so the typical memory usage is best determined by observation.

Run your app for a while, then run passenger-status at different points in time to examine memory usage. Then calculate the average of your data points. In the rest of this section, we'll refer to the amount of memory (in MB) that an application process needs, as RAM_PER_PROCESS.

In our experience, a typical medium-sized single-threaded Rails application process can use 150 MB of RAM on a 64-bit machine, even when the spawning method is set to "smart".

Step 2: determine the system's limits

First, let's define the maximum number of (single-threaded) processes, or the total number of threads, that you can comfortably have given the amount of RAM you have. This is a reasonable upper limit that you can reach without degrading system performance. This number is not the final optimal number, but is merely used for further caculations in later steps.

There are two formulas that we can use, depending on what kind of concurrency model your application is using in production.

Purely single-threaded multi-process formula

If you didn't explicitly configure multithreading, then you are using using this concurrency model. Or, if you are not using Ruby (e.g. if you using Python, Node.js or Meteor), then you are also using this concurrency model, because Passenger only supports multithreading for Ruby apps.

The formula is then as follows:

max_app_processes = (TOTAL_RAM * 0.75) / RAM_PER_PROCESS

It is derived as follows:

  • (TOTAL_RAM * 0.75): We can assume that there must be at least 25% of free RAM that the operating system can use for other things. The result of this calculation is the RAM that is freely available for applications. If your system runs a lot of services and thus has less memory available for Passenger and its apps, then you should lower 0.75 to some constant that you think is appropriate.
  • / RAM_PER_PROCESS: Each process consumes a roughly constant amount of RAM, so the maximum number of processes is a single devision between the aforementioned calculation and this constant.

Multithreaded formula

The formula for multithreaded concurrency is as follows:

max_app_threads_per_process =
  (RAM_PER_PROCESS / 10) /

Here, CHOSEN_NUMBER_OF_PROCESSES is the number of application processes you want to use. In case of Ruby, Python, Node.js and Meteor, this should be equal to NUMBER_OF_CPUS. This is because all these languages can only utilize a single CPU core per process. If you're using a language runtime that does not have a Global Interpreter Lock, e.g. JRuby or Rubinius, then CHOSEN_NUMBER_OF_PROCESSES can be 1.

The formula is derived as follows:

  • (TOTAL_RAM * 0.75): The same as explained earlier.
  • (CHOSEN_NUMBER_OF_PROCESSES * RAM_PER_PROCESS * 0.9): This calculates the amount of memory that all the processes together would consume, assuming they're not running any threads. When this is deducted from TOTAL_RAM * 0.75, we end up with the amount of RAM available to application threads.
  • / (RAM_PER_PROCESS / 10): We estimate that a thread consumes ~10% of the amount of memory a process would, so we divide the amount of RAM available to threads with this number. What we get is the number of threads that the system can handle.

On 32-bit systems, max_app_threads_per_process should not be higher than about 200. Assuming an 8 MB stack size per thread, you will run out of virtual address space if you go much further. On 64-bit systems you don’t have to worry about this problem.

Step 3: derive the applications' needs

The earlier two formulas were not for calculating the number of processes or threads that application needs, but for calculating how much the system can handle without getting into trouble. Your application may not actually need that many processes or threads! If your application is CPU-bound, then you only need a small multiple of the number of CPUs you have. Only if your application performs a lot of blocking I/O (e.g. database calls that take tens of milliseconds to complete, or you call to Twitter) do you need a large number of processes or threads.

Armed with this knowledge, we derive the formulas for calculating how many processes or threads we actually need.

  • If your application performs a lot of blocking I/O then you should give it as many processes and threads as possible:

    # Use this formula for purely single-threaded multi-process scenarios.
    desired_app_processes = max_app_processes
    # Use this formula for multithreaded scenarios.
    desired_app_threads_per_process = max_app_threads_per_process
  • If your application doesn’t perform a lot of blocking I/O, then you should limit the number of processes or threads to a multiple of the number of CPUs to minimize context switching:

    # Use this formula for purely single-threaded multi-process scenarios.
    desired_app_processes = min(max_app_processes, NUMBER_OF_CPUS)
    # Use this formula for multithreaded scenarios.
    desired_app_threads_per_process = min(max_app_threads_per_process, 2 * NUMBER_OF_CPUS)

Step 3: configure Passenger

Purely single-threaded multi-process scenarios


  • passenger_max_pool_size <desired_app_processes>;
  • passenger_min_instances <desired_app_processes>;
  • passenger_pre_start to have your app started automatically at web server boot.

Multithreaded scenarios

In order to use multithreading you must use Passenger Enterprise. The open source version of Passenger does not support multithreading.


  • passenger_max_pool_size <CHOSEN_NUMBER_OF_PROCESSES>;
  • passenger_min_instances <CHOSEN_NUMBER_OF_PROCESSES>;
  • passenger_concurrency_model thread;
  • passenger_thread_count <desired_app_threads_per_process>;
  • passenger_pre_start to have your app started automatically at web server boot.
  • If desired_app_processes is 1, then you should set passenger_spawn_method direct. By using direct spawning instead of smart spawning, Passenger will not keep a Preloader process around, saving you some memory (learn more at Spawn methods). This is because a Preloader process is useless when there's only 1 application process.

Possible step 4: configure Rails

Only if you're using the multithreaded concurrency model do you need to configure Rails. You need to enable thread-safety by setting config.thread_safe! in config/environments/production.rb. In Rails 4.0 this is on by default for the production environment, but in earlier versions you had to enable it manually.

You should also increase the ActiveRecord pool size because it limits concurrency. You can configure it in config/database.yml. Set the pool value to the number of threads per application process. But if you believe your database cannot handle that much concurrency, keep it at a low value.

Example 1: purely single-threaded multi-process scenario with lots of blocking I/O, in a low-memory server

Suppose you have:

  • 1 GB of RAM.
  • 150 MB of memory usage per application process.
  • Lots of blocking I/O in the application.
  • A purely single-threaded multi-process scenario.

Then the calculation is as follows:

# Use this formula for purely single-threaded multi-process deployments.
max_app_processes = (1024 * 0.75) / 150 = 5.12
desired_app_processes = max_app_processes = 5.12

Conclusion: you should use 5 or 6 processes. Passenger should be configured as follows:

passenger_max_pool_size 5;
passenger_min_instances 5;

However a concurrency of 5 or 6 is way too low if your application performs a lot of blocking I/O. You should use a multithreaded deployment instead, or you need to get more RAM so you can run more processes.

Example 2: purely single-threaded multi-process scenario with lots of blocking I/O, in a high-memory server

Suppose you have:

  • 8 CPUs.
  • 32 GB of RAM.
  • 150 MB of memory usage per application process.
  • Lots of blocking I/O in the application.
  • A purely single-threaded multi-process scenario.

Then the calculation is as follows:

# Use this formula for purely single-threaded multi-process deployments.
max_app_processes = (1024 * 32 * 0.75) / 150 = 163.84
desired_app_processes = max_app_processes = 163.84

Conclusion: you should use 163 or 164 processes. This number seems high, but the value is correct. Because your app performs a lot of blocking I/O, you need a lot of I/O concurrency. The more concurrency the better. The amount of concurrency scales linearly with the number of processes, which is why you end up with such a large number.

Passenger should be configured as follows:

passenger_max_pool_size 163;
passenger_min_instances 163;

Note that in this example, 163-164 processes is merely the maximum number of processes that you can run, without overloading your RAM. It does not mean that you have enough concurrency for your application! If you need more concurrency, you should use a multithreaded deployment instead.

Example 3: multithreaded deployment with lots of blocking I/O

Consider the same machine as in example 2:

  • 8 CPUs.
  • 32 GB of RAM.
  • 150 MB of memory usage per application process.
  • Lots of blocking I/O in the application.
  • A purely single-threaded multi-process scenario.

But this time you're using multithreading with 8 application processes (because you have 8 CPUs). How many threads do you need per process?

# Use this formula for multithreaded deployments.
= ((1024 * 32 * 0.75) - (8 * 150)) / (150 / 10) / 8
= 194.8

Conclusion: you should use 195 threads per process.

passenger_max_pool_size 8;
passenger_min_instances 8;
passenger_concurrency_model thread;
passenger_thread_count 195;

Because of the huge number of threads, this only works on a 64-bit platform. If you're on a 32-bit platform, consider lowering the number of threads while raising the number of processes. For example, you can double the number of processes (to 16) and halve the number of threads (to 779).


  • If your application performs a lot of blocking I/O, use lots of processes/threads. You should move away from single-threaded multiprocessing in this case, and start using multithreading.
  • If your application is CPU-bound, use a small multiple of the number of CPUs.
  • Do not exceed the number of processes/threads your system can handle without swapping.

Performance-enhancing features


Passenger supports turbocaching since version 4. Turbocaching is an HTTP cache built inside Passenger. When used correctly, the cache can accelerate your app tremendously. To utilize turbocaching, you only need to set HTTP caching headers.

Learning about HTTP caching headers

The first thing you should do is to learn how to use HTTP caching headers. It's pretty simple and straightforward. Since the turbocache is just a normal HTTP shared cache, it respects all the HTTP caching rules.

Set an Expires or Cache-Control header

To activate the turbocache, the response must contain either an "Expires" header or a "Cache-Control" header.

The "Expires" header tells the turbocache how long to cache a response. Its value is an HTTP timestamp, e.g. "Thu, 01 Dec 1994 16:00:00 GMT".

The Cache-Control header is a more advanced header that not only allows you to set the caching time, but also how the cache should behave. The easiest way to use it is to set the max-age flag, which has the same effect as setting "Expires". For example, this tells the turbocache that the response is cacheable for at most 60 seconds:

Cache-Control: max-age=60

As you can see, a "Cache-Control" header is much easier to generate than an "Expires" header. Furthermore, "Expires" doesn't work if the visitor's computer's clock is wrongly configured, while "Cache-Control" does. This is why we recommend using "Cache-Control".

Another flag to be aware of is the private flag. This flag tells any shared caches – caches which are meant to store responses for many users – not to cache the response. The turbocache is a shared cache. However, the browser's cache is not, so the browser can still cache the response. You should set the "private" flag on responses which are meant for a single user, as you will learn later in this article.

And finally, there is the no-store flag, which tells all caches – even the browser's – not to cache the response.

Here is an example of a response which is cacheable for 60 seconds by the browser's cache, but not by the turbocache:

Cache-Control: max-age=60,private

The HTTP specification specifies a bunch of other flags, but they're not relevant for the turbocache.

Only GET requests are cacheable

The turbocache currently only caches GET requests. POST, PUT, DELETE and other requests are never cached. If you want your response to be cacheable by the turbocache, be sure to use GET requests, but also be sure that your request is idempotent.

Avoid using the "Vary" header

The "Vary" header is used to tell caches that the response depends on one or more request headers. But the turbocache does not implement support for the "Vary" header, so if you output a "Vary" header then the turbocache will not cache your response at all. Avoid using the "Vary" header where possible.

Maximum response size

The turbocache caches only responses that are at most 32 KB, including HTTP headers. This maximum size is currently not configurable, but we are working on it.

Maximum caching time

The turbocache caches a response for a maximum duration of 2 seconds, or whatever is specified as the expiry time according to the HTTP headers, whichever is earliest. The cap of 2 seconds is currently not configurable, but we are working on it.

Out-of-band garbage collection

Passenger supports out-of-band garbage collection for Ruby apps. With this feature enabled, Passenger can run the garbage collector in between requests, so that the garbage collector doesn't delay the app as much. Please refer to the Users Guide for more information about this feature.

Switching to Passenger Standalone with the builtin HTTP engine

In certain situations, using Passenger Standalone in combination with the builtin HTTP engine may yield some performance benefits because it skips a layer of processing.

Passenger normally works by integrating into Nginx or Apache. As described in the Design & Architecture document, requests are first handled by Nginx or Apache, and then forwarded to the Passenger core process and the application process. This architecture provides various benefits, such as security benefits (Nginx and Apache's HTTP connection handling routines are thoroughly battle-tested and secure) and feature benefits (e.g. Gzip compression, superb static file handling).

This is even true if you use the Standalone mode. Although it acts standalone, it is implemented under the hood by running Passenger in a builtin Nginx engine.

However, the fact that all requests go through Nginx or Apache means that there is a slight overhead, which can be avoided. This overhead is small (much smaller than typical application and network overhead), and using Nginx or Apache is very useful, but in certain special situations it may be beneficial to skip this layer.

  • In microbenchmarks, the overhead of Nginx and Apache are very noticeable. Removing Nginx and Apache from the setup, and benchmarking against the Passenger HelperAgent directly, will yield much better results.
  • In some multi-server setups, Nginx and Apache may be redundant. Recall that in typical multi-server setups there is a load balancer which forwards requests to one of the many web servers. Each web server in this setup runs Passenger. But the load balancer is sometimes already responsible for many of the tasks that Nginx and Apache perform, e.g. the secure handling of HTTP connections, buffering, slow client protection or even static file serving. In these cases, removing Nginx and Apache from the web servers and load balancing to the Passenger HelperAgent directly may have a minor improvement on performance.

Nginx and Apache can be removed by using Passenger's builtin HTTP engine. By using this engine, Passenger will listen directly on a socket for HTTP requests, without using Nginx or Apache.

This builtin HTTP engine can be accessed by starting Passenger Standalone using the --engine=builtin parameter, like this:

passenger start --engine=builtin

It should be noted that the builtin HTTP engine has fewer features than the Nginx engine, by design. For example the builtin HTTP engine does not support serving static files, nor does it support gzip compression. Thus, we recommend using the Nginx engine in most situations, unless you have special needs such as documented above.

There is a version of this Optimization Guide for the Standalone mode

You are currently reading the Nginx version of this Optimization Guide. Have you decided to switch to Passenger Standalone? Read the Passenger Standalone version of this guide.

Benchmarking recommendations

Tooling recommendations

  • Use wrk as benchmarking tool.
    • We do not recommend ab because it's slow and buggy.
    • We do not recommend siege and httperf because they cannot utilize multiple CPU cores.
  • Enable HTTP keep-alive in both the server and in your benchmarking tool. Otherwise you will end up benchmarking how quickly the kernel can set up TCP connections, which is a non-trivial part of the request time.
  • Warm up the server before benchmarking. That is, run a mini-benchmark before the actual benchmark, and discard the result of the mini-benchmark.

Operating system recommendations

  • Don't benchmark on macOS. macOS's TCP stack and process scheduler are horrible from a performance point of view. We recommend Linux.
  • When on Linux, be sure to tune your kernel socket settings so that they don't stall the benchmark.

Server and application recommendations

  • If the purpose of your benchmark is to compare against Puma, Unicorn or other app servers, be sure to benchmark against Passenger Standalone, not Passenger for Nginx or Passenger for Apache.
  • Configure Passenger to use a static number of processes. The number of processes should be a multiple of the number of CPU cores if you are using MRI.
  • Ensure that your app outputs HTTP caching headers, so that Passenger's turbocaching can kick in.
  • Pay attention to the actual responses from the app servers, especially error codes (any http code >= 400), and dropped requests. Puma for example drops requests when it cannot keep up with the load in our testing.
  • Pay attention to network congestion and server load as both can introduce noise into your results. We recommend that you do not run your benchmarking tool on the same box as your app server, and do not run your benchmark on a network segment with other traffic present.
  • Mock out sources of performance variation in your webapp, such as external network requests or file/database access.
  • If you are using JRuby pay special attention to tuning your VM for your webapp.