NAME

UniEvent::HTTP::Manager - extremely fast asynchronous preforking / threading event based web server

SYNOPSIS

my $mgr = UniEvent::HTTP::Manager->new({
    min_servers  => 3,
    max_servers  => 20,
    min_load     => 0.2,
    max_load     => 0.7,
    max_requests => 10000,
    worker_model => UniEvent::HTTP::Manager::WORKER_PREFORK,
    server       => { locations => [{host => 'localhost', port => 8080, reuse_port => 1}], },
});

# called in different processes/threads, simple API
$mgr->request_callback(sub {
    my $request = shift;
    $request->respond({
        code => 200,
        body => "hi",
    });
});

# more detailed tuning
# called once per lifetime of worker process/thread
$mgr->spawn_callback(sub {
    my $server = shift; # UniEvent::HTTP::Server object
    # set callbacks in worker's server instance
    $server->route_callback(sub {
        ...
    });
});

$mgr->run;

# run via Plack
plackup -s UniEvent::HTTP -E deployment --port 9090 --min_servers 1 --max_servers 10 --min_load 0.2 --max_load 0.7

DESCRIPTION

This module provides multi-process (or multi-thread) web server build on top of UniEvent::HTTP::Server. It creates an instance of UniEvent::HTTP::Server in each process/thread and maintain certain number of workers defined by config.

UniEvent::HTTP::Server is a full-featured asynchronous RFC-compliant event based http server, so does this module.

Main features are:

REAL high performance, faster than nginx

For example, on AMD Ryzen 3950x + Debian 10, up to 1.500.000 keep-alive http requests per second with 16 workers (up to 100.000 per second per worker).

Automatic workers management

Adds workers when load increases and shutdowns when load decreases.

Multiple listen interfaces support
On-the-fly reconfiguration

Including addresses/ports to listen, number of workers, etc...

SSL support
Asynchronous event loop based

You can create timers, signals, custom tcp connections, etc... and run it in workers or master process. UniEvent should be used as event loop framework.

Support for graceful restart/stop

Without interrupting any http request

Blocking code support

Request processing logic doesn't have to be asynchronous. Load management will nicely work even in this case (if using "min_load/max_load"). This allows for smooth migration from synchronous to asynchronous code base in web servers.

PSGI/Plack compatible

Can run any PSGI applications and frameworks (performance penalties may apply).

See Plack::Handler::UniEvent::HTTP

All the features inherited from UniEvent::HTTP::Server like:
HTTP/1.1 support

Chunked requests and responses, keep-alive and pipeline requests

Compression support

incoming and outgoing

Request/response streaming

Directly into / from target without temporary files or memory consumption

See UniEvent::HTTP for the description of all features

This module is a perl port of C++ panda::unievent::http::manager library.

USING HTTP MANAGER

First thing to do is to create an http manager object

my $manager = UE::HTTP::Manager->new({ ...config... });

Config describes http manager parameters like how many workers there should be depending on various circumstances, how often to check for workers state, minimum worker time to live, maximum requests per worker, etc..., and the http server config itself in server key. Http server config is not described here as it is directly passed to UniEvent::HTTP::Server, so read its docs.

At this point no workers are spawned and thus no http server objects are created yet. To add listeners to http server objects, you need to add callback for spawing a worker.

$manager->spawn_callback(sub {
    my $server = shift;
    ...
});

This callback is called each time a new worker process/thread is created and is called in that worker's process/thread context. It is passed an http server object (UniEvent::HTTP::Server) created with config from server key of main config.

We should add request processing and other callbacks to the server object at this point.

$manager->spawn_callback(sub {
    my $server = shift;
    
    $server->request_callback(sub {
        my $request = shift;
        if ($request->uri->path eq "/hello") {
            $request->respond({
                code    => 200,
                headers => {...},
                body    => "Hi",
            });
        } else {
            $request->respond({code => 404});
        }
    });
    
    $server->error_callback(sub {
        ...
    });
    
    $server->stop_callback(sub {
        # do the cleanup
    });
});

For more control over how request is processed server's route_callback may be used. See UniEvent::HTTP::Server.

If you only need request_callback on server, simple API can be used

$manager->request_callback(sub { ... });

It will set that callback in each http server it spawns.

After setting up things web server should be run

$manager->run;

And this function will not return until server is stopped (it will run the event loop).

To stop the web server stop() can be called at any time.

See docs below for detailed explanation on how it works.

EVENT LOOPS IN HTTP MANAGER

If you're not familiar with event loop programming you should probably first read docs about it. For example UniEvent or AnyEvent.

Manager has two event loops: the one it runs in the master process/thread and the one it runs in worker processes/threads.

Http manager is built on top of UniEvent event loop framework and, unlike AnyEvent, it allows for having multiple event loop objects.

By default, manager creates separate non-default loop object which it runs in master process. It as accessible via $manager->loop(). You should use that loop to create your own event handles that you wish to run in master process.

In worker processes/threads by default, the default loop is used (UniEvent::Loop->default). This is for convenience, as in this case you may not pass an event loop object to event handles constructors as it is defaulted to default loop.

# in worker
my $timer = UE::timer 10, sub { ... }; # runs in default loop

To change this defaults, you may pass a pair of loop objects to manager's constructor (see new()) and it will use them in master/worker processes.

NOTE: if prefork model is used and you create event handles in master process, it is recommended that master and worker loops are different loops. If they are the same, you will have to remove all master process' event handles when worker is spawned (from worker process, in spawn_callback code). Not doing so will lead to master process' code execution in workers. The http manager framework itself always cleans up its master process resources in workers.

NOTE: if thread model is used, workers loop will always be default loop, because loop objects are not thread-safe and the default loop is the only loop that differs in each thread (it is thread-local). If you pass workers loop object different from default loop it will croak. Master thread's loop can be anything (default or not) and no resources should be released in spawn_callback. By default, like in prefork model, master loop is non-default loop.

Example showing how to add timers in master process and worker process:

my $manager = UE::HTTP::Manager->new({...});
my $master_timer = UE::timer 1, sub {
    # master's maintenance work
}, $manager->loop;

$manager->spawn_callback(sub {
    my $server = shift;
    ...
    my $worker_timer = UE::timer 1, sub {
        # worker's maintenance work
    };
    
    # as this callback returns before server is run, we need to hold $worker_timer somewhere, otherwise it will immediately be "garbage-collected"
    # for example, we hold by closure
    # also stopping timer in server's stop callback becomes mandatory if "force_worker_stop" config parameter is set to false, see docs for new()
    $server->stop_callback(sub {
        $worker_timer->stop; 
    });
    
});

$manager->run;

METHODS

new(\%config, [$master_loop = private loop], [$worker_loop = default loop])

Creates web server instance. Config is a hashref with the following supported parameters (square brackets marks required parameters or its default values if not passed):

server [required]

Config for UniEvent::HTTP::Server object which is created in each worker when it is spawned. See UniEvent::HTTP::Server docs for description.

NOTE: UniEvent::HTTP::Manager keeps an eye on UniEvent::HTTP::Server's server.locations[].reuse_port parameter.

This controls how manager will create listening sockets.

If reuse_port is false, manager will create listening socket for this location in master process once and use the same sockets in each worker. So that they will be accepting the same socket. The advantage of this method is that it allows to start web server with some advanced privileges, bind socket to ports like 80, 443, etc... in master process and then drop privileges, for example, in manager's start_callback (see docs below). The disadvantage is that it may impact performance significantly in case of many running workers and very high number of new connections per second (i.e. when keep-alive is not utilized much and your requests handlers are small and fast). This is not a technical disadvantage of this library, but the OS itself.

If reuse_port is true, manager will not create any sockets for this location, instead each worker will create listening socket by itself and bind it to the same addresses and ports making use of REUSE_PORT feature. This way each worker will accept its own socket and the OS will distribute the load between them. This method has a WAY LESS performance penalties when a lot of processes/threads are accepting a lot of new connections per second on the same addresses/ports.

Alternatively, you can create sockets by yourself and pass it in sock property of a location. In this case reuse_port is ignored (set to false). You can even create sockets in worker processes (in spawn_callback) and reconfigure worker's http server there. Keep in mind that in these cases sockets you create must be bound to an address/port but not listening yet (listen() must not be called). See UniEvent::HTTP::Server docs for more info.

reuse_port is only meaningful with host and port combination and has no effect with path (unix sockets / named pipes) and custom sock parameter (the effect is as it was false). Also reuse_port is always false on windows as it is not supported there. For non-windows systems it is enabled by default.

min_servers [1]

Minimum number of workers at any time. Number of workers will not drop below this value despite of any circumstances.

max_servers [max(number of CPUs, min_servers)]

Maximum number of active workers. Default value is number of CPUs on the machine or min_servers if it is higher.

NOTE: Number of workers MAY temporarily exceed this value when some workers get restarted (due to max requests or bulk workers restart). In such cases manager starts a new worker to replace some old worker, waits until new worker gets fully initialized and ready to process requests, then it signals the old worker to stop (which can take time if old worker is currently active, i.e. processing some request). Usually this process takes fractions of a second (however it is a subject to change if your code for example in spawn_callback is slow) and occurs rarely so it's not a big deal.

min_spare_servers

The minimum number of servers to have waiting for requests. Minimum and maximum numbers should not be set to close to each other or the server may spawn and kill workers too often.

This option is added for compability with other web servers only, as it is unclear what "waiting for requests" means for asynchronous web server. In UniEvent::HTTP::Manager a worker is spare if number of requests it is currently processing is 0. Because in asynchronous server any worker is waiting for requests at any time. This option is NOT very USEFUL, especially if you code is asynchronous or partially asynchronous. Use it only if all of your request processing handlers are blocking.

NOTE: it is recommended to use min_load/max_load instead.

max_spare_servers [min(min_servers + min_spare_servers, max_servers), only if min_spare_servers is set]

The maximum number of servers to have waiting for requests. See min_spare_servers for explanation.

NOTE: it is recommended to use min_load/max_load instead.

max_load [0.7, if min_spare_servers is not set]

Maximum average workers busy load, before spawning more workers (fractional, 1 = 100% load). This value is not a CPU load, it is somewhat more useful, see UniEvent::Loop, track_load_average() for details what this value means. If average busy load on workers becomes higher that configured value, manager will start more workers until busy load gets lower than configured value.

min_load [max_load*0.5, if max_load is set]

Minimum average workers load, before stopping excess workers. If average busy load on workers becomes lower that configured value, manager will start stopping workers until busy load gets higher than configured value.

load_average_period [3]

How many last seconds to collect busy load for. If you use min/max_spare_servers instead of min/max_load this value is not used.

max_requests [0]

Maximum number of requests per worker. If worker reaches this number, new replacing worker is started and after that this worker gets stopped. 0 means do not limit number of requests.

min_worker_ttl [60]

Minimum number of seconds a worker will live. This property has higher priority than min/max_load, max_requests, etc... That means that if, for example, a worker has reached maximum number of requests but started less than min_worker_ttl seconds ago, it will live until it reaches min_worker_ttl. Second example is when manager wants to stop some workers due to min_load setting and no workers are older than min_worker_ttl, then no workers will be stopped until someone reaches min_worker_ttl.

check_interval [1]

Interval in seconds between main manager checks (to see if we can kill off some excess workers or if we need to spawn more workers, etc...). Can be fractional.

activity_timeout [0]

This setting helps dealing with hanged workers. If worker is not responding to manager during this number of seconds it gets killed. If you set this setting, ensure that you don't have heavy handlers that lasts more than activity_timeout without returning to loop (this is especially important if you have code that does blocking I/O). 0 means disable this feature.

termination_timeout [10]

Maximum number of seconds for a worker to finish its stuff after manager requested it to stop. If this amount of time is exceeded then worker gets killed. 0 means unlimited time for worker to finish.

force_worker_stop [true]

This option controls what to do when a worker (previously requested to stop) has stopped (finished all active requests and sent all responses).

If set to true, worker will call $loop->stop() and thus exit immediately.

If set to false worker will do nothing and the loop will bail out of execution (i.e. worker will exit) if (or when) there are no more active event loop handles. This allows for possible user's custom event handles (not related to http server) to finish. However keep in mind that if there will be any endless and strong (non-weak) event handle (for example, a non-weak timer), loop (and worker) will never finish and will be killed by termination_timeout (or never if termination_timeout is 0).

worker_model [WORKER_PREFORK on Unix, WORKER_THREAD on Windows]

MPM (multiprocessing module). Should be either UniEvent::HTTP::Manager::WORKER_PREFORK or UniEvent::HTTP::Manager::WORKER_THREAD. UniEvent::HTTP::Manager::WORKER_PREFORK is not supported on Windows. UniEvent::HTTP::Manager::WORKER_THREAD is supported only on threaded perls. UniEvent::HTTP::Manager::WORKER_THREAD support for perl is highly experimental for now and should not be used yet.

loop()

Returns UniEvent::Loop object in which manager runs (event loop of master process).

run()

Runs the web server. This function will run the event loop and thus will not return until the server is stopped.

restart_workers()

Restarts all workers. At first, manager will start new workers for replacing the old ones and thus double the workers count. When corresponding new working is initialized and ready, corresponding old worker will be sent a signal to stop. It will finish all active requests and exit. After this process the number of workers will become as before. So that this process may exceed max_workers configuration parameter temporarily.

If this method is called shortly after previous call, when previous process hasn't yet finished, then another new workers set will be spawned tripling the number of workers that was before the first call to restart_workers().

reconfigure(\%config)

Reconfigures the web server. Supported parameters are the same as in new(), except for worker_model and server.locations[].reuse_port which can't be changed. All other parameters can be changed on-the-fly. Manager may gracefully restart all workers during this process (or may not, if significant parameters hasn't been changed).

This method may return error

start_callback($sub)

start_event()

Callbacks set via these methods will be invoked right after server is started, i.e. created and bound all sockets, right before entering the event loop.

It is a good place, for example, to drop privileges if process has been started with advanced privileges.

Callback doesn't receive anything.

See "EVENT CALLBACKS" in UniEvent for differences between _callback and _event versions of methods.

spawn_callback($sub)

spawn_event()

Callbacks set via these methods will be invoked when a new worker is spawned, in proccess/thread of worker.

Callback signature:

my $server = shift;

Where $server is personal UniEvent::HTTP::Server instance which will run in this worker.

This callback is usually used for setting callbacks in server instance and for creating custom event handles in worker (timers, signal handlers, etc...)

See "EVENT CALLBACKS" in UniEvent for differences between _callback and _event versions of methods.

request_callback($sub)

request_event()

Callbacks set via these methods will be directly passed to worker's UniEvent::HTTP::Server instance when spawning a new worker.

It is the same as

$manager->spawn_callback(sub {
    my $server = shift;
    $server->request_callback($sub);
    $server->request_event->...;
});

See UniEvent::HTTP::Server docs for API.

MULTI-THREAD WORKER MODEL FOR THREADED PERL

This support is HIGHLY EXPERIMENTAL and should not be used in production for now. Because of numerious issues in dependency XS modules with perl-threads, it will likely crash in perl destruction phase (when any worker dies), however it may work if number of workers is fixed.

This will be fixed in the future as multi-thread model is the only possible model for MS Windows (it doesn't support prefork model). However it is not a short task because supporting and handling ugly perl threads in XS modules and making it work with normal C-threads in C++ framework is anything but not a simple task.

LOGS

Logs are accessible via XLog framework as "UniEvent::HTTP::Manager" module.

XLog::set_logger(XLog::Console->new);
XLog::set_level(XLog::DEBUG, "UniEvent::HTTP::Manager");

You might also need logs from "UniEvent::HTTP" layer. See UniEvent::HTTP

SEE ALSO

Plack::Handler::UniEvent::HTTP

UniEvent::HTTP

UniEvent

AUTHOR

Pronin Oleg <syber@cpan.org>

Crazy Panda LTD

LICENSE

You may distribute this code under the same terms as Perl itself.