upstream-proxy

Route requests to Node.js apps by hostname.


Keywords
proxy, upstream, router, dispatcher, nginx, adapter, connector, http, https, tls, http/2, websocket, server-sent events, sse
License
MIT
Install
npm install upstream-proxy@2.0.7

Documentation

Upstream Proxy

Route requests to Node.js apps by hostname.

Plug it to an NGINX upstream, connect it to any other service - or expose it directly to the internet (port 80 and 443).

Upstream Proxy standalone NGINX with Upstream Proxy

It works with all kinds of web traffic...

  • HTTP
  • TLS (HTTPS)
  • HTTP/2
  • WebSocket
  • Server-Sent Events

...and is

  • easy to use
  • robust
  • fast

In the future there will be examples covering some use cases.

Installation

npm install upstream-proxy

Usage

Basic example:

const upstreamProxy = require('upstream-proxy');

let myConfig = [
    {
        name: 'app-1',
        hostnames: [ 'localhost' ],
        endpoint: { host: '127.0.0.1', port: 3001 }
    }
];

let proxy = new upstreamProxy(myConfig);
proxy.listen(3000).start();

If you want to catch all requests that don't match any given host name, you can use an asterisk:

let myConfig = [
    ...
    ...
    ...
    {
        name: 'catch-all',
        hostnames: [ '*' ],
        endpoint: { host: '127.0.0.1', port: 3999 }
    }
];

API Methods

start()

Starts the proxy.

Example:

let result = proxy.start();

console.log( result );
// OK

Once created, the proxy listens for connections immediately.
If nothing is configured, each request returns a "503 Service Unavailable" (see getStatus()).

stop()

Stops the proxy.

Example:

let result = proxy.stop();

console.log( result );
// OK

New requests will be answered with "503 Service Unavailable".
Existing connections are NOT affected, to disconnect the clients use disconnectAllClients().

getStatus()

Gets current status.

Possible return values are:

  • active => proxy routes incoming requests
  • passive => each request returns a "503 Service Unavailable"

Example:

let proxyStatus = proxy.getStatus();

console.log( proxyStatus );
// active

setConfig(obj)

Sets the configuration and generates a routing map - completely overwrites current configuration and routes.
Existing connections are not affected.

Example:

const upstreamProxy = require('upstream-proxy');

let proxy = new upstreamProxy();
proxy.listen(3000).start();

let myConfig = [
    {
        name: 'app-1',
        hostnames: [ 'localhost' ],
        endpoint: { host: '127.0.0.1', port: 3001 }
    }
];

let result = proxy.setConfig(myConfig);

console.log( result );
// OK

getConfig()

Gets the current configuration.

Example:

let liveConfig = proxy.getConfig();

console.log( JSON.stringify(liveConfig, null, 2) );
/*
[
    {
        "name": "app-1",
        "hostnames": [ "localhost" ],
        "endpoint": { "host": "127.0.0.1", "port": 3001 }
    }
]
*/

getRoutes()

Gets the current routing map.

Example:

let liveRoutes = proxy.getRoutes();

//liveRoutes is of type Map()
console.log( JSON.stringify([...liveRoutes], null, 2) ); 
/*
[
  [ "localhost", { "host": "127.0.0.1", "port": 3001 } ]
]
*/

setCallbacks(obj)

Sets callbacks (hooks) for individual error handling.

Example:

let myError503 = (socket, host_header) => {
  // individual error handling
}

let myCallbacks = {
  503: myError503
}

let result = proxy.setCallbacks(myCallbacks);

console.log( result );
// OK

Here, each request that normally would be closed by the proxy with a HTTP status 503 is handed back to "myError503" - passing back the socket object and the host name string. Now it's your responsibility to handle the request, for example:

let htmlPage = "<h1>I'm on vacation at the moment - please try again later...</h1>";

socket.end('HTTP/1.1 503 Service Unavailable\r\n\r\n' + htmlPage);

getCallbacks()

Gets currently configured callbacks.

Example:

let liveCallbacks = proxy.getCallbacks();

console.log( JSON.stringify(liveCallbacks, null, 2) );
/*
{
  503: myError503
}
*/

disconnectClients(str)

Disconnects clients for the specified host name, returns the number of terminated connections.

Example:

let nr = proxy.disconnectClients("localhost");

console.log( nr );
// 7

disconnectAllClients()

Disconnects all clients, returns the number of terminated connections.

Example:

let nr = proxy.disconnectAllClients();

console.log( nr );
// 102

Further Information

Serving the web

In a classic NGINX or Apache web server configuration you would use virtual hosting to host multiple apps/sites on a single IP address. But this is complicated and troublesome to configure/automate.

Proxying requests through Upstream Proxy to multiple Node.js processes is my favorite route to go, because it allows me to...

  • write apps that are simple to develop and test - because they're stand alone. All they need to know for proper operation is an endpoint for listening.
  • do hassle-free configuration of routes
  • apply monitoring and clustering to my apps via process manager (e.g. PM2) - avoid using a web server for it!
  • intentionally reload or restart single apps

Upstream Proxy is purpose built: It takes the place of a regular web server when you don't need it for anything else except proxying requests. If you want to plug it into an NGINX upstream for routing requests you are fine too - the configuration is much simpler and you get Javascript truly all the way down.

FYI: There is no reason you can't use Upstream Proxy as a gateway to non-Node.js apps or services.

TLS (HTTPS) and HTTP/2

SNI is supported for detecting the hostname of a secure request. Today all modern browsers support SNI.

Upstream Proxy does not need to know about your certificates. It peeks at the TLS/SNI extension headers (which are not encrypted) and forwards the encrypted data as-is. There is no need to read or modify the encrypted payload.

WebSocket(s)

They just work!

The initial negotation of for a WebSocket is done via HTTP(S), so Upstream Proxy routes it normally. Once a socket has been established, data is passively/transparently piped between the upstream and downstream sockets, including all WebSocket negotiation and data.

The same applies to Server-Sent Events.

Benchmarks

There are none published yet (internally I have done a lot of testing...).

If you create one and want to share, I would be happy to include some.