Have you been experimenting with Server-Side Swift? Or perhaps, have you considered it and want to know what is involved? One of the biggest questions with a new backend technology is: how easy is it to host? I have deployed a few websites and services, including this blog, so I would like to share with with you what has worked best for me.
For the purposes of this post, I will assume you know how to develop and build the Server-Side Swift. It doesn't matter what framework you use as long as it produces a binary that can listen to a custom port on the network (more on that later). I will also assume that you have at least some experience managing a linux server from the command line.
If those things are still a challenge for you, I recommend you check out my other posts and subscribe to my future posts as I will be writing a lot about how to develop Server-Side Swift in the future.
Quick Summary of HTTP
Before I get into my configuration, I want to describe quickly how a website or service communicates with users. If you are already familar with this, feel free to skip to the next section. The web is built on a protocol called Hypertext Transfer Protocol or HTTP for short. It provides a framework for how a web client should make a request to a server and how that server should respond.
If you are developing a server based application, you will almost certainly be using a framework to handle the nitty-gritty of this protocol. All you really need to know is that the job of your server is to listen for requests from the outside world and respond.
The requests are also sent to a specific port which you can think of like a door that traffic can pass through represented by a number. By default, insecure web requests sent over HTTP are sent to port 80 while secure web requests over HTTPS are sent to port 443. You will see how these facts are important as I talk about setting up your server.
So to summarize, the relevant parts of a web request for us are:
- Domain name (or IP Address)
- Secure or Unsecure known as HTTPS or HTTP
The rest is handled once the request is propertly routed to your service. Ok, now onto the actual deployment.
Before we can even talk about configuration, we first have to discuss the server. If you already have a Linux server, feel free to jump to the next section.
There are many options for a web server but while Linux is in third place when it comes to desktop operating systems, it dominates the backbone of the internet. You could try to run your website from your Mac at home or possibly rent out a Mac online, but there are many more services available to use Linux and driving traffic through your home internet is not a great idea.
To host Swift I highly recommend renting out a Virtual Private Server or a VPS. Essentially you rent part of a physical computer hosted and managed by a company that specializes in that. You are given remote access to it and it appears as if you have a full computer all to yourself. There are many Linux web hosts out there, but the one I recommned is Linode. I have multiple servers (or "Linodes") hosted with them and they have been very reliable and the company is very respectful of its customers, including with the prices they charge. They also have fantastic documenation, especially on properly securing your server.
Once you have a server setup and secured, you need to install Swift. A great tool for managing different versions of Swift on Linux is Swift Version Manager otherwise known as Swiftenv. Follow the instructions on their page about installing via a git clone. Once installed, logout of your SSH session and log back in.
Now you can install Swift by running the following command:
swiftenv install 4.0.2
For the purposes of this post, I will be setting up a production version of my blog at drewag.me and a development version of it at dev.drewag.me
(no, this subdomain does not actually exist). To do this, I will create two different NGINX configuration files. They should both be created in
/etc/nginx/sites-available/. Having two separate files allows us to start and stop each site independently. I like to name my configuration files after
the domain name they are intended for so let's start with the development one.
sudo nano /etc/nginx/sites-available/dev.drewag.me.conf
In this file we want to setup a virtual host at dev.drewag.me that will proxy all of the requests to our local port. I am going to use port 8081 for the development service.
# /etc/nginx/sites-available/dev.drewag.me <VirtualHost *:80> ServerName dev.drewag.me ProxyPreserveHost On ProxyPass / http://localhost:8081 ProxyPassReverse / http://localhost:8081 </VirtualHost>
Let's go through this line by line. Apache configuration files use an XML style system to specify sections. Here we are defining a virtual host that will listen
for requests from any IP address (that is what the
* means) on port 80. On the next line, we specify that the domain name we want for this virtual host is
dev.drewag.me. The rest of the configurations are for the proxying.
First we turn on
ProxyPreserveHost. This ensures that the request seen by our service appear as if they are coming directly from its original source instead of
Apache. Next, we are specifying that we want to pass all requests made to the root level and below, back to this server at port 8080. That is the most critical line
for our purposes. However, on the next line we also set
ProxyPassReverse. This ensures that once our service responds, Apache will update the response to look
like it is coming from our domain instead of the localhost address. Finally our last line closes that virtual host configuration section.
Before you enable this configuration we must first ensure that mod_proxy, the component of Apache responsible for the proxy commands, is enabled. You can do so with the following command:
sudo a2enmod proxy
Now you can enable your virtual host site:
sudo a2ensite dev.drewag.me
To make sure both of those commands take effect, you need to restart apache:
sudo service apache2 restart
Now, as long as you have your domain name pointing at your server, you can start up your service to listen on port 8081 and send requests to it at your domain name! Then
you can create as many additional site configuration files with the appropriate
ServerName and a different internal port. With this setup, they will all work simultaneously.
Note that if you secured your server properly, port 8081 should not be open to the public (it should be blocked by your firewall). The only reason traffic is able to get there through this configuration is that it is being passed from port 80 by Apache.
Securing Your Services with SSL
I choose to run my site over HTTPS to ensure security for my visitors. I am not dealing with any real secure information, other than my donate page, but I still feel "better safe than sorry" is appropriate here. To do this, we need to make a small adjustment to our configuration file. This time, I will show you my configuration file for the production version of drewag.me. First, we are going to setup a virtual host on port 80, the HTTP port, that will redirect all traffic to the HTTPS version of our site:
# /etc/apache2/sites-available/dev.drewag.me <VirtualHost *:80> ServerName drewag.me Redirect permanent / https://drewag.me/ </VirtualHost>
That is pretty simple. All I am specifying is that all traffic to the root path should be permanently redirected to the HTTPS domain. Warning: pay careful attention
that you include the forward slash (
/) at the end of the domain name.
Next, in the same file, we are going to add a virtual host definition for handling the HTTPS requests. This time it will be on port 443:
# /etc/apache2/sites-available/dev.drewag.me <VirtualHost *:443> ServerName drewag.me ProxyPreserveHost On ProxyPass / http://localhost:8080 ProxyPassReverse / http://localhost:8080 SSLEngine on SSLCertificateFile /etc/ssl/certs/drewag_me.crt SSLCertificateKeyFile /etc/ssl/private/drewag.me.key SSLCertificateChainFile /etc/ssl/certs/drewag_me.ca-bundle </VirtualHost>
This looks similar to our HTTP version but with some extra configurations. The proxy configurations are the same except this time, they are redirecting to port 8080. The other group of configurations are specific to the secure part of HTTPS, the SSL configuration. SSL is a method of encrypting all communication between two locations so that only the intended participants can read the contents. This mechanism requres that you buy an SSL certificate from a trusted Certificate Authority. You can create and sign your own certificates but then browsers will warn visitors to your site that they cannot trust it. I use the site namecheap because they have cheap certificates and, for my purposes, I have not needed incredibly secure certificates. I have only had good experiences with them.
The first SSL configuration
SSLEngine On is simply turning on the encryption. The remaining configurations specify the necessary files for doing the encryption. Your Certificate Authority
will have instructions on how to generate or download each of the files.
Now you just need to make sure the SSL module is enabled along with this new site config and restart Apache:
sudo a2enmod proxy sudo a2ensite drewag.me sudo service apache2 restart
You should now be able to securely communicate with your web service if it is running on port 8080.
One great part of this configuration is that all SSL encryption and decryption is being handled by Apache and you don't have to rely on the implementation inside whatever Swift framework you are using.
Running Your Binary as a Service
We have a great setup so far, but at this point you still need to be logged in and manually running your program for it to handle the requests. We want our binaries to run automatically without us having to be logged in. Even more than that, if they happen to crash, we want them to restart automatically. To do this, we will run them as a service.
On Ubuntu 16.04 I use the service manager called systemd. This allows me to configure a service defined by an arbitrary command. It also allows the automatically automatic boot start and restart we are looking for.
To create a new service you need to create a new service file inside /etc/systemd/system.
I like to name my services after the domain they are intented for so I created a file called drewag.me.service:
sudo nano /etc/systemd/system/drewag.me.service
For our purpose, there are three parts to this file. First we have the "Unit" section:
# /etc/systemd/system/drewag.me.service [Unit] Description=Drewag.me Blog After=network.target
Here we set a human readable description of the service, in this case I have it set as "Drewag.me Blog". Next, with
After variable, we indicate that we don't want this service to start until after the network interfaces are
The following section is the "Service" section that defines the most critical configurations:
# /etc/systemd/system/drewag.me.service [Service] User=www-data Restart=on-failure RestartSec=90 WorkingDirectory=/var/www/drewag.me Environment=LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/var/www/drewag.me/.build/release/ ExecStart=/var/www/drewag.me/.build/release/drewag.me server 8080
With the first variable I indicate that I want the service to run as the "www-data" user. This addresses the concern of what permissions the program will have.
I lock down the permissions of www-data as much as I can. Next, I ask that the service automatically restart when it fails and specify it should restart after 90
WorkingDirectory variable defines where the program is run from. I like to have my web services run from the /var/www/ directory as that is a
conventional place to put websites and web services. It is important that my program is run from that directory because there are assets in that directory that
the program needs access to. I have also found that some of the third party packages I use need a properly defined LDLIBRARYPATH variable, presumably because
they have dynamic libraries they are linking to. That is what the
Environment variable is setting. Finally, I use the
ExecStart variable to define the command
I want executed to start the service. I specify the full path to where the binary is compiled and pass any relevant arguments. My example command starts up a server
on port 8080, but these arguments are specific to how I developed my program; yours will almost certainly differ.
The final section we need is the "Install" section.
# /etc/systemd/system/drewag.me.service [Install] WantedBy=multi-user.target
This specifies when we want our service installed. To be honest, I only have
multi-user.target because I have seen it used in lots of examples and it has
worked for me.
Once that is setup, you can start your service by running:
sudo service drewag.me start
Now your service will run in the background and automatically restart if necessary.
Gracefully Handling Downtime
Last is the part of my configuration that I am most proud of. Our services may restart automatically but there will still be 90 seconds there will it will be down and it is possible that it will fail to start back up. During those down times, I want a better response for users than the default "503 Service Unavailable" that apache will return. Instead, I have Apache serve up a static page telling my users that the site is down and should be up shortly.
To do this, I will use a feature of Apache called balancers. They are primarily intended for larger websites that have so much traffic that they need multiple servers to handle the load. Those sites put a balancer in front of all their servers in order to distrubute traffic evenly, which minimizing the possiblity of a single server being overwhelmed. For our purposes, we want to use a feature of balancers that allows specifying a fallback server in the scenario where the main server does not respond.
For my example, I will show an HTTP configuration but the same applies for the SSL configuration. First thing we want to do is setup our static page. We can do this in Apache:
# /etc/apache2/sites-available/dev.drewag.me Listen 8082 <VirtualHost *:8082> AliasMatch "^/.+$" "/var/www/dev.drewag.me/site-down.html" </VirtualHost>
Now all we need to do is pass our web requests through a balancer that will use port 8082 as a backup:
# /etc/apache2/sites-available/dev.drewag.me <VirtualHost *:80> ServerName dev.drewag.me <Proxy balancer://hotcluster> BalancerMember http://localhost:8081 timeout=15 retry=30 BalancerMember http://localhost:8082 status=+H </Proxy> ProxyPreserveHost On ProxyPass / balancer://hotcluster/ ProxyPassReverse / balancer://hotcluster/ </VirtualHost>
Here we have added in a configuration for a balancer called "hotcluster" and we are directing traffic through
the proxy there instead of directly to our service (notice the changes to the
In the balancer we specified two members. The first is the real service. It is set to timeout at 15 seconds and if it is detected
down, the balancer will try to use it again after 30 seconds. All traffic in the meantime will fallback to our static site on port
status=+H part of the command specifies that it is a "hot standby" simply meaning that traffic should not be sent to
it unless all the main members are not returning a response.
For this configuration to work, I had to make sure the following Apache mods were enabled:
sudo a2enmod http proxy_http proxy_balancer lbmethod_byrequests
After enabling them, restart apache again:
sudo service apache2 restart
Now you should be able to take down your service and when you visit your domain name, you should see your static page. To stop your service you can issue the command:
sudo service drewag.me stop
This technique allows us to handle the reality that compiled programs, like those written in Swift, are more likely to crash than traditional web technologies. Swift may be a very safe language but other web technologies tend to avoid completely crashing at all costs because of this very problem. That leads me to my last piece of advice.
Tracking Your Server's Reliablity
This is relevant for all web services, but it is especially important when we are on the cutting edge of new technologies. We want to be immediately aware of any problem with our service before many users experience it. We don't want to be hearing from users that our site is down. There are several services that will monitor the status of your site but my favorite is called Pingdom. It has other more advanced features, but I use its ability to make a request to my domain every minute and notify me if it ever recieves an error. I had a bug at one point that was causing my service to crash occasionally, but because the service automatically restarts I was never seeing the problem. It was Pingdom that showed me that it was occasionally recieving an error and that lead me to look through my logs and figure out what the problem was. I discussed the problem I ran into on my post about Filling Holes in Swift on Linux.
Ultimately, there are a lot of ways you could deploy your services. Golang is another compiled language that has been around longer than Swift and so that community will also have solutions. This is a very large field with people who dedicate their careers to managing servers for online websites and services. However, this confuguration is relatively simple, it has served me well in small scale environments, and it definitely has the potential to scale up much larger. I hope you found some value in it!