I'm currently researching the best method for running a static website from Docker.
The site consists of one single HTML file, a bunch of CSS files, and a few JS files. On server-side nothing needs to be preprocessed. The website uses JS to request some JSON files, though. Handling of the files is doing via client-side JS, the server only need to - serve the files.
The website is intended to be used as selfhosted web application and is quite niche so there won't be much load and not many concurrent users.
For all of the variants I found information online. From the options I found I actually prefer the BusyBox route because it seems the cleanest with the least amount of overhead (I just need to serve the files, the rest is done on the client).
Do you have any other ideas? How do you host static content?
Just go nginx, anything else is faffing about. Busybox may not be security tested, so best to avoid on the internet. Php is pointless when its a static site with no php. Id avoid freenginx until its clear that it is going to be supported. There is nothing wrong with stock nginx, the fork is largely political rather than technical.
Php is pointless when its a static site with no php
Absolutely, but it has a built-in webserver that can serve static files, too (I constantly use that in my dev environment).
But I guess you're mostly right about just using Nginx. I already have multiple containers running it, though. Most of them just serving static files. But it's ca. 50 megabytes compressed size each container just for Nginx alone.
Absolutely, but it has a built-in webserver that can serve static files, too (I constantly use that in my dev environment).
How about Python? You can get an HTTP server going with just python3 -m http.server from the dir where the files are. Worth remembering because Python is super common and probably already installed in many places (be it on host or in containers).
I personally package the files in a scratch or distroless image and use https://github.com/static-web-server/static-web-server, which is a rust server, quite tiny. This is very similar to nginx or httpd, but the static nature of the binary removes clutter, reduces attack surface (because you can use smaller images) and reduces the size of the image.
Thanks, this looks actually pretty great. From the description it's basically BusyBox httpd but with Nginx stability and production-readiness and functionality. It also seems to be actively developed.
I just use nginx in docker. It runs from a Pi4 so needs to be lightweight. I'm sure there are lighter httpd servers to use, but it works for me. I also run nginx proxy manager to create a reverse proxy and to manage the certificate renewal that comes from Let's Encrypt.
My setup already has Nginx Proxy Manager to handle SSL. This is specifically about serving files from within a docker container with as little overhead as possible.
I see from your other comments that you're already running nginx in other containers. The simplest solution would be to make use of one of them. Zero overhead since you're not adding any new container. 🙂
You mentioned you're using NPM, well NPM already has a built-in nginx host that you can reach by making a proxy host pointed at http://127.0.0.1:80 and adding the following to the "Advanced" tab:
location / {
root /data/nginx/local_static;
index index.html;
}
Replace the root location with whatever dir you want, use a volume option on the NPM container to map the dir to the host, put your files in there and that's it.
Yeah it's not exactly an obvious feature. I don't even remember how I stumbled onto it, I think I was looking at the /data dirs and noticed the default one.
I haven't tried using it for more than one site but I think that if you add multiple domain names to the same proxy host they go to the same server instance and you might be able to tweak the "Advanced" config to serve all of them as virtual hosts.
It's not necessarily a bad thing to have a separate nginx host. For example I have a PHP app that has its own nginx container because I want to keep all the containers for it in one place and not mix it up with NPM.
The busybox one seems great as it comes with shells. php looks like it would add some issues.
Personally since I use go, I would create a go embedded app, which I would make a deb, rpm, and a dockerfile using "goreleaser"
package main
import (
"embed"
"net/http"
)
//go:embed static/*
var content embed.FS
func main() {
http.HandleFunc("/", func(w http.ResponseWriter, r *http.Request) {
// Serve index.html as the default page
http.ServeContent(w, r, "index.html", nil, content)
})
// Serve static files
http.Handle("/static/", http.StripPrefix("/static/", http.FileServer(http.FS(content))))
// Start the server
http.ListenAndServe(":8080", nil)
}
Would be all the code but allows for expansion later. However the image goreleaser builds doesn't come with busybox on it so you can't docker exec into it. https://goreleaser.com/customization/docker/
Most of the other options including the PHP one seem to include a scripting language or a bunch of other system tools etc. I think that's overkill
I would consider the lack of a shell a benefit in this scenario. You really don't want the extra attack surface and tooling.
Considering you also manage the host, if you want to see what's going on inside the container (which for such a simple image can be done once while building it the first time more likely), you can use unshare to spawn a bash process in the container namespaces (e.g., unshare -m -p [...] -t PID bash, or something like this - I am going by memory).
I’ve always used an nginx alpine image and have been very happy with it.
Not sure how this fork business is turning out and I have also heard conflicting opinions on wether to care or not…
If you do wish for something simple that is not nginx I’m also very happy with caddy, which can also handle ssl certificates for you, if you plan to make it publicly reachable.
If your looking for a small size you could build a custom image with buildroot and lighttpd. It is way, way overkill but it would be the smallest.
For something easier use the latest image of your web server of choice and then pass though a directory with the files. From there you can automate patching with watch tower.
First thing you mention is such a fun and useful exercise. But as you point out, way overkill. Might even be dangerous to expose it. I got mine to 20kb on top of busybox.
There is something that tickles the right spots when a complete container image significantly smaller than the average js payload in "modern" websites.
The simplest way is certainly to use a hosted service like GitHub Pages. These make it so easy to create static websites.
If you’re not flexible on that detail, then I next recommend Go actually. You could write a tiny web server and embed the static files into the app at build time. In the end, you’d have a single binary that acts as a web server and has your content. Super easy to dockerize.
Things like authentication will complicate the app over time. If you need extra features like this, then I recommend using common tools like nginx as suggested by others.
Containers are a perfectly suitable use-case for serving static sites. You get isolation and versioning at the absolutely negligible cost of duplicating a binary (the webserver - which in case of the one I linked in my comment, it's 5MB of space). Also, you get autostart of the server if you use compose, which is equivalent to what you would do with a Systemd unit, I suppose.
You can then use a reverse-proxy to simply route to the different containers.
But it you already have an nginx or other web server otherwise required to start up (which is in all likelihood the case), you don't need any more auto startup, the "reverse proxy" already started can just serve it. I would say that container orchestration versioning can be helpful in some scenarios, but a simple git repository for a static website is way more useful since it's got the right tooling to annotate changes very specifically on demand.
That reverse proxy is ultimately also a static file server. There's really no value in spinning up more web servers for a strictly static site.
Folks have gone overboard assuming docker or similar should wrap every little thing. It sometimes adds complexity without making anything simpler. It can simplify some scenarios, but adding a static site to a webserver is not a scenario that enjoys any benefit.
Because serving static files doesn't really require any flexibility in web serving code.
If your setup has an nginx or similar as a reverse proxy entry point, you can just tell it to serve the directory. Why bother making an entire new chroot and proxy hop when you have absolutely zero requirements beyond what the reverse proxy already provides. Now if you don't have that entry point, fine, but at least 99% of the time I see some web server as initial arbiter into services that would have all the capability to just serve the files.
I already have a fully set up docker environment that serves all sorts of things (including some containers that serve special static content using Nginx).
Yes, Freenginx should/would/will be a drop-in replacement, at least int he beginning. We'll see how this works out over time. Forks purely out of frustration never lived long enough to gain a user base and attract devs. But it's an "anti corporate bullshit" fork and this alone puts it on my watchlist.
The old age of the Docker image is a bit of a red flag to me.
I settled with SWS since the Docker image and a locally installable version are actively maintained by the creator. It just serves static files and optionally directory listing as JSON (which comes in quite handy).