Your repo should be easy to build, and how
Whenever I look at a new repository, I have a simple smell test: how long does it take me to clone, build and get the code running?
Here, I’m usually counting the steps I have to do, the commands I have to run. The clock time is less important (although fast builds are still nice!). Ideally, there’s a single command which takes me from a fresh checkout to a complete build — and without me having to fiddle with too many dependencies first.
Once I have a working build, I can start fiddling with code and find my own way around. Getting that first build is key.
Making it easy to do a clean build has many benefits.
The obvious one is time saved — I run one command, then I can walk away while the computer does all the slow bits. Downloading dependencies, compiling code, setting up the local environment, that sort of thing. It might take a while to finish, but I don’t need to supervise it while that happens. I can spend that time doing something more useful.
It’s also more reliable. Remembering “make build” is easy. Remembering eight calls to different shell scripts, and their associated arguments, is much harder. If the build is simple, there’s less to get wrong, and it’s more likely I’ll get it right first time. Automating the build process makes it faster and more reliable.
And finally, first impressions count! Being able to start working quickly is a pleasant experience. If writing and testing my first patch is easy, I’m more likely to do it a second time. And a third. And so on. This is particularly important in open source repos where patches often come from people giving up their time for free.
In the last year, I’ve spent a lot of time simplifying my build processes, both in my work and my personal repos. Most of my current repos now have a single-step build. It’s not perfect, but I’m very pleased with the results.
In this post, I’ll explain my typical setup, and how I use Make and Docker to get fast and reliable builds.
Starting work on a freshly installed Mac: I clone our main repo, type "make id_minter-build", and it starts fetching dependencies.
Ten minutes later, with nothing else from me, I've got a complete build.
I really felt the benefits last Monday. Normally I commute to London on the train, but there was snow and ice at the weekend, and cold weather wreaks havoc on the railways. I decided to work from home, rather than sit on a train that would probably be delayed or cancelled.
Because I wasn’t expecting to work from home when I left on Friday, I’d left my work laptop in the office. And even worse, I recently had to reinstall my home computer, so I didn’t have any of my existing setup.
Not to worry: all our projects are on GitHub so I could easily get the code. (And they’re open source with an MIT license, so working on a personal computer doesn’t cause legal issues.) We run all our dev processes in Docker containers, so I only had to install Docker. Then we use Make to manage Docker itself: building the containers, keeping them up-to-date, and running the test commands.
So after I’d cloned the repo, I could run my tests with just one command:
$ make api-test
followed by ten minutes of dependencies compiling in the background, which I could spend catching up on email. When I came back, the code had compiled, the tests run, and I was ready to start work.
Other projects I’ve worked on took much longer to get set up. At a previous job, the instructions for your first build ran to two sides of A4 (!). Tweaking compiler flags, laying out submodules, getting my network settings ~just so – I’d been working for a fortnight before I saw “build: ok”.
If I had to go through something like that, I’d have wasted the whole day. Instead, I was back up-and-running in less than half an hour.
How to set up Make and Docker for one-step builds
Let’s start without Make or Docker. Our API is a Scala application, with tests and builds managed by a tool called sbt. We also run mocks/stubs of a few AWS services in our tests, which are Docker containers orchestrated by Docker Compose.
So if you have the right versions of sbt, Scala and Docker/Compose installed, you run our tests with the command:
$ sbt "project api" ";dockerComposeUp;test;dockerComposeStop"
My first Makefile
The first thing we can do is write a Makefile, and a Make rule that wraps this command:
# Makefile api-test: sbt "project api" ";dockerComposeUp;test;dockerComposeStop"
Now we can run
make api-test, and that has the same effect as running the sbt command, but we’ve hidden the details behind a human-friendly name.
You don’t have to remember the exact sbt command, just api-test. Tests for our other applications have a similar naming scheme: transformer-test, loris-test, ingestor-test, and so on — you don’t have to worry how to invoke sbt, or even if those tests use sbt at all! And if the test command ever changes, you only have to edit the Makefile once, and everybody picks up the change.
If you already know how Make works, you can skip to the next section. If you’re unfamiliar with Make, there’s a quick primer below.
target: [dependency dependency ...] [command 1] ... [command n]
A Makefile is a series of rules. Each rule starts with a dependency line, with a target before the colon, and a list of dependencies after the colon. In the first example, the target is
api-test, and there are no dependencies. Often the target is the name of the file you want to create, but a rule doesn’t have to create a file.
Each dependency line is followed by a series of tab-indented commands which will be run to build this target (and the indents have to be tabs, not spaces). In the first example, the rule has a single command: the sbt call that runs the API tests.
Here’s what happens when you run
Make looks to see if the target already exists.
If the target already exists, it looks at the modified dates: is the target newer than all the dependencies? If so, it’s already up-to-date, and there’s nothing to do. If not, the target is out-of-date, and needs to rebuilt.
If it doesn’t exist, it needs to be built from scratch.
If the target needs to be built, Make runs the commands in sequence. If any of the commands fail (have a non-zero exit code), the build has failed and Make will return an error.
We’ll see some more complicated examples in the rest of the post. Check you understand what they’re doing.
Pushing it inside Docker
If I want to run these tests, I need to install Scala, sbt, and docker-compose on my local machine. I’d better hope they’re available in my package manager, and I can install them easily, and they don’t break something else I’ve installed… ick.
This is precisely the sort of problem that can be solved with Docker. Rather than installing dependencies on our main machine, we can install them in a Docker image, and run our tests there. Everything runs in a pristine, reproducible environment, and the installation process can be entirely automated.
Here’s a Dockerfile for installing our dependencies:
# sbt_wrapper.Dockerfile FROM pvansia/scala-sbt:0.13.13 RUN apk update && apk add docker py-pip RUN pip install docker-compose WORKDIR /repo ENTRYPOINT ["sbt"]
This file defines a new Docker image:
- It starts from an existing image that has sbt and Scala installed
- Then it runs apk and pip (package managers for Alpine Linux and Python) to install our extra dependencies
- It sets a working directory
- It prepares to run
sbtwhen we run this container; any arguments we put on the end of
docker runwill be passed to sbt
The Dockerfile reference explains the syntax in much more detail.
We can build this image with
docker build, which we’ll wrap inside a Make rule:
build-sbt-docker-image: docker build --tag sbt_wrapper --file sbt_wrapper.Dockerfile .
And then we can modify our first Make rule to use the Docker image:
api-test: docker run \ --volume $$(pwd):/repo \ --volume /var/run/docker.sock:/var/run/docker.sock \ --net host \ sbt_wrapper "project api" ";dockerComposeUp;test;dockerComposeStop"
docker run as follows:
- Sharing our repo into the container, so it has all our code
- Sharing the Docker socket with the container, so processes inside the container can run Docker commands on the host machine (in particular, starting the stub containers)
- Share the host network with the container, so it can see the new Docker containers it starts
- Then the name of the image to run, and the two final arguments which are passed through to
So now we have to build the image, then run the tests:
$ make build-sbt-docker-image $ make api-test
We’re improving: we only need to install Docker, dependency installation is automated, and we still have human-friendly ways to run our tasks. But we have to remember to build the Docker image first, or this doesn’t work. Can we automate that as well?
Make dependency management, take 1
Remember that Make allows us to declare dependencies for a rule. If we add dependencies after the colon, Make will try to build those first before it builds the main target.
So we can tell Make that api-test depends on build-sbt-docker-image like so:
api-test: build-sbt-docker-image ...
And we’re back to running:
$ make api-test
When you run that command, Make will build a new Docker image, then run your tests inside the image. Hooray!
So we’re done… right?
Make dependency management, part 2
If you try this out, you’ll find that Make rebuilds your Docker image every time you run your tests. Because Make never sees a file called build-sbt-docker-image, it assumes it has to rebuild it, and re-runs the rule. Docker caching means the rebuild is fairly fast on subsequent runs, but it still clutters up your console with logs. Could we do better?
We need Make to know when we’ve built the Docker image, so let’s modify the Make rule to drop a marker after it’s build the image. The next time Make runs, it’ll see the marker, know the image has already been built, and skip rebuilding it.
This is the pattern I typically use:
.docker/sbt_wrapper: sbt_wrapper.Dockerfile docker build --tag sbt_wrapper --file sbt_wrapper.Dockerfile . mkdir -p .docker touch .docker/sbt_wrapper api-test: .docker/sbt_wrapper ...
(Notice that I’ve also added a dependency on the Dockerfile, so if the Dockerfile does change, we’ll get a one-off rebuild.)
.docker directory is gitignored, and it serves as a collection of markers for Docker images.
This pattern makes a small difference to build times, but a big difference to console noise.
Putting this into practice
I’ve been using this system for about half a year, and it’s worked remarkably well. I have one-step rebuilds, human-friendly names for my build process, and a really mature and powerful dependency management system. You can see some examples in the platform repo’s Makefiles, or the Makefile for this blog.
It works particularly well for polyglot repos, where devs may have varying levels of familiarity with any given language. Somebody can get started without needing to learn the intricacies of the install process for a new language.
If you’re looking for a way to simplify or improve your build processes, I’d really encourage giving these technique a look.