Container management with Docker
Container Shipyard
Docker is an economical alternative to conventional virtualization. Because each Docker container shares the underlying operating system, it enjoys the resource isolation and allocation benefits of VMs but is much more portable and efficient.
Rarely has such young software raised so much attention in such a short time as the Docker [1] container manager. Docker is one of the fastest growing open source developments in the history of free software. Released in March 2013, it made it to version 1.0 in June 2014 and was declared ready for production. In the same month, Docker was already part of Red Hat's RHEL 7 release. Further cooperative agreements exist with Canonical, openSUSE, CoreOS, VMware, IBM and others. Moreover, Docker is a member of Google's Kubernetes project [2].
The image associated with the name Docker sums it up: a ship (computer) is loaded with containers that you can stack, transport, isolate, and protect (Figure 1). Docker containers in IT are isolated virtual environments for isolated execution of processes and applications. Docker is currently used primarily for developing, testing, and deploying applications and can isolate them along with their dependencies in containers. These applications have practically no overhead and can run nearly as fast as native applications.
Unlike the well-known hypervisors [3] VirtualBox, VMware Workstation, HyperV, Parallels Desktop, or Xen, a Docker container doesn't need its own operating system. Each container uses the basic functions of its host system and is therefore self-sufficient and isolated from other containers. Only the host's kernel needs to match that of the guest.
The Basics
Docker uses the group (cgroup)
[4] and kernel namespace
[5] technologies. Namespaces serve to isolate and protect, whereas cgroups ensure that the container can use the resources selectively granted to it by the host. Until version 0.8 (1.3 is the current version) Docker also availed itself of the functions of Linux Containers (LXCs) [6] but eventually developed its own API named LibContainer
in version 0.9 that is also available to other projects.
Currently, Docker is mainly used as a Platform as a Service (PaaS) [7], but it is becoming quite suitable as a resource-friendly alternative to desktop virtualization (Figure 2). Graphical tools are being developed to handle containers, and I will present one of them, Panamax, later in this article as one of the first interfaces its kind.
The Components
The Docker system consists of three main components: a daemon, a client, and images. The daemon is responsible for orchestrating the containers; the client is what you use to control Docker with commands. The client and daemon can run on the same machine or the client can contact the daemon on a data center server or over the cloud. Docker images are the equivalent of virtual machines (VMs) for VirtualBox or appliances in VMware. Compared with these, Docker images are extremely space-saving, can be stacked, and can be versioned in a CVS [8].
The Docker hub is laid out as Software as a Service (SaaS) [9] and includes the repositories that now have around 15,000 complete image recipes (Figure 3). These are different from base images, which represent a base operating system along with related images containing one or more applications. Creating your own images isn't rocket science after a bit of training, but you need to know the command language of the distribution. An example would be to create an image using WordPress, PHP, MariaDB, SSH, or Nginx.
Overlaid Layers
Thanks to the overlay functions supplied by Aufs [10] or the Device mapper [11], Docker has an efficient back end to store the containers. Each container, to put it simply, is stored as a transparent layer. These functions are the same ones that allow us to seemingly write to a Live CD (Figure 4). Quite recently, Btrfs [12] has been playing a role as well with its snapshot function. Details on the different filesystem functions for storing and managing containers can be found in Red Hat's documentation for their Atomic [13] project, a container operating system [14] still being developed.
Before I start going into details, I'll clarify things a bit. A Docker image is just a read-only template, a kind of recipe. A base image [15] provides rudimentary operating system functions and can have a parent-child relationship with other images that interact with the base image on various layers. For example, this could be an ownCloud image, which I'll get to later.
As a rule, containers consist of several images that contain all the functionality and dependencies that the content of the container (an operating system, application, or service) needs at run time. The basic handling of Docker containers is through a command-line interface resembling Git [16] and is easy to learn. The "Containers, Images, and Docker Files" box provides more information on how things work.
Containers, Images, and Docker Files
To understand better how an image is created, look at the description of the ownCloud image in the Docker hub [17]. The Dockerfile
tab shows the important parts of the image. An image also includes licensing information and a README file that goes into specifics.
It's pretty clear what this image does and why it's a template. The FROM
line at the top identifies the base image being used – in this case, Debian Stable (Figure 5). When you download Dockerfile, each line beginning with RUN
is run in sequence. First, wget
is installed with an archive key, the ownCloud repository, and its dependencies.
The template then sets pre-installations of the Apache web server, which, in this case, is in the base image or needs to be installed later. You can then load this complete image into a container where the base image is also loaded. Containers are by default isolated through "deny all" policies and can't communicate with the outside world. If you do want this, use EXPOSE
parameters with port numbers to expose. The VOLUME
parameter allows connecting with the host's filesystems or another container at a specified location. Since Docker 1.2, the containers' rights and privileges can be controlled granularly [18] so that the combination of cgroups and namespaces can meet their full potential.
A container thus consists of a base operating system with apps and metadata added by users. The image tells Docker what the container should include and, where appropriate, what command is used for startup. In the simplest case the container starts right away with a shell, although far more complicated processes can occur, such as automated test runs.
Because images are read-only, Docker opens a layer for the containers with help from a Union filesystem, such as Aufs or Device mapper. These layers are writeable (wr
), and the application runs within them. If a change occurs, such as an update, another layer is applied. Instead of changing the entire image, changes at run time are integrated only when you convert the container into an image again using:
docker commit
Docker containers come two ways: interactively started or daemonized. If the container is to execute a predetermined sequence and then stop, start it interactively with:
docker run -t -i debian:sid </command_1;command_2>
A container started with
docker run -d debian:sid /bin/bash
waits in the background for input.
Buy this article as PDF
(incl. VAT)
Buy Linux Magazine
Subscribe to our Linux Newsletters
Find Linux and Open Source Jobs
Subscribe to our ADMIN Newsletters
Support Our Work
Linux Magazine content is made possible with support from readers like you. Please consider contributing when you’ve found an article to be beneficial.
News
-
New Slimbook EVO with Raw AMD Ryzen Power
If you're looking for serious power in a 14" ultrabook that is powered by Linux, Slimbook has just the thing for you.
-
The Gnome Foundation Struggling to Stay Afloat
The foundation behind the Gnome desktop environment is having to go through some serious belt-tightening due to continued financial problems.
-
Thousands of Linux Servers Infected with Stealth Malware Since 2021
Perfctl is capable of remaining undetected, which makes it dangerous and hard to mitigate.
-
Halcyon Creates Anti-Ransomware Protection for Linux
As more Linux systems are targeted by ransomware, Halcyon is stepping up its protection.
-
Valve and Arch Linux Announce Collaboration
Valve and Arch have come together for two projects that will have a serious impact on the Linux distribution.
-
Hacker Successfully Runs Linux on a CPU from the Early ‘70s
From the office of "Look what I can do," Dmitry Grinberg was able to get Linux running on a processor that was created in 1971.
-
OSI and LPI Form Strategic Alliance
With a goal of strengthening Linux and open source communities, this new alliance aims to nurture the growth of more highly skilled professionals.
-
Fedora 41 Beta Available with Some Interesting Additions
If you're a Fedora fan, you'll be excited to hear the beta version of the latest release is now available for testing and includes plenty of updates.
-
AlmaLinux Unveils New Hardware Certification Process
The AlmaLinux Hardware Certification Program run by the Certification Special Interest Group (SIG) aims to ensure seamless compatibility between AlmaLinux and a wide range of hardware configurations.
-
Wind River Introduces eLxr Pro Linux Solution
eLxr Pro offers an end-to-end Linux solution backed by expert commercial support.