Skip to content

PrivateBin/docker-unit-alpine

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

PrivateBin on Nginx Unit & Alpine

PrivateBin is a minimalist, open source online pastebin where the server has zero knowledge of pasted data. Data is encrypted and decrypted in the browser using 256bit AES in Galois Counter mode.

This repository contains the Dockerfile and resources needed to create a docker image with a pre-installed PrivateBin instance in a secure default configuration. The images are based on the docker hub Alpine image, extended with the GD module required to generate discussion avatars and the Nginx Unit application server to serve static JavaScript libraries, CSS & the logos as well as dynamic PHP rendered HTML. All logs of Nginx Unit (access & errors) are forwarded to docker logs.

Limitations

Compared to the Nginx web server, php-fpm & Alpine images, these are smaller, but lack the following features:

  • no automatic gzip compression of static text files
  • nginx, the webserver and nginx unit are not the same thing. "Unit is a lightweight and versatile application runtime [and] was created by nginx team members from scratch [...]."

You can use a front end webserver that addresses these limitations or use the other images that use a full nginx webserver, offering these functions out of the box.

Image variants

This is the all-in-one image (Docker Hub / GitHub) that can be used with any storage backend supported by PrivateBin - file based storage, databases, Google Cloud or S3 Storage. We also offer dedicated images for each backend:

Image tags

All images contain a release version of PrivateBin and are offered with the following tags:

  • latest is an alias of the latest pushed image, usually the same as nightly, but excluding edge
  • nightly is the latest released PrivateBin version on an upgraded Alpine release image, including the latest changes from the docker image repository
  • edge is the latest released PrivateBin version on an upgraded Alpine edge image
  • stable contains the latest PrivateBin release on the latest tagged release of the docker image git repository - gets updated when important security fixes are released for Alpine or upon new Alpine releases
  • 1.5.1 contains PrivateBin version 1.5.1 on the latest tagged release of the docker image git repository - gets updated when important security fixes are released for Alpine or upon new Alpine releases, same as stable
  • 1.5.1-... are provided for selecting specific, immutable images

If you update your images automatically via pulls, the stable, nightly or latest are recommended. If you prefer to have control and reproducability or use a form of orchestration, the numeric tags are probably preferable. The edge tag offers a preview of software in future Alpine releases and serves as an early warning system to detect image build issues in these.

Image registries

These images are hosted on the Docker Hub and the GitHub container registries:

Running the image

Assuming you have docker successfully installed and internet access, you can fetch and run the image from the docker hub like this:

$ docker run -d --restart="always" --read-only -p 8080:8080 -v $PWD/privatebin-data:/srv/data privatebin/unit-alpine

The parameters in detail:

  • -v $PWD/privatebin-data:/srv/data - replace $PWD/privatebin-data with the path to the folder on your system, where the pastes and other service data should be persisted. This guarantees that your pastes aren't lost after you stop and restart the image or when you replace it. May be skipped if you just want to test the image or use database or Google Cloud Storage backend.
  • -p 8080:8080 - The Nginx webserver inside the container listens on port 8080, this parameter exposes it on your system on port 8080. Be sure to use a reverse proxy for HTTPS termination in front of it in production environments.
  • --read-only - This image supports running in read-only mode. Using this reduces the attack surface slightly, since an exploit in one of the images services can't overwrite arbitrary files in the container. Only /tmp, /var/tmp, /var/run & /srv/data may be written into.
  • -d - launches the container in the background. You can use docker ps and docker logs to check if the container is alive and well.
  • --restart="always" - restart the container if it crashes, mainly useful for production setups

Note that the volume mounted must be owned by UID 65534 / GID 82. If you run the container in a docker instance with "userns-remap" you need to add your subuid/subgid range to these numbers.

Custom configuration

In case you want to use a customized conf.php file, for example one that has file uploads enabled or that uses a different template, add the file as a second volume:

$ docker run -d --restart="always" --read-only -p 8080:8080 -v $PWD/conf.php:/srv/cfg/conf.php:ro -v $PWD/privatebin-data:/srv/data privatebin/unit-alpine

Note: The Filesystem data storage is supported out of the box. The image includes PDO modules for MySQL and PostgreSQL, required for the Database one, but you still need to keep the /srv/data persisted for the server salt and the traffic limiter when using a release before 1.4.0.

Environment variables

The following variables do get passed down to the PHP application to support various scenarios. This allows changing some settings via the environment instead of a configuration file. Most of these relate to the storage backends:

Amazon Web Services variables used by the S3 backend
  • AWS_ACCESS_KEY_ID
  • AWS_CONTAINER_AUTHORIZATION_TOKEN
  • AWS_CONTAINER_CREDENTIALS_FULL_URI
  • AWS_CONTAINER_CREDENTIALS_RELATIVE_URI
  • AWS_DEFAULT_REGION
  • AWS_PROFILE
  • AWS_ROLE_ARN
  • AWS_ROLE_SESSION_NAME
  • AWS_SECRET_ACCESS_KEY
  • AWS_SESSION_TOKEN
  • AWS_STS_REGIONAL_ENDPOINTS
  • AWS_WEB_IDENTITY_TOKEN_FILE
  • AWS_SHARED_CREDENTIALS_FILE
Google Cloud variables used by the GCS backend
  • GCLOUD_PROJECT
  • GOOGLE_APPLICATION_CREDENTIALS
  • GOOGLE_CLOUD_PROJECT
  • PRIVATEBIN_GCS_BUCKET
Custom backend settings

The following variables are not used by default, but can be enabled in your custom configuration file, to keep sensitive information out of it:

  • STORAGE_HOST
  • STORAGE_LOGIN
  • STORAGE_PASSWORD
  • STORAGE_CONTAINER
Configuration folder
  • CONFIG_PATH
Timezone settings

The image supports the use of the following two environment variables to adjust the timezone. This is most useful to ensure the logs show the correct local time.

  • TZ
  • PHP_TZ

Note: The application internally handles expiration of pastes based on a UNIX timestamp that is calculated based on the timezone set during its creation. Changing the PHP_TZ will affect this and leads to earlier (if the timezone is increased) or later (if it is decreased) expiration then expected.

Adjusting Nginx Unit or PHP settings

You can attach your own php.ini the folder /etc/php/conf.d/. You can dynamically change the Nginx Unit configuration at runtime via it's Unix socket at /run/control.unit.sock - if you want to persist the Unit configuration changes, you need to attach a persistent volume to /var/lib/unit. This would for example let you adjust the maximum size the service accepts for file uploads, if you need more than the default 10 MiB.

Kubernetes deployment

Below is an example deployment for Kubernetes.

---
apiVersion: apps/v1
kind: Deployment
metadata:
  name: privatebin-deployment
  labels:
    app: privatebin
spec:
  replicas: 3
  selector:
    matchLabels:
      app: privatebin
  template:
    metadata:
      labels:
        app: privatebin
    spec:
      securityContext:
        runAsUser: 65534
        runAsGroup: 82
        fsGroup: 82
      containers:
      - name: privatebin
        image: privatebin/unit-alpine:stable
        ports:
        - containerPort: 8080
        env:
        - name: TZ
          value: Antarctica/South_Pole
        - name: PHP_TZ
          value: Antarctica/South_Pole
        securityContext:
          readOnlyRootFilesystem: true
          privileged: false
          allowPrivilegeEscalation: false
        livenessProbe:
          httpGet:
            path: /
            port: 8080
        readinessProbe:
          httpGet:
            path: /
            port: 8080
        volumeMounts:
        - mountPath: /srv/data
          name: privatebin-data
          readOnly: False
        - mountPath: /run
          name: run
          readOnly: False
        - mountPath: /tmp
          name: tmp
          readOnly: False
  volumes:
    - name: run
      emptyDir:
        medium: "Memory"
    - name: tmp
      emptyDir:
        medium: "Memory"

Note that the volume privatebin-data has to be a shared, persisted volume across all nodes, i.e. on an NFS share. As of PrivateBin 1.4.0 it is no longer required, when using a database or Google Cloud Storage.

Running administrative scripts

The image includes two administrative scripts, which you can use to migrate from one storage backend to another, delete pastes by ID, removing empty directories when using the Filesystem backend, to purge all expired pastes and display statistics. These can be executed within the running image or by running the commands as alternative entrypoints with the same volumes attached as in the running service image, the former option is recommended.

# assuming you named your container "privatebin" using the option: --name privatebin

$ docker exec -t privatebin administration --help
Usage:
  administration [--delete <paste id> | --empty-dirs | --help | --purge | --statistics]

Options:
  -d, --delete      deletes the requested paste ID
  -e, --empty-dirs  removes empty directories (only if Filesystem storage is
                    configured)
  -h, --help        displays this help message
  -p, --purge       purge all expired pastes
  -s, --statistics  reads all stored pastes and comments and reports statistics

$ docker exec -t privatebin migrate --help
migrate - Copy data between PrivateBin backends

Usage:
  migrate [--delete-after] [--delete-during] [-f] [-n] [-v] srcconfdir
          [<dstconfdir>]
  migrate [-h|--help]

Options:
  --delete-after   delete data from source after all pastes and comments have
                   successfully been copied to the destination
  --delete-during  delete data from source after the current paste and its
                   comments have successfully been copied to the destination
  -f               forcefully overwrite data which already exists at the
                   destination
  -h, --help       displays this help message
  -n               dry run, do not copy data
  -v               be verbose
  <srcconfdir>     use storage backend configuration from conf.php found in
                   this directory as source
  <dstconfdir>     optionally, use storage backend configuration from conf.php
                   found in this directory as destination; defaults to:
                   /srv/bin/../cfg/conf.php

Note that in order to migrate between different storage backends you will need to use the all-in-one image called privatebin/unit-alpine, as it comes with all the drivers and libraries for the different supported backends. When using the variant images, you will only be able to migrate within two backends of the same storage type, for example two filesystem paths or two database backends.

Rolling your own image

To reproduce the image, run:

$ docker build -t privatebin/unit-alpine .

Behind the scenes

Nginx Unit serves static files and caches them, too. Requests to the index.php (which is the only PHP file exposed in the document root at /var/www) are also processed by it using PHP as a SAPI module. All other PHP files and the data are stored under /srv.

The Nginx setup supports only HTTP, so make sure that you run a reverse proxy in front of this for HTTPS offloading and reducing the attack surface on your TLS stack. The Nginx in this image is set up to deflate/gzip text content.

During the build of the image, the PrivateBin release archive is downloaded from Github. All the downloaded Alpine packages and the PrivateBin archive are validated using cryptographic signatures to ensure they have not been tempered with, before deploying them in the image.