Skip to content

minT(oolkit): Mint awesome, secure and production ready containers just the way you need them! Don't change anything in your container image and minify it by up to 30x (and for compiled languages even more) making it secure too! (free and open source)

License

Notifications You must be signed in to change notification settings

mintoolkit/mint

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Follow

Discord chat

Gitpod ready-to-code

Install Mint

Get Examples

Optimize Your Experience with Containers. Make Your Containers Better, Smaller, More Secure and Do Less to Get There (free and open source!)

Note that DockerSlim is now MinToolkit or just Mint (it was also called SlimToolkit, but that name was too similar to a commercial company name and using a different name was a CNCF requirement for the project). It's no longer limited only to the Docker runtime and its support for additional container runtimes and tools is growing. The "min" part of the name represents the ability to create minimal container images, which is now the unofficial defacto name to call slim images. "min" is also short for minify. The "T" part is short for "toolkit". Together min and t is a good way to describe the core functionality where the tool mints/creates new and improved container images.

Mint was created by Kyle Quest and it's been improved by many contributors. New contributors are always welcome and the project is commited to providing more resources and support to make it easy to contribute.

Overview

Mint allows developers to inspect, optimize and debug their containers using its xray, slim (aka build), debug, lint, run, images, imagebuild, merge, registry, vulnerability (and other) commands. It simplifies and improves your developer experience building, customizing and using containers. It makes your containers better, smaller and more secure while providing advanced visibility and improved usability working with the original and minified containers.

Don't change anything in your container image and minify it by up to 30x making it secure too! Optimizing images isn't the only thing it can do though. It can also help you understand and author better container images.

Keep doing what you are doing. No need to change anything. Use the base image you want. Use the package manager you want. Don't worry about hand optimizing your Dockerfile. You shouldn't have to throw away your tools and your workflow to have small container images.

Don't worry about manually creating Seccomp and AppArmor security profiles. You shouldn't have to become an expert in Linux syscalls, Seccomp and AppArmor to have secure containers. Even if you do know enough about it wasting time reverse engineering your application behavior can be time-consuming.

Mint will optimize and secure your containers by understanding your application and what it needs using various analysis techniques. It will throw away what you don't need, reducing the attack surface of your container. What if you need some of those extra things to debug your container? You can use dedicated debugging side-car containers for that (more details below).

Mint has been used with Node.js, Python, Ruby, Java, Go, Rust, Elixir and PHP (some app types) running on Ubuntu, Debian, CentOS, Alpine and even Distroless.

Note that some application stacks do require advanced container probing to make sure that all dynamically loaded components are detected. See the --http-probe* flags for more details to know how you can define custom probe commands. In some cases you might also need to use the --include-path flag to make sure everything your application needs is included (e.g., ubuntu.com python SPA app container image example where the client side template files are explicitly included).

It's also a good idea to use your app/environment tests when you run the Mint app. See the --continue-after flag for more details about integrating your tests with the temporary container Mint creates when it's doing its dynamic analysis. Running tests in the target container is also an option, but it does require you to specify a custom ENTRYPOINT/CMD with a custom wrapper to start your app and to execute your tests.

Minify How

Interactive CLI prompt screencast:

asciicast

Watch this screencast to see how an application image is minified by more than 30x.

asciicast

When you run the slim (aka build) or profile commands in Mint it gives you an opportunity to interact with the temporary container it creates. By default, it will pause and wait for your input before it continues its execution. You can change this behavior using the --continue-after flag.

If your application exposes any web interfaces (e.g., when you have a web server or an HTTP API), you'll see the port numbers on the host machine you will need to use to interact with your application (look for the port.list and target.port.info messages on the screen). For example, in the screencast above you'll see that the internal application port 8000 is mapped to port 32911 on your host.

Note that Mint will interact with your application for you when HTTP probing is enabled (enabled by default; see the --http-probe* flag docs for more details). Some web applications built with scripting languages like Python or Ruby require service interactions to load everything in the application. Enable HTTP probing unless it gets in your way.

You can also interact with the temporary container via a shell script or snippet using --exec-file or --exec. For example, you can create a container which is only capable of using curl.

>> mint slim --target archlinux:latest --tag archlinux:curl --http-probe=false --exec "curl checkip.amazonaws.com"
...

>> docker run archlinux:curl curl checkip.amazonaws.com
...

>> docker images
archlinux                 curl                ...        ...         17.4MB
archlinux                 latest              ...        ...         467MB
...

Community

Feel free to join any of these channels or just open a new Github issue if you want to chat or if you need help.

Mint on the Internet

Books:

Minification Examples

You can find the examples in a separate repository: https://github.com/mintoolkit/examples

Node.js application images:

  • from ubuntu:14.04 - 432MB => 14MB (minified by 30.85X)
  • from debian:jessie - 406MB => 25.1MB (minified by 16.21X)
  • from node:alpine - 66.7MB => 34.7MB (minified by 1.92X)
  • from node:distroless - 72.7MB => 39.7MB (minified by 1.83X)

Python application images:

  • from ubuntu:14.04 - 438MB => 16.8MB (minified by 25.99X)
  • from python:2.7-alpine - 84.3MB => 23.1MB (minified by 3.65X)
  • from python:2.7.15 - 916MB => 27.5MB (minified by 33.29X)
  • from centos:7 - 647MB => 23MB (minified by 28.57X)
  • from centos/python-27-centos7 - 700MB => 24MB (minified by 29.01X)
  • from python2.7:distroless - 60.7MB => 18.3MB (minified by 3.32X)

Ruby application images:

  • from ubuntu:14.04 - 433MB => 13.8MB (minified by 31.31X)
  • from ruby:2.2-alpine - 319MB => 27MB (minified by 11.88X)
  • from ruby:2.5.3 - 978MB => 30MB (minified by 32.74X)

Go application images:

  • from golang:latest - 700MB => 1.56MB (minified by 448.76X)
  • from ubuntu:14.04 - 531MB => 1.87MB (minified by 284.10X)
  • from golang:alpine - 258MB => 1.56MB (minified by 165.61X)
  • from centos:7 - 615MB => 1.87MB (minified by 329.14X)

Rust application images:

  • from rust:1.31 - 2GB => 14MB (minified by 147.16X)

Java application images:

  • from ubuntu:14.04 - 743.6 MB => 100.3 MB

PHP application images:

  • from php:7.0-cli - 368MB => 26.6MB (minified by 13.85X)

Haskell application images:

  • (Scotty service) from haskell:8 - 2.09GB => 16.6MB (minified by 125.32X)
  • (Scotty service) from haskell:7 - 1.5GB => 21MB (minified by 71X)

Elixir application images:

  • (Phoenix service) from elixir:1.6 - 1.1 GB => 37 MB (minified by 29.25X)

RECENT UPDATES

Latest version: 1.41.7 (9/24/2024)

The 1.41.7 version adds the imagebuild command to build container images using different build engines and it also adds a number of improvements with HTTP probing.

For more info about the latest release see the CHANGELOG.

INSTALLATION

If you already have Mint installed use the update command to get the latest version:

mint update

Downloads

  1. Download the zip package for your platform.

    • Latest Mac binaries (curl -L -o ds.zip https://github.com/mintoolkit/mint/releases/download/1.41.7/dist_mac.zip)

    • Latest Mac M1 binaries (curl -L -o ds.zip https://github.com/mintoolkit/mint/releases/download/1.41.7/dist_mac_m1.zip))

    • Latest Linux binaries (curl -L -o ds.tar.gz https://github.com/mintoolkit/mint/releases/download/1.41.7/dist_linux.tar.gz)

    • Latest Linux ARM binaries (curl -L -o ds.tar.gz https://github.com/mintoolkit/mint/releases/download/1.41.7/dist_linux_arm.tar.gz)

    • Latest Linux ARM64 binaries (curl -L -o ds.tar.gz https://github.com/mintoolkit/mint/releases/download/1.41.7/dist_linux_arm64.tar.gz)

  2. Unzip the package and optionally move it to your bin directory.

Linux (for non-intel replace dist_linux with the platform-specific extracted path):

tar -xvf ds.tar.gz
mv  dist_linux/mint /usr/local/bin/
mv  dist_linux/mint-sensor /usr/local/bin/

Mac:

unzip ds.zip
mv  dist_mac/mint /usr/local/bin/
mv  dist_mac/mint-sensor /usr/local/bin/
  1. Add the location where you unzipped the package to your PATH environment variable (optional).

If the directory where you extracted the binaries is not in your PATH then you'll need to run your Mint app binary from that directory.

Scripted Install

You can also use this script to install the current release of Mint on Linux (x86 and ARM) and macOS (x86 and Apple Silicon)

curl -sL https://raw.githubusercontent.com/mintoolkit/mint/master/scripts/install-mint.sh | sudo -E bash -

Homebrew

brew install docker-slim

The Homebrew installer: https://formulae.brew.sh/formula/docker-slim

Docker

docker pull mintoolkit/mint

See the RUNNING CONTAINERIZED section for more usage info.

BASIC USAGE INFO

mint [global flags] [xray|slim|debug|profile|imagebuild|run|lint|merge|images|registry|vulnerability|app|help] [command-specific flags] <IMAGE_ID_OR_NAME>

If you don't specify any command mint will start in the interactive prompt mode.

COMMANDS

  • slim - Create a minimal container image for your selected image generating the supported security profiles. This is the most popular command. (aka build).
  • debug - Debug minimal or regular container images running in Docker, Podman, Kubernetes and ContainerD.
  • xray - Performs static analysis for the target container image (including 'reverse engineering' the Dockerfile for the image). Use this command if you want to know what's inside of your container image and what makes it fat.
  • lint - Analyzes container instructions in Dockerfiles (Docker image support is WIP).
  • registry - Execute registry operations (pull, push, copy, server).
  • profile - Performs basic container image analysis and dynamic container analysis, but it doesn't generate an optimized image.
  • run - Runs one or more containers (for now runs a single container similar to docker run).
  • merge - Merge two container images (optimized to merge minified images).
  • imagebuild - Build container image using selected build engine.
  • images - Get information about container images (example: mint --quiet images).
  • vulnerability - Execute vulnerability related tools and operations (epss).
  • app - Execute app management, maintenance, debugging and query operations (bom, version, remove-sensor-volumes, update, install operations).
  • help - Show the available commands and global flags.

Example: mint slim my/sample-app

See the USAGE DETAILS section for more details. Run mint help to get a high level overview of the available commands. Run mint COMMAND_NAME without any parameters and you'll get more information about that command (e.g., mint slim).

If you run mint without any parameters you'll get an interactive prompt that will provide suggestions about the available commands and flags. Tabs are used to show the available options, to autocomplete the parameters and to navigate the option menu (which you can also do with Up and Down arrows). Spaces are used to move to the next parameter and Enter is used to run the command. For more info about the interactive prompt see go-prompt.

USAGE DETAILS

mint [global options] command [command options] <target image ID or name>

Commands:

  • slim - Create a minimal container image for your selected image generating the supported security profiles. (aka build).
  • debug - Debug minimal or regular container images running in Docker, Podman, Kubernetes and ContainerD.
  • xray - Show what's in the container image and reverse engineer its Dockerfile.
  • lint - Lint the target Dockerfile (or image, in the future)
  • registry - Execute registry operations (pull, push, copy, server).
  • profile - Collect fat image information and generate a fat container report.
  • merge - Merge two container images (optimized to merge minified images).
  • imagebuild - Build container image using selected build engine.
  • images - Get information about container images.
  • vulnerability - Execute vulnerability related tools and operations (epss).
  • app - Execute app management, maintenance, debugging and query operations (bom, version, remove-sensor-volumes, update, install operations).
  • help - Show help info.

Global options:

  • --report - command report location (target location where to save the executed command results; slim.report.json by default; set it to off to disable)
  • --check-version - check if the current version is outdated
  • --version - print the version
  • --debug - enable debug logs
  • --verbose - enable info logs
  • --log-level - set the logging level ('debug', 'info', 'warn' (default), 'error', 'fatal', 'panic')
  • --log-format - set the format used by logs ('text' (default), or 'json')
  • --crt-api-version - Container runtime API version, right now applies only to the Docker runtime API version (default: 1.32)
  • --quiet - less verbose CLI execution mode
  • --output-format - set the output format to use ('text' (default), or 'json')
  • --log - log file to store logs
  • --host - Docker host address or socket (prefix with tcp:// or unix://)
  • --crt-connection - Container runtime connection (for non-Docker runtimes / for Docker user --host)
  • --crt-context - Container runtime context name if supported (for Docker similar to setting '--context' or DOCKER_CONTEXT)
  • --crt-io-inactivity-timeout - CRT I/O general inactivity timeout.
  • --crt-save-inactivity-timeout - CRT save image operation inactivity timeout (overrides the general I/O timeout). If not set, defaults to 30 seconds.
  • --crt-copy-inactivity-timeout - CRT copy from container operation inactivity timeout (overrides the general I/O timeout). If not set, defaults to 30 seconds.
  • --tls - use TLS connecting to Docker
  • --tls-verify - do TLS verification
  • --tls-cert-path - path to TLS cert files
  • --state-path value - Mint state base path (must set it if the Mint binaries are not in a writable directory!)
  • --archive-state - Archives Mint state to the selected Docker volume (default volume - mint-state). By default, enabled when Mint is running in a container (disabled otherwise). Set it to off to disable explicitly.
  • --in-container - Set it to true to explicitly indicate that Mint is running in a container (if it's not set Mint will try to analyze the environment where it's running to determine if it's containerized)

To disable the version checks set the global --check-version flag to false (e.g., --check-version=false) or you can use the DSLIM_CHECK_VERSION environment variable.

LINT COMMAND OPTIONS

  • --target - target Dockerfile path (or Docker image, in the future; if you don't use this flag you must specify the target as the argument to the command)
  • --target-type - explicitly specify the command target type (values: dockerfile, image)
  • --skip-build-context - don't try to analyze build context
  • --build-context-dir - explicitly specify the build context directory
  • --skip-dockerignore - don't try to analyze .dockerignore
  • --include-check-label - include checks with the selected label key:value
  • --exclude-check-label - exclude checks with the selected label key:value
  • --include-check-id - check ID to include
  • --include-check-id-file - file with check IDs to include
  • --exclude-check-id - check ID to exclude
  • --exclude-check-id-file - file with check IDs to exclude
  • --show-nohits - show checks with no matches
  • --show-snippet - show check match snippet (default value: true)
  • --list-checks - list available checks (don't need to specify the target flag if you just want to list the available checks)

XRAY COMMAND OPTIONS

  • --runtime - Runtime environment type (values: docker, podman and a special meta runtime auto, which auto-selects the runtime based on the installed runtime; defaults to auto)
  • --target - Target container image (name or ID)
  • --pull - Try pulling target if it's not available locally (default: false).
  • --docker-config-path - Set the docker config path used to fetch registry credentials (used with the --pull flag).
  • --registry-account - Account to be used when pulling images from private registries (used with the --pull flag).
  • --registry-secret - Account secret to be used when pulling images from private registries (used with the --pull and --registry-account flags).
  • --show-plogs - Show image pull logs (default: false).
  • --changes value - Show layer change details for the selected change type (values: none, all, delete, modify, add).
  • --changes-output value - Where to show the changes (values: all, report, console).
  • --layer value - Show details for the selected layer (using layer index or ID)
  • --add-image-manifest - Add raw image manifest to the command execution report file
  • --add-image-config - Add raw image config object to the command execution report file
  • --layer-changes-max - Maximum number of changes to show for each layer
  • --all-changes-max - Maximum number of changes to show for all layers
  • --add-changes-max - Maximum number of add changes to show for all layers
  • --modify-changes-max - Maximum number of modify changes to show for all layers
  • --delete-changes-max - Maximum number of delete changes to show for all layers
  • --change-path value - Include changes for the files that match the path pattern (Glob/Match in Go and **). Value formats: <path pattern> | dump:<output type>:<path pattern> | ::<path pattern> where output type is console or a directory name. If value starts with dump: the match will be 'dumped' to the selected output type. [can use this flag multiple times]
  • --change-data value - Include changes for the files that match the data pattern (regex). Value formats: <data regex> | dump:<output type>:<path pattern>:<data regex> | ::<path pattern>:<data regex> | :::<data regex> where output type is console or a directory name. If value starts with dump: the match will be 'dumped' to the selected output type. [can use this flag multiple times]
  • --change-data-hash value - Include changes for the files that match the provided data hashes (sha1). Value formats: <sha1 hash> | dump:<output type>:<sha1 hash> | ::<sha1 hash> where output type is console or a directory name. If value starts with dump: the match will be 'dumped' to the selected output type. [can use this flag multiple times]
  • --reuse-saved-image - Reuse saved container image (default: true).
  • --top-changes-max - Maximum number of top changes to track (defalt: 20).
  • --hash-data - Generate file data hashes (default: false).
  • --detect-duplicates - Detect duplicate files based on their hashes (default: true).
  • --show-duplicates - Show all discovered duplicate file paths (default: false).
  • --show-special-perms - Show files with special permissions (setuid,setgid,sticky) (default: true)
  • --detect-utf8 - Detect utf8 files and optionally extract the discovered utf8 file content (possible values: "true" or "dump" or "dump:output_target.tgz" or "dump:output_target.tgz::max_size_bytes" or "dump:output_target.tgz:::max_size_bytes").
  • --detect-all-certs - Detect all certifcate files
  • --detect-all-cert-pks - Detect all certifcate private key files
  • --detect-identities - Detect system identities (users, groups) and their properties (default: true)
  • --change-match-layers-only - Show only layers with change matches (default: false).
  • --export-all-data-artifacts - TAR archive file path to export all text data artifacts (if value is set to . then the archive file path defaults to ./data-artifacts.tar)
  • --remove-file-artifacts - Remove file artifacts when command is done (note: you'll loose the reverse engineered Dockerfile)
  • --command-params-file - JSON file with all command parameters - the JSON file can use a snake case formatted commands example --docker-config-path would be docker_config_path

Change Types:

  • none - Don't show any file system change details in image layers (the top changes from the corresponding layer are still shown)
  • all - Show all file system change details in image layers
  • delete - Show only delete file system change details in image layers
  • modify - Show only modify file system change details in image layers
  • add - Show only 'add' file system change details in image layers

In the interactive CLI prompt mode you must specify the target image using the --target flag while in the traditional CLI mode you can use the --target flag or you can specify the target image as the last value in the command.

SLIM COMMAND OPTIONS

  • --target - Target container image (name or ID). It's an alternative way to provide the target information. The standard way to provide the target information is by putting the target image reference as the last value in the slim (aka build) command CLI call.

  • --pull - Try pulling target if it's not available locally (default: true).

  • --docker-config-path - Set the docker config path used to fetch registry credentials (used with the --pull flag).

  • --registry-account - Account to be used when pulling images from private registries (used with the --pull flag).

  • --registry-secret - Account secret to be used when pulling images from private registries (used with the --pull and --registry-account flags).

  • --show-plogs - Show image pull logs (default: false).

  • --compose-file - Load container info from selected compose file

  • --target-compose-svc - Target service from compose file

  • --target-compose-svc-image - Override the container image name and/or tag when targetting a compose service using the target-compose-svc parameter (format: tag_name or image_name:tag_name)

  • --target-compose-svc-no-ports - Do not publish ports for target service from compose file

  • --dep-exclude-compose-svc-all - Do not start any compose services as target dependencies

  • --dep-include-compose-svc - Include specific compose service as a target dependency (only selected services will be started)

  • --dep-exclude-compose-svc - Exclude specific service from the compose services that will be started as target dependencies

  • --dep-include-compose-svc-deps - Include all dependencies for the selected compose service (excluding the service itself) as target dependencies

  • --dep-include-target-compose-svc-deps - Include all dependencies for the target compose service (excluding the service itself) as target dependencies. This is a shortcut flag to avoid repeating the service name (it's a pretty long flag name though :-))

  • --compose-svc-start-wait - Number of seconds to wait before starting each compose service

  • --compose-net - Attach target to the selected compose network(s) otherwise all networks will be attached

  • --compose-env-nohost - Don't include the env vars from the host to compose

  • --compose-env-file - Load compose env vars from file (host env vars override the values loaded from this file)

  • --compose-workdir - Set custom work directory for compose

  • --compose-project-name - Use custom project name for compose

  • --container-probe-compose-svc - Container test/probe service from compose file

  • --prestart-compose-svc - placeholder for now

  • --poststart-compose-svc - placeholder for now

  • --http-probe - Enables/disables HTTP probing (ENABLED by default; you have to disable the probe if you don't need it by setting the flag to false: --http-probe=false)

  • --http-probe-off - Alternative way to disable HTTP probing

  • --http-probe-client-timeout - Probe network client timeout in seconds (defaults to 30 seconds)

  • --http-probe-client-timeout-crawl - Crawl probe network client timeout in seconds (defaults to probe network client timeout if set or internal default if not, 10 seconds)

  • --http-probe-cmd - User defined HTTP probe(s) as [[[[\"crawl\":]PROTO:]METHOD:]PATH] [can use this flag multiple times]

  • --http-probe-cmd-file - File with user defined HTTP probe commands

  • --http-probe-cmd-upload - User defined HTTP probe(s) to submit form data as [[[[[PROTO:]FORM_FILE_NAME:]FORM_FIELD_NAME:]FILE_OR_GENERATE_TYPE:]PATH] [can use this flag multiple times

  • --http-probe-start-wait - Number of seconds to wait before starting HTTP probing

  • --http-probe-retry-off - Disable retries for probes (alternative way to disable probes)

  • --http-probe-retry-count - Number of retries for each probe (set to -1 to disable probe retries / default value: 5)

  • --http-probe-retry-wait - Number of seconds to wait before retrying HTTP probe (doubles when target is not ready; default value: 8)

  • --http-probe-concurrency - Max number of concurrent requests executing HTTP probes (default value: 5)

  • --http-probe-ports - Explicit list of ports to probe (in the order you want them to be probed; excluded ports are not probed!)

  • --http-probe-full - Do full HTTP probe for all selected ports (if false, finish after first successful scan; default value: true)

  • --http-probe-exit-on-failure - Exit when all HTTP probe commands fail (default value: true)

  • --http-probe-exit-on-failure-count - Exit when selected number of probe call failures happens

  • --http-probe-fail-on-status-5xx - Treat 5xx HTTP status codes as errors during HTTP probing

  • --http-probe-crawl - Enable crawling for the default HTTP probe command (default value: true)

  • --http-crawl-max-depth - Max depth to use for the HTTP probe crawler (default value: 3)

  • --http-crawl-max-page-count - Max number of pages to visit for the HTTP probe crawler (default value: 1000)

  • --http-crawl-concurrency - Number of concurrent workers when crawling an HTTP target (default value: 10)

  • --http-max-concurrent-crawlers - Number of concurrent crawlers in the HTTP probe (default value: 1)

  • --http-probe-apispec - Run HTTP probes for API spec where the value represents the target path where the spec is available (supports Swagger 2.x and OpenAPI 3.x) [can use this flag multiple times]

  • --http-probe-apispec-file - Run HTTP probes for API spec from file (supports Swagger 2.x and OpenAPI 3.x) [can use this flag multiple times]

  • --http-probe-exec - App to execute when running HTTP probes. [can use this flag multiple times]

  • --http-probe-exec-file - Apps to execute when running HTTP probes loaded from file.

  • --publish-port - Map container port to host port analyzing image at runtime to make it easier to integrate external tests (format => port | hostPort:containerPort | hostIP:hostPort:containerPort | hostIP::containerPort )[can use this flag multiple times]

  • --publish-exposed-ports - Map all exposed ports to the same host ports analyzing image at runtime (default value: false)

  • --show-clogs - Show container logs (from the container used to perform dynamic inspection)

  • --show-blogs - Show build logs (when the minified container is built)

  • --copy-meta-artifacts - Copy meta artifacts to the provided location

  • --remove-file-artifacts - Remove file artifacts when command is done (note: you'll loose autogenerated Seccomp and Apparmor profiles unless you copy them with the copy-meta-artifacts flag or if you archive the state)

  • --tag - Use a custom tag for the generated image (instead of the default value: <original_image_name>.slim) [can use this flag multiple times if you need to create additional tags for the optimized image]

  • --entrypoint - Override ENTRYPOINT analyzing image at runtime

  • --cmd - Override CMD analyzing image at runtime

  • --mount - Mount volume analyzing image (the mount parameter format is identical to the -v mount command in Docker) [can use this flag multiple times]

  • --include-path - Include directory (and what's in it) or file from image [can use this flag multiple times] (optionally overwriting the artifact's permissions, user and group information; full format: targetPath:octalPermFlags#uid#gid, mimimal format: targetPath ; see the non-default USER FAQ section for more details)

  • --include-path-file - Load directory or file includes from a file (optionally overwriting the artifact's permissions, user and group information; full format: targetPath:octalPermFlags#uid#gid, mimimal format: targetPath ; see the non-default USER FAQ section for more details)

  • --include-paths-creport-file - Keep files from the referenced creport

  • --include-bin value - Include binary from image (executable or shared object using its absolute path)

  • --include-bin-file - Load shared binary file includes from a file (similar to --include-path-file)

  • --include-dir-bins value - Include binaries in the target directory and include their dependencies, which could be in other locations (executables or shared objects using its absolute path)

  • --include-exe value - Include executable from image (by executable name)

  • --include-exe-file - Load executable file includes from a file (similar to --include-path-file)

  • --include-shell - Include basic shell functionality (default value: false)

  • --include-workdir - Keep files in working directory (default value: false)

  • --include-last-image-layers - Keep all files from the last number of layers in the image

  • --include-healthcheck - Keep the HEALTHCHECK related artifacts (default value: true)

  • --include-cert-all - Keep all discovered cert files (default: true)

  • --include-cert-bundles-only - Keep only cert bundles

  • --include-cert-dirs - Keep known cert directories and all files in them

  • --include-cert-pk-all - Keep all discovered cert private keys

  • --include-cert-pk-dirs - Keep known cert private key directories and all files in them

  • --include-new - Keep new files created by target during dynamic analysis (default value: true)

  • --include-distro-info - Keep OS distro metadata (default value: false)

  • --include-oslibs-net - Keep the common networking OS libraries (default value: true)

  • --include-ssh-client - Keep the common SSH client components and configs

  • --include-zoneinfo - Keep the OS/libc zoneinfo data (default value: false)

  • --include-app-nuxt-dir - Keep the root Nuxt.js app directory (default value: false)

  • --include-app-nuxt-build-dir - Keep the build Nuxt.js app directory (default value: false)

  • --include-app-nuxt-dist-dir - Keep the dist Nuxt.js app directory (default value: false)

  • --include-app-nuxt-static-dir - Keep the static asset directory for Nuxt.js apps (default value: false)

  • --include-app-nuxt-nodemodules-dir - Keep the node modules directory for Nuxt.js apps (default value: false)

  • --include-app-next-dir - Keep the root Next.js app directory (default value: false)

  • --include-app-next-build-dir - Keep the build directory for Next.js app (default value: false)

  • --include-app-next-dist-dir - Keep the static SPA directory for Next.js apps (default value: false)

  • --include-app-next-static-dir - Keep the static public asset directory for Next.js apps (default value: false)

  • --include-app-next-nodemodules-dir - Keep the node modules directory for Next.js apps (default value: false)

  • --include-node-package - Keep node.js package by name [can use this flag multiple times]

  • --preserve-path - Keep path from orignal image in its initial state (changes to the selected container image files when it runs will be discarded). [can use this flag multiple times]

  • --preserve-path-file - File with paths to keep from original image in their original state (changes to the selected container image files when it runs will be discarded).

  • --path-perms - Set path permissions/user/group in optimized image (format: target:octalPermFlags#uid#gid ; see the non-default USER FAQ section for more details)

  • --path-perms-file - File with path permissions to set (format: target:octalPermFlags#uid#gid ; see the non-default USER FAQ section for more details)

  • --exclude-pattern - Exclude path pattern (Glob/Match in Go and **) from image (useful when --include-path keeps a directory and you need to exclude / filter out some of the files in that directory)

  • --exclude-varlock-files - Exclude the files in the var and run lock directory (default value: true)

  • --exclude-mounts - Exclude mounted volumes from image (default value: true)

  • --label - Override or add LABEL analyzing image at runtime [can use this flag multiple times]

  • --volume - Add VOLUME analyzing image at runtime [can use this flag multiple times]

  • --env - Add ENV analyzing target image at runtime [can use this flag multiple times]

  • --env-file - Load multiple environment variables from a file when analyzing target image at runtime.

  • --workdir - Override WORKDIR analyzing image at runtime

  • --network - Override default container network settings analyzing image at runtime

  • --expose - Use additional EXPOSE instructions analyzing image at runtime [can use this flag multiple times]

  • --link - Add link to another container analyzing image at runtime [can use this flag multiple times]

  • --hostname - Override default container hostname analyzing image at runtime

  • --etc-hosts-map - Add a host to IP mapping to /etc/hosts analyzing image at runtime [can use this flag multiple times]

  • --container-dns - Add a dns server analyzing image at runtime [can use this flag multiple times]

  • --container-dns-search - Add a dns search domain for unqualified hostnames analyzing image at runtime [can use this flag multiple times]

  • --image-overrides - Save runtime overrides in generated image (values is all or a comma delimited list of override types: entrypoint, cmd, workdir, env, expose, volume, label). Use this flag if you need to set a runtime value and you want to persist it in the optimized image. If you only want to add, edit or delete an image value in the optimized image use one of the --new-* or --remove-* flags (define below).

  • --continue-after - Select continue mode: enter | signal | probe | exec | timeout-number-in-seconds | container.probe (default value if http probes are disabled: enter). You can also select probe and exec together: 'probe&exec' (make sure to use quotes around the two modes or the & will break the shell command).

  • --dockerfile - The source Dockerfile name to build the fat image before it's optimized.

  • --tag-fat - Custom tag for the fat image built from Dockerfile.

  • --cbo-add-host - Add an extra host-to-IP mapping in /etc/hosts to use when building an image (Container Build Option).

  • --cbo-build-arg - Add a build-time variable (Container Build Option).

  • --cbo-label - Add a label when building from Dockerfiles (Container Build Option).

  • --cbo-target - Target stage to build for multi-stage Dockerfiles (Container Build Option).

  • --cbo-network - Networking mode to use for the RUN instructions at build-time (Container Build Option).

  • --cbo-cache-from - Add an image to the build cache (Container Build Option).

  • --cro-runtime - Runtime to use with the created containers (Container Runtime Option).

  • --cro-host-config-file - File to load the Docker host configuration data (JSON format) to use when running the container. See the HostConfig struct definition from the go-dockerclient package for configuration details. Note that Mint will automatically add SYS_ADMIN to the list of capabilities and run the container in privileged mode, which are required to generate the seccomp profiles. The host config parameters specified using their standalone build or profile command flags overwrite the values in the host config file (volume binds are merged).

  • --cro-sysctl - Set namespaced kernel parameters in the created container (Container Runtime Option).

  • --cro-shm-size - Shared memory size for /dev/shm in the created container (Container Runtime Option).

  • --use-local-mounts - Mount local paths for target container artifact input and output (off, by default)

  • --use-sensor-volume - Sensor volume name to use (set it to your Docker volume name if you manage your own Mint sensor volume).

  • --keep-tmp-artifacts - Keep temporary artifacts when command is done (off, by default).

  • --keep-perms - Keep artifact permissions as-is (default: true)

  • --run-target-as-user - Run target app (in the temporary container) as USER from Dockerfile (true, by default)

  • --new-entrypoint - New ENTRYPOINT instruction for the optimized image

  • --new-cmd - New CMD instruction for the optimized image

  • --new-expose - New EXPOSE instructions for the optimized image

  • --new-workdir - New WORKDIR instruction for the optimized image

  • --new-env - New ENV instructions for the optimized image

  • --new-label - New LABEL instructions for the optimized image

  • --new-volume - New VOLUME instructions for the optimized image

  • --remove-volume - Remove VOLUME instructions for the optimized image

  • --remove-env - Remove ENV instructions for the optimized image

  • --remove-label - Remove LABEL instructions for the optimized image

  • --remove-expose - Remove EXPOSE instructions for the optimized image

  • --exec - A shell script snippet to run via Docker exec

  • --exec-file - A shell script file to run via Docker exec

  • --sensor-ipc-mode - Select sensor IPC mode: proxy | direct (useful for containerized CI/CD environments)

  • --sensor-ipc-endpoint - Override sensor IPC endpoint

  • --rta-onbuild-base-image - Enable runtime analysis for onbuild base images (default: false)

  • --rta-source-ptrace - Enable PTRACE runtime analysis source (default: true)

  • --image-build-engine - Select image build engine: internal | docker | none (internal - build the output image without using Docker [default behavior], docker - build the output image with Docker, none - don't build the output image, allows you to do your own build with the tools you want to use, which you'll be able to do by pointing to the artifact directory where the files.tar and Dockerfile artifacts are located for the output image)

  • --image-build-arch - Select output image build architecture (use the standard container image names for the architectures without the OS part)

  • --obfuscate-metadata - Obfuscate the operating system and application metadata to make it more challenging to identify the image components including OS and application packages. As a side effect using this flag will hide vulnerabilities from vulnerability scanners that rely on the OS and programming language specific package metadata to identify vulnerabilities in container images. This flag can be used to prevent vulnerability scanners from detecting vulnerabilities. This flag also prevents some SBOM generation tools from detecting operating system and application packages.

  • --obfuscate-app-package-names - Select the obfuscation mode for the application package names. Available modes: none | empty | prefix | random (none - Do no app package name obfuscation; empty - Replace the app package names with empty values; prefix - Prefix app package names with a string; random - Replace app package names with random values). This flag changes package names, so the packages can't be recognized by vulnerability scanning and SBOM generation tools. The "empty" mode completely hides the packages from vulnerability scanners and from SBOM tools. This flag can be used with the --obfuscate-metadata flag.

  • --enable-mondel - Enable monitor data event log for sensor monitors to log/stream the events captured by those monitors (default: false)

  • --command-params-file - JSON file with all command parameters - the JSON file can use a snake case formatted commands example --docker-config-path would be docker_config_path

In the interactive CLI prompt mode you must specify the target image using the --target flag while in the traditional CLI mode you can use the --target flag or you can specify the target image as the last value in the command.

The --include-path option is useful if you want to customize your minified image adding extra files and directories. The --include-path-file option allows you to load multiple includes from a newline delimited file. Use this option if you have a lot of includes. The includes from --include-path and --include-path-file are combined together. You can also use the --exclude-pattern flag to control what shouldn't be included.

The --continue-after option is useful if you need to script the Mint app. If you pick the probe option then Mint will continue executing the build command after the HTTP probe is done executing. If you pick the exec options then Mint will continue executing the build command after the container exec shell commands (specified using the --exec-file or --exec flags) are done executing. If you pick the timeout option Mint will allow the target container to run for 60 seconds before it will attempt to collect the artifacts. You can specify a custom timeout value by passing a number of seconds you need instead of the timeout string. If you pick the signal option you'll need to send a USR1 signal to the Mint app process. The signal option is useful when you want to run your own tests against the temporary container Mint creates. Your test automation / CI/CD pipeline will be able to notify the Mint app that it's done running its test by sending the USR1 to it.

You can also combine multiple continue-after modes. For now only combining probe and exec is supported (using either probe&exec or exec&probe as the --continue-after flag value). Other combinations may work too. Combining probe and signal is not supported.

The --include-shell option provides a simple way to keep a basic shell in the minified container. Not all shell commands are included. To get additional shell commands or other command line utilities use the --include-exe and/or --include-bin options. Note that the extra apps and binaries might missed some of the non-binary dependencies (which don't get picked up during static analysis). For those additional dependencies use the --include-path and --include-path-file options.

The --dockerfile option makes it possible to build a new minified image directly from source Dockerfile. Pass the Dockerfile name as the value for this flag and pass the build context directory or URL instead of the docker image name as the last parameter for the build command: mint slim --dockerfile Dockerfile --tag my/custom_minified_image_name . If you want to see the console output from the build stages (when the fat and slim images are built) add the --show-blogs build flag. Note that the build console output is not interactive and it's printed only after the corresponding build step is done. The fat image created during the build process has the .fat suffix in its name. If you specify a custom image tag (with the --tag flag) the .fat suffix is added to the name part of the tag. If you don't provide a custom tag the generated fat image name will have the following format: slim-tmp-fat-image.<pid_of_slim>.<current_timestamp>. The minified image name will have the .slim suffix added to that auto-generated container image name (slim-tmp-fat-image.<pid_of_slim>.<current_timestamp>.slim). Take a look at this python examples to see how it's using the --dockerfile flag.

The --use-local-mounts option is used to choose how the Mint sensor is added to the target container and how the sensor artifacts are delivered back to the master. If you enable this option you'll get the original Mint app behavior where it uses local file system volume mounts to add the sensor executable and to extract the artifacts from the target container. This option doesn't always work as expected in the dockerized environment where Mint itself is running in a Docker container. When this option is disabled (default behavior) then a separate Docker volume is used to mount the sensor and the sensor artifacts are explicitly copied from the target container.

DEBUG COMMAND OPTIONS

Debug minimal or regular container images running in Docker, Podman, Kubernetes and ContainerD.

  • --runtime - Runtime environment type (values: docker, k8s, containerd, podman and a special meta runtime auto, which auto-selects the runtime based on the installed runtime; defaults to auto)
  • --debug-image - Debug image to use for the debug side-car container (default value for this flag is busybox). If you are using the interactive prompt mode you'll get a drop down list of all preselected debug images.
  • --list-debug-images - Show the built-in list of debug images to use for the debug sidecar container (with the --debug-image flag). You can use other images too. Use this flag by itself (no need to specify the --target flag).
  • --target - Target container name or ID (this can also be provided as the last param in the command line invocation of the debug command). Note that the target container must be running. If you are using the interactive prompt mode you'll get a drop down list of the currently running containers in the selected runtime.
  • --namespace - Namespace to target [k8s runtime] (defaults to default)
  • --pod - Pod to target [k8s runtime]
  • --cmd - (Optional) custom CMD to use for the debug side-car container (if you need to run CMD value in a shell use --shell-cmd or '--' instead).
  • --shell-cmd - (Optional) custom CMD to use as a shell command for the debug sidecar container (alternatively pass custom CMD params after '--').
  • --entrypoint - (Optional) custom ENTRYPOINT to use for the debug side-car container.
  • --terminal - Attach interactive terminal to the debug container (default: true). When the interactive terminal is not enabled the debug container output will be printed out to the screen when the debug command exits.
  • --kubeconfig - Kubeconfig file location [k8s runtime]
  • --workdir - Custom WORKDIR to use for the debug sidecar container.
  • --env - Environment variable to add to the debug sidecar container.
  • --load-target-env-vars - Load all (container spec) environment variables from the target container into the debug sidecar container (true by default; set it to false if you don't want this behavior).
  • --mount - Volume to mount in the debug sidecar container (format: name:path or name:path:ro).
  • --mount-target-volumes - Mount all volumes mounted in the target container to the debug sidecar container.
  • --uid - UID to use for the debugging sidecar container.
  • --gid - GID to use for the debugging sidecar container.
  • --run-privileged - Run the debug sidecar as a privileged container (true by default).
  • --security-context-from-target - Use the security context params from the target container with the debug sidecar container.
  • --fallback-to-target-user - Fallback to using target container user if it's non-root (true by default; set it to false to disable this behavior).
  • --run-as-target-shell - Attach an interactive terminal to the debug container and run shell as if it's running in the target container environment (true by default).
  • --list-sessions - List all debug sessions for the selected target (pod and optionally selected container for k8s or container for other runtimes).
  • --show-session-logs - Show logs for the selected debug session (using namespace, pod, target container or debug session container name for k8s or debug session container name for other runtimes).
  • --session - Debug session container name (used for debug session actions).
  • --connect-session - Connect to existing debug session.
  • --list-namespaces - List names for available namespaces (use this flag by itself) [k8s runtime].
  • --list-pods - List names for running pods in the selected namespace (use this flag by itself) [k8s runtime].
  • --list-debuggable-containers - List container names for active containers that can be debugged (use this flag by itself).
  • --help show help (default: false)

See the "Debugging Using the debug Command" section for more information about this command.

IMAGEBUILD COMMAND OPTIONS

Build container image using selected build engine. The created image will be saved as a tar file. You can also load the new image into one of the supported runtimes.

USAGE: mint [GLOBAL FLAGS] imagebuild [FLAGS] [IMAGE]

Flags:

  • --engine - Container image build engine to use: docker (Native Docker container build engine), podman (Native Podman/Buildah container build engine), buildkit (BuildKit container build engine), depot (Depot.dev cloud-based container build engine), simple (built-in simple image build engine)
  • --image-name - Container image name to use (including tag).
  • --image-archive-file - Local file path for the image tar archive file (used for the depot and buildkit engines).
  • --dockerfile - Local Dockerfile path (for buildkit and depot) or a relative to the build context directory (for docker or podman). Default: Dockerfile.
  • --context-dir - Local build context directory. Default: ..
  • --build-arg - Build time variable (ARG). [can use this flag multiple times]
  • --label - Image label to add. [can use this flag multiple times]
  • --architecture - Build architecture (amd64 or arm64).
  • --engine-endpoint - Build engine endpoint address (for buildkit).
  • --engine-token - Build engine specific API token (for depot).
  • --engine-namespace - Build engine specific namespace (for depot).
  • --runtime-load - Container runtime where to load the created image: none, docker, podman.
  • --base - simple build engine: base image to use (from selected runtime, docker by default, or pulled if not available)
  • --base-tar - simple build engine: base image from a local tar file
  • --base-with-certs - simple build engine: boolean flat to use the static-debian12 distroless base image - contains only certs and timezone info
  • --exe-path - simple build engine: local (linux) executable file that will be used as the entrypoint for the new image (added to the selected base image or scratch image if no base image is provided)

Examples:

An example showing how to use the cloud-based depot.dev build engine:

export DEPOT_TOKEN=your_token
export DEPOT_PROJECT_ID=your_project
mint imagebuild --engine depot --dockerfile Dockerfile --context-dir .

Another depot.dev engine example that also loads the created image into the local Docker instance:

mint imagebuild --engine depot --engine-token your_depot_token --engine-namespace your_depot_project --dockerfile Dockerfile --context-dir . --runtime-load docker

A BuildKit engine example that uses a local BuildKit instance running on tcp port 1234:

mint imagebuild --engine buildkit --engine-endpoint tcp://localhost:12345 --image-name imagebuild-buildkit-app:latest --dockerfile ./node_app/Dockerfile --context-dir ./node_app --runtime-load docker

With this BuildKit example the new image name is set to imagebuild-buildkit-app:latest. Make sure to use the --image-name flag if you don't want to use the default container image name.

Note that the --dockerfile flag is a path to the Dockerfile you want to use and it's not related to the context directory path provided with the --context-dir. This works for the depot and buildkit engines, but not for the docker or podman engines where the --dockerfile flag value is relative to the context directory.

A Podman engine example:

mint imagebuild --engine podman --image-name imagebuild-podman-app:latest --build-arg BA=1 --label LONE=value --label LTWO="label two value" --dockerfile Dockerfile --context-dir node_app

Note that this example also building the same image in the node_app directory, but here the --dockerfile flag value is set to Dockerfile. In this case, setting the --dockerfile flag was not necessary, because Dockerfile is the default --dockerfile flag value.

RUN COMMAND OPTIONS

Run one or more containers

USAGE: mint [GLOBAL FLAGS] run [FLAGS] [IMAGE]

Flags:

  • --target - Target container image to run. Same as specifying the target container image as the last value for the command. Used mostly for the interactive prompt mode where you need to select flag names.
  • --pull - Pull the target image before trying to run it.
  • --docker-config-path - Docker config path (used to fetch registry credentials).
  • --registry-account - Target registry account used when pulling images from private registries.
  • --registry-secret - Target registry secret used when pulling images from private registries.
  • --show-plogs - Show image pull logs.
  • --entrypoint - Override ENTRYPOINT running the target image.
  • --cmd - Override CMD running the target image.
  • --live-logs - Show live logs for the container (cant use with --terminal).
  • --terminal - Attach interactive terminal to the container.
  • --publish - Map container port to host port (format => port | hostPort:containerPort | hostIP:hostPort:containerPort | hostIP::containerPort ).
  • --rm - Remove the container when it exits.
  • --detach - Start the container and do not wait for it to exit.

MERGE COMMAND OPTIONS

Merge two container images. Optimized to merge minified images.

Flags:

  • --image - Image to merge. Flag instance position determines the merge order. The command supports two instances of this flag.

  • --use-last-image-metadata - Use only the last image metadata for the merged image.

  • --tag - Custom tags for the output image (multiple instances).

REGISTRY COMMAND OPTIONS

For the operations that require authentication you can reuse the registry credentials from Docker (do docker login first and then use the --use-docker-credentials flag with the registry command) or you can specify the auth info using the --account and --secret flags).

Current sub-commands: pull, push, image-index-create, server.

There's also a placeholder for copy, but it doesn't do anything yet. Great opportunity to contribute ;-)

Shared Command Level Flags:

  • --use-docker-credentials - Use the registry credentials from the default Docker config file.
  • --account - Registry credentials account.
  • --secret - Registry credentials secret.

PULL SUBCOMMAND OPTIONS

USAGE: mint [GLOBAL FLAGS] registry [SHARED FLAGS] pull [FLAGS] [IMAGE]

Flags:

  • --target value - Target container image (name or ID) [$DSLIM_TARGET]
  • --save-to-docker- Save pulled image to docker (default: true) [$DSLIM_REG_PULL_SAVE_TO_DOCKER]

PUSH SUBCOMMAND OPTIONS

USAGE: mint [GLOBAL FLAGS] registry [SHARED FLAGS] push [FLAGS] [IMAGE]

Flags:

  • --docker -- Push local docker image.
  • --tar -- Push image from a local tar file.
  • --as -- Tag the selected image with the specified name before pushing.

Note that mint registry push LOCAL_DOCKER_IMAGE_NAME is a shortcut for mint registry push --docker LOCAL_DOCKER_IMAGE_NAME.

Normally you have to explicitly tag the target image to have a name that's appropriate for the destination registry. The --as flag is a convinient way to tag the image while you are pushing it. Here's an example pushing a local Docker nginx image to a local registry: mint registry push --docker nginx --as localhost:5000/nginx

You can create a local registry using the server subcommand. See the server sub-command section below for more details.

COPY SUBCOMMAND OPTIONS

USAGE: mint registry copy [SRC_IMAGE] [DST_IMAGE]

NOTE: Just a placeholder for now (TBD)

IMAGE-INDEX-CREATE SUBCOMMAND OPTIONS

USAGE: mint registry image-index-create --image-index-name [MULTI-ARCH_IMAGE_TAG] --image-name [IMAGE_ONE] --image-name [IMAGE_TWO]

Flags:

  • --image-index-name - Image index name to use.
  • --image-name - Target image name to include in image index.
  • --as-manifest-list - Create image index with the manifest list media type instead of the default OCI image index type.
  • --dump-raw-manifest - Dump raw manifest for the created image index.
  • --insecure-refs - Allow the referenced images from insecure registry connections.

SERVER SUBCOMMAND OPTIONS

Starts a server which implements the OCI API spec on port 5000 by default.

USAGE: mint [GLOBAL FLAGS] registry server [FLAGS]

Flags:

  • --address - Registry server address to listen on (default: 0.0.0.0)
  • --port - Registry server port (default: 5000)
  • --https - Use HTTPS.
  • --cert-path - Cert path for use with HTTPS (for use when not using autocert).
  • --key-path - Key path for use with HTTPS (for use when not using autocert).
  • --domain - Domain to use for registry server (to get certs). Only works if the registry is internet accessible (see autocert Go docs for more details).
  • --referrers-api - Enables the referrers API endpoint (OCI 1.1+). Enabled by default (set to false to disable).

VULNERABILITY COMMAND OPTIONS

USAGE: mint [GLOBAL FLAGS] vulnerability [SHARED FLAGS] [SUBCOMMAND] [FLAGS]

Current sub-commands:

  • epss - Gets EPPS information for the target vulnerabilities or based on the selected vulnerability filters.

Shared Command Level Flags:

  • --cve - Target vulnerability CVE ID (can specify multiple times to target multiple vulnerabilities).

EPSS SUBCOMMAND OPTIONS

USAGE: mint [GLOBAL FLAGS] vulnerability [SHARED FLAGS] epss [FLAGS]

Flags:

  • --op - EPSS operation (lookup | list).
  • --date - Date for the EPSS information (YYYY-MM-DD format). Works with the lookup and list operations.
  • --with-history - Return EPSS results with historical data. Works with the lookup and list operations.
  • --limit - Limit the number of returned records.
  • offset - Offset where to start returning records.
  • filter-cve-id-pattern - 'CVE ID pattern' ESPP list operation filter.
  • filter-days-since-added - 'days since added' ESPP list operation filter.
  • filter-score-gt - 'score is greater than' ESPP list operation filter.
  • filter-score-lt - 'score is less than' ESPP list operation filter.
  • filter-percentile-gt - 'percentile is greater than' ESPP list operation filter.
  • filter-percentile-lt - 'percentile is less than' ESPP list operation filter.
  • filter-order-records - 'order returned records' ESPP list operation filter ('score-desc' | 'score-asc' | 'percentile-desc' | 'percentile-asc').

Examples:

  • mint --quiet vulnerability --cve CVE-2021-21315 epss
  • mint --output-format=json vulnerability --cve CVE-2021-21315 epss
  • mint --quiet --output-format=json vulnerability --cve CVE-2021-21315 --cve CVE-2023-49070 epss
  • mint --quiet vulnerability --cve CVE-2021-21315 epss --with-history --date 2022-12-13
  • mint --quiet vulnerability epss --op list --date 2024-01-05
  • mint --quiet vulnerability epss --op list --filter-cve-id-pattern 2023 --filter-score-gt 0.92 --limit 2 --offset 3

RUNNING CONTAINERIZED

The current version of Mint is able to run in containers. It will try to detect if it's running in a containerized environment, but you can also tell Mint explicitly using the --in-container global flag.

You can run Mint in your container directly or you can use the Mint container image in your containerized environment. If you are using the Mint container image make sure you run it configured with the Docker IPC information, so it can communicate with the Docker daemon. The most common way to do it is by mounting the Docker unix socket to the Mint app container. Some containerized environments (like Gitlab and their dind service) might not expose the Docker unix socket to you, so you'll need to make sure the environment variables used to communicate with Docker (e.g., DOCKER_HOST) are passed to the Mint app container. Note that if those environment variables reference any kind of local host names those names need to be replaced or you need to tell the Mint app about them using the --etc-hosts-map flag. If those environment variables reference local files those local files (e.g., files for TLS cert validation) will need to be copied to a temporary container, so that temporary container can be used as a data container to make those files accessible by the Mint app container.

When Mint app runs in a container it will attempt to save its execution state in a separate Docker volume. If the volume doesn't exist it will try to create it (mint-state, by default). You can pick a different state volume or disable this behavior completely by using the global --archive-state flag. If you do want to persist the Mint app execution state (which includes the seccomp and AppArmor profiles) without using the state archiving feature you can mount your own volume that maps to the /bin/.mint-state directory in the Mint app container.

By default, the Mint app will try to create a Docker volume for its sensor unless one already exists. If this behavior is not supported by your containerized environment you can create a volume separately and pass its name to the Mint app using the --use-sensor-volume flag.

Here's a basic example of how to use the containerized version of the Mint app: docker run -it --rm -v /var/run/docker.sock:/var/run/docker.sock mintoolkit/mint slim your-docker-image-name

Here's a GitLab example for their dind .gitlab-ci.yml config file: docker run -e DOCKER_HOST=tcp://$(grep docker /etc/hosts | cut -f1):2375 mintoolkit/mint slim your-docker-image-name

Here's a CircleCI example for their remote docker .circleci/config.yml config file (used after the setup_remote_docker step):

docker create -v /dcert_path --name dcert alpine:latest /bin/true
docker cp $DOCKER_CERT_PATH/. dcert:/dcert_path
docker run --volumes-from dcert -e DOCKER_HOST=$DOCKER_HOST -e DOCKER_TLS_VERIFY=$DOCKER_TLS_VERIFY -e DOCKER_CERT_PATH=/dcert_path mintoolkit/mint slim your-docker-image-name

Different CI/CD services have different containerized environment designs that impose various restrictions that may impact the ability of the main app to communicate with the sensor app embedded in the temporary container Mint creates. Try adjusting the values for the --sensor-ipc-mode and --sensor-ipc-endpoint flags. This Google Cloud Build blog post by Márton Kodok is a good reference for both of those flags.

Using *-file Flags

  • There are several flags that accept file paths (--include-path-file, --compose-file, --http-probe-cmd-file, etc). You need volume mount the location of the referenced paths or the file paths themselves when you use the containerized version of Mint because the Mint app container won't have accept to the referenced files otherwise.

CI/CD INTEGRATIONS

Integrating Mint in Jenkins

Prerequisites:

  • Spin up a virtual machine(e.g.EC2 Instance, Azure VM, GCE) which has an Ubuntu OS via your desired cloud platform(AWS, Azure, GCP), SSH into the machine, update the machine packages and install docker. An example of this step is highlighted below given you are running an AWS EC2 Instance.
sudo apt update -y
sudo apt install docker -y
sudo systemctl start docker
sudo usermod -aG docker ec2-user
  • Install Jenkins on the virtual machine using docker as stipulated by the Jenkins Documentation, this step pulls Jenkins Image from DockerHub, runs Jenkins as a container via port 8080 and creates an explicit docker volume on the host machine to retain Jenkins data. Given you are running an AWS EC2 Instance, create a TCP rule with port 8080 in the Instance security group rules which allows only your Internet Protocol(IP) address to access the Jenkins server.
docker run -p 8080:8080 -p 50000:50000 -d -v jenkins_home:/var/jenkins_home jenkins/jenkins:lts
  • Given Jenkins is now running as a containerized environment in the virtual machine, you need to make docker available in the Jenkins container, you can do this by bind mounting the virtual machine docker unix socket onto the jenkins container, note that to carry out this step you need to stop the running jenkins container, you can find the jenkins container ID by using the docker ps command, the commands to execute are highlighted below. This step is essential as it makes docker available in the Jenkins container, and with docker you can pull Mint app image which is to be used in furthur steps.
docker ps 
docker stop [jenkins_container_id]
docker run -p 8080:8080 -p 50000:50000 -d \  
-v jenkins_home:/var/jenkins_home \ 
-v /var/run/docker.sock:/var/run/docker.sock \ 
-v $(which docker):/usr/bin/docker jenkins/jenkins:lts
  • Enable Docker permissions in the new jenkins container, such that Jenkins can perform docker commands and pull the Official Image in the container. To do this, you need to get into the Jenkins container as a root user, you can find the jenkins container ID by using the docker ps command, the commands to execute are highlighted below:
docker exec -u 0 -it [jenkins_container_id] bash
chmod 666 /var/run/docker.sock 
docker pull mintoolkit/mint

Jenkinsfile Slim Stage

Given you have completed the prerequisite steps above, you can build a docker image and minify the image size using Mint via the snippet stage below which should be highlighted in your Jenkinsfile stages.

stage("Build and Slim Docker Image") {
  steps {
      script {
          echo "building and minifying docker image..."
          sh 'docker build -t IMAGE_NAME:$BUILD_NUMBER .'
          sh 'docker run --rm -v /var/run/docker.sock:/var/run/docker.sock mintoolkit/mint \
              build --target IMAGE_NAME:$BUILD_NUMBER --tag IMAGE_NAME:slim-$BUILD_NUMBER \
              exit'
      }
  }
}
  • The snippet stage above allows for customization, you should replace the image name--IMAGE_NAME with your desired image name, the environment variable tag--$BUILD_NUMBER represents a unique incremental number allocated by Jenkins each time your jenkins pipeline runs.
  • The docker build command builds a Docker Image of your application from a Dockerfile.
  • The docker run command runs Mint in a non-interactive mode via the docker unix socket, minifies the built(target) image--IMAGE_NAME:$BUILD_NUMBER, and adjusting it to a new slimmed image with the image/tag--IMAGE_NAME:slim-$BUILD_NUMBER.
  • You should put the Mint stage before a docker tag/push stage and after a build/test artifact in your Jenkinsfile, an example pipeline is highlighted below for a sample nodejs application; The first stage test and builds an artifact of the application; The second stage builds a docker image and a slimmed version of the docker image; The third stage tags the slimmed docker image with a DockerHub account remote repository and pushes the image to the remote repository.
pipeline {
    agent any
    stages {
        stage("building nodejs app") {
            steps{
                script {
                    echo "building nodejs app..."
                    sh 'npm run test'
                    sh 'npm pack'
                }
            }
        }
        stage("Build and Slim Docker Image") {
            steps {
                script {
                    echo "building and slimming docker image..."
                    sh 'docker build -t node_alpine:$BUILD_NUMBER .'
                    sh 'docker run --rm -v /var/run/docker.sock:/var/run/docker.sock mintoolkit/mint \
                        build --target node_alpine:$BUILD_NUMBER --tag node_alpine:slim-$BUILD_NUMBER \
                        exit'
                }
            }
        }
        stage("Push Slim Image to Regristy") {
            steps {
                script {
                    echo 'pushing image to docker regristry...'
                    withCredentials([usernamePassword(credentialsId: 'docker-hub-credentials', passwordVariable: 'PASS', usernameVariable: 'USER')]) {
                        sh 'docker tag node_alpine:slim-$BUILD_NUMBER $USER/node_alpine:slim-$BUILD_NUMBER'
                        sh 'echo $PASS | docker login -u $USER --password-stdin'
                        sh 'docker push $USER/node_alpine:slim-$BUILD_NUMBER'
                    }
                }
            }
        }
    }
}   

Integrating Mint in Github Actions

Docker-Slim Github Action

Integrating Slim in Github Actions in your CI/CD workflow involves using the Docker-Slim Github Action, this Action(snippet below) minifies a target docker image--IMAGE_NAME:latest in your workflow, making it smaller and adjusting the new slimmed image as IMAGE_NAME:slim.

# Slim it!
- uses: kitabisa/docker-slim-action@v1
  env:
    DSLIM_HTTP_PROBE: false
  with:
    target: IMAGE_NAME:latest
    tag: "slim"

Github Actions Workflow

You can integrate the Docker-Slim Github Action in your workflow by inserting the Action after a Docker Build/Push Github Action, before Docker Login Github Action and docker tag/push commands, a customized example workflow is highlighted below. Note that the environment variable tag--{{github.run_number}} in the workflow represents a unique incremental number allocated by Github Actions each time your workflow runs.

# Build the Docker image first
- uses: docker/build-push-action@v4
  with:
    push: false
    tags: IMAGE_NAME:{{github.run_number}} 

# Slim the Image
- uses: kitabisa/docker-slim-action@v1
  env:
    DSLIM_HTTP_PROBE: false
  with:
    target: IMAGE_NAME:{{github.run_number}}
    tag: "slim-{{github.run_number}}"

# Docker Hub Login
  uses: docker/login-action@v2
  with:
    username: ${{ secrets.DOCKERHUB_USERNAME }}
    password: ${{ secrets.DOCKERHUB_TOKEN }}

# Push to the registry
- run: | 
   docker tag IMAGE_NAME:slim-{{github.run_number}} ${{ secrets.DOCKERHUB_USERNAME }}/IMAGE_NAME:slim-{{github.run_number}}
   docker push ${{ secrets.DOCKERHUB_USERNAME }}/IMAGE_NAME:slim-{{github.run_number}}

The workflow above indicates four steps:

  • A Docker Build/Push Github Action for building a docker image with the image name/tag--IMAGE_NAME:{{github.run_number}}, you should give replace IMAGE_NAME with your desired image name. Note that this Action must have a false option to push the built image--given that you need the image slimmed/minified before pushing it to a container registry.
  • A Docker-Slim Github Action which minifies the target image--IMAGE_NAME:{{github.run_number}}, this Action has the "slim-{{github.run_number}}" tag and adds this tag to the slimmed/minified docker image such that the image name/tag becomes IMAGE_NAME:slim-{{github.run_number}}.
  • A Docker Login Github Action which logs into your DockerHub container regristry account, you should store your DockerHub username and personal access token as secrets in the github repository meant for the workflow. Suppose your container registry is not DockerHub, you can check the Docker Login Github Action documentation for the use case of logging into your desired container registry.
  • A docker tag command for naming/tagging the slimmed image with your DockerHub account remote repository name which could be the same name(IMAGE_NAME) as the slimmed image; A docker push command to push the slimmed image to your Dockerhub account remote repository.

DOCKER CONNECT OPTIONS

If you don't specify any Docker connect options the Mint app expects to find the Docker Unix socket (/var/run/docker.sock) or the following environment variables: DOCKER_HOST, DOCKER_TLS_VERIFY (optional), DOCKER_CERT_PATH (required if DOCKER_TLS_VERIFY is set to "1"). Note that the DOCKER_HOST environment variable can be used to point to a Unix socket address (in case the default Unix socket isn't there). This is useful when you use Docker Desktop and you haven't configured Docker Desktop to create the default Unix socket.

If the Docker environment variables are configured to use TLS and to verify the Docker cert (default behavior), but you want to disable the TLS verification you can override the TLS verification behavior by setting the --tls-verify to false:

mint --tls-verify=false build my/sample-node-app-multi

You can override all Docker connection options using these flags: --host, --tls, --tls-verify, --tls-cert-path, --crt-context. These flags correspond to the standard Docker options (and the environment variables). Note that you can also use the --host flag (similar to DOCKER_HOST) to point to a Unix socket (e.g., --host=unix:///var/run/docker.sock).

The --crt-context flag is currently supported with the Docker runtime and it's similar to using the --context flag or DOCKER_CONTEXT. Note that Mint will use DOCKER_CONTEXT if it's configured.

If you want to use TLS with verification:

mint --host=tcp://192.168.99.100:2376 --tls-cert-path=/Users/youruser/.docker/machine/machines/default --tls=true --tls-verify=true build my/sample-node-app-multi

If you want to use TLS without verification:

mint --host=tcp://192.168.99.100:2376 --tls-cert-path=/Users/youruser/.docker/machine/machines/default --tls=true --tls-verify=false build my/sample-node-app-multi

If the Docker environment variables are not set and if you don't specify any Docker connect options Mint will try to use the default unix socket.

DOCKER DESKTOP

You may not have the default Unix socket (/var/run/docker.sock) configured if you use Docker Desktop. By default, Docker Desktop uses ~/.docker/run/docker.sock as the Unix socket.

You can either use --host or DOCKER_HOST to point to the Docker Desktop's Unix socket or you can configure Docker Desktop to create the default/traditional Unix socket (creating the /var/run/docker.sock symlink manually is an option too).

To configure Docker Desktop to create the default Unix socket open its UI and go to Settings -> Advanced where you need to check the Enable default Docker socket (Requires password) option.

HTTP PROBE COMMANDS

If the HTTP probe is enabled (note: it is enabled by default) it will default to running GET / with HTTP and then HTTPS on every exposed port. You can add additional commands using the --http-probe-cmd, --http-probe-cmd-upload and --http-probe-cmd-file options.

If you want to disable HTTP probing set the --http-probe flag to false (e.g., --http-probe=false). You can also use the --http-probe-off flag to do the same (simply use the flag without any parameters).

The --http-probe-cmd option is good when you want to specify a small number of simple commands where you select some or all of these HTTP command options: crawling (defaults to false), protocol, method (defaults to GET), resource (path and query string).

The --http-probe-cmd-upload option is good when you want to upload a file. The file upload is done using a multipart form data request (the most common way to upload files). The flag value has the following format: [[[[[PROTO:]FORM_FILE_NAME:]FORM_FIELD_NAME:]FILE_OR_GENERATE_TYPE:]PATH]. The most basic flag value is a PATH (e.g., --http-probe-cmd-upload /uploadpath). When no file to upload or data generate instruction is provided a simple text file is generated. The FILE_OR_GENERATE_TYPE section can either point to a local file path for the file to upload or it can be one of the data generate instructions: generate.text, generate.text.json, generate.image, generate.image.png, generate.image.jpeg, generate.image.gif. You can specify a custom form field name to use for the request using the FORM_FIELD_NAME section (defaults to file). You can also specify the file name for the upload using the FORM_FILE_NAME section (defaults to file.data or the base path of the file name when FILE_OR_GENERATE_TYPE points to a local file). The PROTO section is used to scope the request to a specific protocol (http, https, etc). Note that you can also specify an upload command using a JSON record provided with the --http-probe-cmd-file flag (see below).

If you only want to use custom HTTP probe command and you don't want the default GET / command added to the command list you explicitly provided you'll need to set --http-probe to false when you specify your custom HTTP probe command. Note that this inconsistency will be addressed in the future releases to make it less confusing.

Possible field combinations:

  • /path - runs GET /path
  • crawl:/path - runs GET /path and then crawls the pages referenced by the target page
  • post:/path - runs POST /path
  • crawl:get:/path - runs GET /path and then crawls the pages referenced by the target page
  • https:get:/path runs GET /path only on https
  • crawl:http:get:/path - runs GET /path and then crawls the pages referenced by the target page

Here are a couple of examples:

Adds two extra probe commands: GET /api/info and POST /submit (tries http first, then tries https): mint slim --show-clogs --http-probe-cmd /api/info --http-probe-cmd POST:/submit my/sample-node-app-multi

Adds one extra probe command: POST /submit (using only http): mint slim --show-clogs --http-probe-cmd http:POST:/submit my/sample-node-app-multi

The --http-probe-cmd-file option is good when you have a lot of commands and/or you want to select additional HTTP command options.

Available HTTP command options:

  • method - HTTP method to use
  • resource - target resource URL
  • port - port number
  • protocol - http, https, http2, http2c (cleartext version of http2), ws, wss (secure websocket)
  • headers - array of strings with column delimited key/value pairs (e.g., "Content-Type: application/json")
  • body - request body as a string
  • body_file - request body loaded from the provided file
  • body_generate - auto-generate request body using one of the instructions: generate.text, generate.text.json, generate.image, generate.image.png, generate.image.jpeg, generate.image.gif
  • body_is_form - request body is a multipart form (request data from body, body_file or body_generate will be form encoded)
  • form_field_name - form field name to use (for form submissions)
  • form_file_name - form file name to use (for form submissions)
  • username - username to use for basic auth
  • password - password to use for basic auth
  • crawl - boolean to indicate if you want to crawl the target (to visit all referenced resources)
  • fastcgi - Fast CGI config params (todo: document sub-fields, see code for details)

Here's a probe command file example:

mint slim --show-clogs --http-probe-cmd-file probeCmds.json my/sample-node-app-multi

Commands in probeCmds.json:

{
  "commands":
  [
   {
     "resource": "/api/info"
   },
   {
     "method": "POST",
     "resource": "/submit"
   },
   {
     "procotol": "http",
     "resource": "/api/call?arg=one"
   },
   {
     "protocol": "http",
     "method": "POST",
     "resource": "/submit2",
     "body": "key=value"
   },
   {
     "protocol": "http",
     "method": "POST",
     "resource": "/submit3",
     "body_file": "mydata.json",
     "headers": ["Content-Type: application/json"]
   }
  ]
}

The HTTP probe command file path can be a relative path (relative to the current working directory) or it can be an absolute path.

For each HTTP probe call Mint will print the call status. Example: info=http.probe.call status=200 method=GET target=http://127.0.0.1:32899/ attempt=1 error=none.

You can execute your own external HTTP requests using the target.port.list field in the container info message Mint prints when it starts its test container: mint[build]: info=container name=<your_container_name> id=<your_container_id> target.port.list=[<comma_separated_list_of_port_numbers_to_use>] target.port.info=[<comma_separated_list_of_port_mapping_records>]. Example: mint[build]: info=container name=mintk_42861_20190203084955 id=aa44c43bcf4dd0dae78e2a8b3ac011e7beb6f098a65b09c8bce4a91dc2ff8427 target.port.list=[32899] target.port.info=[9000/tcp => 0.0.0.0:32899]. With this information you can run curl or other HTTP request generating tools: curl http://localhost:32899.

The current version also includes an experimental crawling capability. To enable it for the default HTTP probe use the --http-probe-crawl flag. You can also enable it for the HTTP probe commands in your command file using the crawl boolean field.

When crawling is enabled the HTTP probe will act like a web crawler following the links it finds in the target endpoint.

Probing based on the Swagger/OpenAPI spec is another experimental capability. This feature introduces two new flags:

  • http-probe-apispec - value: <path_to_fetch_spec>:<api_endpoint_prefix>
  • http-probe-apispec-file - value: <local_file_path_to_spec>

You can use the --http-probe-exec and --http-probe-exec-file options to run the user provided commands when the http probes are executed. This example shows how you can run curl against the temporary container created by Mint when the http probes are executed.

mint slim --http-probe-exec 'curl http://localhost:YOUR_CONTAINER_PORT_NUM/some/path' --publish-port YOUR_CONTAINER_PORT_NUM your-container-image-name

DEBUGGING MINIFIED CONTAINERS

Debugging Using the debug Command

Most users should run the debug command in the interactive mode, which can be done by executing the main app (mint) without any parameters. The interactive mode shows you the commands and their flags. It also helps you interactively select many important debug command flag values (e.g., target container, namespace, pod).

The current version of the debug command supports multiple container runtimes: docker, podman, kubernetes and containerd.

Use the --runtime flag to explicitly select your target container runtime. By default, the flag value is set to auto and the command will try to auto-detect the installed runtime(s). If the auto-select isn't picking up your installation set the flag explicitly.

By default the debug command will provide you with an interactive terminal when it attaches the debuging sidecar image to the target container. The interactive terminal will allow you navigate the target container filesystem as if you were connected to the target container. To disable this behavior set --run-as-target-shell to false.

Other default behavior:

  • The debugging sidecar container runs as a privileged container by default. To disable the behavior set --run-privileged to false.
  • The environment variables from the target container (defined at the container image/spec level) will be loaded into the debugging sidecar container. For now, this is only implemented for k8s. To disable the behavior set --load-target-env-vars to false.

Podman Runtime

Depending on how the Podman container runtime is installed and if you are using Podman Desktop you may need to start Podman services to expose the API the debug command is using. Right now it's your responsibility, but the future versions will auto-start the Podman service.

Here's what you'll need to do:

  1. Start the service (it can be done with a unix socket too): podman system service --time=0 tcp://localhost:8899
  2. Provide the same connection string to the debug command using the global --crt-connection flag: mint --crt-connection tcp://localhost:8899 debug --runtime podman --target YOUR_CONTAINER_NAME_OR_ID

ContainerD Runtime

You'll be able to use the ContainerD runtime only on Linux. At this point in time it's not possible to run the debug command on Macs with Lima, Finch or similar Mac options.

The Debug Images

You can use any container image as a debug image, but there's a list of pre-selected debug images you can choose.

You can list all pre-selected debug images with the --list-debug-images and if you are using the interactive prompt mode there'll be an auto-complete dropdown menu for the --debug-image flag.

Here's the current list of debug images:

  • cgr.dev/chainguard/min-toolkit-debug:latest - a general purpose Mint debug image created by Chainguard
  • cgr.dev/chainguard/wolfi-base:latest - a basic lightweight Wolfi image
  • busybox:latest - a lightweight image with common unix utilities
  • nicolaka/netshoot - a network trouble-shooting swiss-army container
  • lightruncom/koolkits:node - a debug image for Node.js applications
  • lightruncom/koolkits:python - a debug image for Python applications
  • lightruncom/koolkits:golang - a debug image for Go applications
  • lightruncom/koolkits:jvm - a debug image for Java applications
  • digitalocean/doks-debug:latest - a kubernetes troubleshooting debug image
  • public.ecr.aws/zinclabs/debug-ubuntu-base:latest - an image with common debugging utilities

Steps to Debug Your Container (Kubernetes Runtime)

  1. Make sure the target environment you want to debug is up (the example k8s manifest creates a pod with the minimal nginx image from Chainguard and it has no shell):
>> kubectl apply -f examples/k8s_nginx_cgr/manifest.yaml
  1. Run the debug command:
>> mint debug --runtime=k8s --pod=example-pod example-container

or

>> mint debug --runtime=k8s --pod=example-pod --target=example-container

For a very simple setup where you have only one namespace, only one pod and only one container or if you want to target the first pod and the first container in that pod your debug command can be simplified to look like this:

>> mint debug --runtime k8s

Now you should have an interactive shell into the debug container started by mint and you can type your regular shell commands.

By default the debug command will connect the interactive terminal to the debugged container and it will run a shell as if it's running in the target container environment, so you will see the file system of the target container as if you are directly connected (you won't have to go through the proc file system). You can change this behavior by using the --run-as-target-shell (which is true by default). For example, this call will connect you to the debug container in a more traditional way: mint debug --runtime=k8s --run-as-target-shell=false example-container

Also note that if you use the interactive prompt mode (when you run mint with no command line parameters) you will get auto-complete behavior for a number of flags: --target, --namespace, --pod, --session.

Each time you try to debug an image mint will have a session that represents it. You'll be able to reconnect to the existing active debug sessions and you'll be able to get logs from all available sessions.

Steps to Debug Your Container (Docker Runtime)

  1. Start the target container you want to debug:
>> docker run -it --rm -p 80:80 --name mycontainer nginx
  1. Run the debug command:
>> mint debug mycontainer

or

>> mint debug --target=mycontainer

Now you should have an interactive shell into the debug container started by mint and you can type your regular shell commands.

By default the debug command will connect the interactive terminal to the debugged container and it will run a shell as if it's running in the target container environment, so you will see the file system of the target container as if you are directly connected (you won't have to go through the proc file system). You can change this behavior by using the --run-as-target-shell (which is true by default). For example, this call will connect you to the debug container in a more traditional way: mint debug --run-as-target-shell=false mycontainer

Also note that if you use the interactive prompt mode (when you run mint with no command line parameters) you will get auto-complete behavior for a number of flags: --target, --session.

Each time you try to debug an image mint will have a session that represents it. You'll be able to reconnect to the existing active debug sessions and you'll be able to get logs from all available sessions.

Debugging the "Hard Way" (Docker Runtime)

You can create dedicated debugging side-car container images loaded with the tools you need for debugging target containers. This allows you to keep your production container images small. The debugging side-car containers attach to the running target containers.

Assuming you have a running container named node_app_alpine you can attach your debugging side-car with a command like this: docker run --rm -it --pid=container:node_app_alpine --net=container:node_app_alpine --cap-add sys_admin alpine sh. In this example, the debugging side-car is a regular alpine image. This is exactly what happens with the node_alpine app sample (located in the node_alpine directory of the examples repo) and the run_debug_sidecar.command helper script.

If you run the ps command in the side-car you'll see the application from the target container:

# ps
PID   USER     TIME   COMMAND
    1 root       0:00 node /opt/my/service/server.js
   13 root       0:00 sh
   38 root       0:00 ps

You can access the target container file system through /proc/<TARGET_PID>/root:

# ls -lh /proc/1/root/opt/my/service
total 8
drwxr-xr-x    3 root     root        4.0K Sep  2 15:51 node_modules
-rwxr-xr-x    1 root     root         415 Sep  8 00:52 server.js

Some of the useful debugging commands include cat /proc/<TARGET_PID>/cmdline, ls -l /proc/<TARGET_PID>/cwd, cat /proc/1/environ, cat /proc/<TARGET_PID>/limits, cat /proc/<TARGET_PID>/status and ls -l /proc/<TARGET_PID>/fd.

Talks and Presentations about Debugging

MINIFYING COMMAND LINE TOOLS

Unless the default CMD instruction in your Dockerfile is sufficient you'll have to specify command line parameters when you execute the build command in Mint. This can be done with the --cmd option.

Other useful command line parameters:

  • --show-clogs - use it if you want to see the output of your container.
  • --mount - use it to mount a volume when Mint inspects your image.
  • --entrypoint - use it if you want to override the ENTRYPOINT instruction when Mint inspects your image.

Note that the --entrypoint and --cmd options don't override the ENTRYPOINT and CMD instructions in the final minified image.

Here's a sample slim (aka build) command:

mint slim --show-clogs=true --cmd docker-compose.yml --mount $(pwd)/data/:/data/ mintoolkit/container-transform

It's used to minify the container-transform tool. You can get the minified image from Docker Hub.

QUICK SECCOMP EXAMPLE

If you want to auto-generate a Seccomp profile AND minify your image use the slim (aka build) command. If you only want to auto-generate a Seccomp profile (along with other interesting image metadata) use the profile command.

Step one: run Mint

mint slim your-name/your-app

Step two: use the generated Seccomp profile

docker run --security-opt seccomp:<mint directory>/.images/<YOUR_APP_IMAGE_ID>/artifacts/your-name-your-app-seccomp.json <your other run params> your-name/your-app

Feel free to copy the generated profile :-)

You can use the generated Seccomp profile with your original image or with the minified image.

USING AUTO-GENERATED SECCOMP PROFILES

You can use the generated profile with your original image or with the minified image Mint created:

docker run -it --rm --security-opt seccomp:path_to/my-sample-node-app-seccomp.json -p 8000:8000 my/sample-node-app.slim

ORIGINAL DEMO VIDEO

DockerSlim demo

Demo video on YouTube

DEMO STEPS

The demo runs on Mac OS X, but you can build a linux version. Note that these steps are different from the steps in the demo video.

  1. Get the Mint app binaries:

Unzip them and optionally add their directory to your PATH environment variable if you want to use the app from other locations.

The extracted directory contains two binaries (and now it also contains a symlink for the old n ame):

  • mint <- the main Mint application binary
  • mint-sensor <- the sensor application used to collect information from running containers
  • docker-slim <- the symlink to mint, the new main app binary (useful if you are still using the old name in your scripts)
  • slim <- the symlink to mint, the new main app binary (useful if you are still using the older name in your scripts)
  1. Clone the examples repo to use the sample apps (note: the examples have been moved to a separate repo). You can skip this step if you have your own app.

git clone https://github.com/mintoolkit/examples.git

  1. Create a Docker image for the sample node.js app in examples/node_ubuntu. You can skip this step if you have your own app.

cd examples/node_ubuntu

docker build -t my/sample-node-app .

  1. Run the Mint app:

./min build my/sample-node-app <- run it from the location where you extraced the Mint app binaries (or update your PATH env var to include the directory where the Mint app binaries are located)

Mint creates a special container based on the target image you provided. It also creates a resource directory where it stores the information it discovers about your image: <min directory>/.images/<TARGET_IMAGE_ID>.

By default, the Mint app will run its http probe against the temporary container. If you are minifying a command line tool that doesn't expose any web service interface you'll need to explicitly disable http probing (by setting --http-probe=false).

  1. Use curl (or other tools) to call the sample app (optional)

curl http://<YOUR_DOCKER_HOST_IP>:<PORT>

This is an optional step to make sure the target app container is doing something. Depending on the application it's an optional step. For some applications it's required if it loads new application resources dynamically based on the requests it's processing (e.g., Ruby or Python).

You'll see the mapped ports printed to the console when the Mint app starts the target container. You can also get the port number either from the docker ps or docker port <CONTAINER_ID> commands. You can also map the exposed ports to specific host ports you want to use.

  1. Press and wait until the Mint app says it's done

By default or when http probing is enabled explicitly the Mint app will continue its execution once the http probe is done running. If you explicitly picked a different continue-after option follow the expected steps. For example, for the enter continue-after option you must press the enter button on your keyboard.

If http probing is enabled (when http-probe is set) and if continue-after is set to enter and you press the enter key before the built-in HTTP probe is done the probe might produce an EOF error because the Mint app will shut down the target container before all probe commands are done executing. It's ok to ignore it unless you really need the probe to finish.

  1. Once Mint is done check that the new minified image is there

docker images

You should see my/sample-node-app.slim in the list of images. Right now all generated images have .slim at the end of its name.

  1. Use the minified image

docker run -it --rm --name="slim_node_app" -p 8000:8000 my/sample-node-app.slim

FAQ

Is it safe for production use?

Yes! Either way, you should test your Docker images.

How can I contribute if I don't know Go?

You don't need to read the language spec and lots of books :-) Go through the Tour of Go and optionally read 50 Shades of Go and you'll be ready to contribute!

What's the best application for Mint?

Mint will work for any containerized application; however, Mint automates app interactions for applications with an HTTP API. You can use Mint even if your app doesn't have an HTTP API. You'll need to interact with your application manually to make sure Mint can observe your application behavior.

Can I use Mint with dockerized command line tools?

Yes. The --cmd, --entrypoint, and --mount options will help you minify your image. The container-transform tool is a good example.

Notes:

You can explore the artifacts Mint generates when it's creating a slim image. You'll find those in <mint directory>/.images/<TARGET_IMAGE_ID>/artifacts. One of the artifacts is a "reverse engineered" Dockerfile for the original image. It'll be called Dockerfile.reversed.

If you don't want to create a minified image and only want to "reverse engineer" the Dockerfile you can use the info command.

What if my Docker images uses the USER command?

The current version of Mint does include support for non-default users (take a look at the non-default user examples (including the ElasticSearch example located in the 3rdparty directory) in the examples repo. Please open tickets if something doesn't work for you.

Everything should work as-is, but for the special cases where the current behavior don't work as expected you can adjust what Mint does using various build command parameters: --run-target-as-user, --keep-perms, --path-perms, --path-perms-file (along with the --include-* parameters).

The --run-target-as-user parameter is enabled by default and it controls if the application in the temporary container is started using the identity from the USER instruction in the container's Dockerfile.

The --keep-perms parameter is also enabled by default. It tells Mint to retain the permissions and the ownership information for the files and directories copied to the optimized container image.

The --path-perms and --path-perms-file parameters are similar to the --include-path and --include-path-file parameters. They are used to overwrite the permission and the user/group information for the target files and directories. Note that the target files/directories are expected to be in the optimized container image. If you don't know if the target files/directories will be in the optimized container you'll need to use one of the --include-* parameters (e.g., --include-path-file) to explicitly require those artifacts to be included. You can specify the permissions and the ownership information in the --include-* parameters too (so you don't need to have the --path-* parameters just to set the permissions).

The --path-* and --include-* params use the same format to communicate the permission/owernship info: TARGET_PATH_OR_NAME:PERMS_IN_OCTAL_FORMAT#USER_ID#GROUP_ID.

You don't have to specify the user and group IDs if you don't want to change them.

Here's an example using these parameters to minify the standard nginx image adding extra artifacts and changing their permissions: mint slim --include-path='/opt:770#104#107' --include-path='/bin/uname:710' --path-perms='/tmp:700' nginx.

This is what you'll see in the optimized container image:

drwx------  0 0      0           0 Feb 28 22:15 tmp/
-rwx--x---  0 0      0       31240 Mar 14  2015 bin/uname
drwxrwx---  0 104    107         0 Feb 28 22:13 opt/

The uname binary isn't used by nginx, so the --include-path parameter is used to keep it in the optimized image changing its permissions to 710.

The /tmp directory will be included in the optimized image on its own, so the --path-perms parameter is used to change its permissions to 700.

When you set permissions/user/group on a directory the settings are only applied to that directory and not to the artifacts inside. The future versions will allow you to apply the same settings to everything inside the target directory too.

Also note that for now you have to use numeric user and group IDs. The future versions will allow you to use user and group names too.

Nginx fails in my minified image

If you see nginx: [emerg] mkdir() "/var/lib/nginx/body" failed it means your nginx setup uses a non-standard temporary directory. Nginx will fail if the base directory for its temporary folders doesn't exist (they won't create the missing intermediate directories). Normally it's /var/lib/nginx, but if you have a custom config that points to something else you'll need to add an --include-path flag as an extra flag when you run the Mint app.

Mint fails with a 'no permission to read from' error

This problem shouldn't happen anymore because the exported artifacts are saved in a tar file and the master app doesn't need to access the files directly anymore.

If you run older versions of Mint you can get around this problem by running Mint from a root shell. That way it will have access to all exported files.

Mint copies the relevant image artifacts trying to preserve their permissions. If the permissions are too restrictive the master app might not have sufficient priviledge to access these files when it's building the new minified image.

EXPLORE SOURCE CODE WITH AI

You can explore the project code and ask questions about the code using a Perplexity-like Code Sage portal created by Storia AI: https://sage.storia.ai/mintoolkit

BUILD PROCESS

Build Options

Pick one of the build options that works best for you.

Containerized

Run make build_in_docker on linux or make build_m1_in_docker on Macs (or ./scripts/docker-builder.run.sh or click on ./scripts/mac/docker-builder.run.command on Macs) from the project directory (builds Mint in a Docker container; great if you don't want to install Go on your local machine and if you already have Docker).

Native

Run make build on linux or make build_m1 on Macs (or ./scripts/src.build.sh or click on ./scripts/mac/src.build.command on Macs) to build Mint natively (requires Go installed locally).

Note:

Try using the latest version of Go building the Mint app. The current version of Go used to build the Mint app is 1.22.

Gitpod

If you have a web browser, you can get a fully pre-configured development environment in one click:

Open in Gitpod

Additional Tools
  • license-bill-of-materials - Optional tool to track dependencies and their licenses.
  • golint - Optional tool for code analysis. See https://github.com/golang/lint for more details.

You can install these tools using the tools.get.sh shell script in the scripts directory.

Notes:

  • Make sure you have golint if you intend to run the src.inspect.sh or mac.src.inspect.command scripts.

CONTRIBUTING

If the project sounds interesting or if you found a bug see CONTRIBUTING.md and submit a PR or open an issue! Non-code contributions including docs are highly appreciated! Open an issue even if you have a question or something is not clear.

CORE CONCEPTS

  1. Inspect container metadata (static analysis)
  2. Inspect container data (static analysis)
  3. Inspect running application (dynamic analysis)
  4. Build an application artifact graph
  5. Use the collected application data to build small images
  6. Use the collected application data to auto-generate various security framework configurations.

DYNAMIC ANALYSIS OPTIONS

  1. Instrument the container image (and replace the entrypoint/cmd) to collect application activity data
  2. Use kernel-level tools that provide visibility into running containers (without instrumenting the containers)
  3. Disable relevant namespaces in the target container to gain container visibility (can be done with runC)

SECURITY

The goal is to auto-generate Seccomp, AppArmor, (and potentially SELinux) profiles based on the collected information.

  • AppArmor profiles
  • Seccomp profiles

CHALLENGES

Some of the advanced analysis options require a number of Linux kernel features that are not always included. The kernel you get with Docker Machine / Boot2docker is a great example of that.

ORIGINS

DockerSlim was a Docker Global Hack Day #dockerhackday project. It barely worked at the time, but it did get a win in Seattle and it took the second place in the Plumbing category overall :-)

DHD3

Since then it's been improved and it works pretty well for its core use cases. It can be better though. That's why the project needs your help! You don't need to know much about the container internals, container runtimes and you don't need to know anything about Go. You can contribute in many different ways. For example, use Mint on your images and open Github issues documenting your experience even if it worked just fine :-)

LICENSE

Apache License v2, see LICENSE for details.

CODE OF CONDUCT

The project follows the CNCF Code of Conduct.


Go Report Card