aboutsummaryrefslogtreecommitdiff
path: root/src/cpt-lib.in
diff options
context:
space:
mode:
Diffstat (limited to 'src/cpt-lib.in')
-rw-r--r--src/cpt-lib.in2013
1 files changed, 2013 insertions, 0 deletions
diff --git a/src/cpt-lib.in b/src/cpt-lib.in
new file mode 100644
index 0000000..7728ae8
--- /dev/null
+++ b/src/cpt-lib.in
@@ -0,0 +1,2013 @@
+#!/bin/sh -ef
+# @DOCSTRING@
+# shellcheck source=/dev/null
+#
+# This is the Carbs Packaging Toolchain written for Carbs Linux.
+# It was originally forked from the kiss package manager by
+# Dylan Araps.
+#
+# Currently maintained by Cem Keylan.
+
+version() {
+ log "Carbs Packaging Tools" @VERSION@
+ exit 0
+}
+
+out() {
+ # Print a message as is.
+ printf '%s\n' "$@"
+}
+
+log() {
+ # Print a message prettily.
+ #
+ # All messages are printed to stderr to allow the user to hide build
+ # output which is the only thing printed to stdout.
+ #
+ # '${3:-->}': If the 3rd argument is missing, set prefix to '->'.
+ # '${2:+colorb}': If the 2nd argument exists, set text style of '$1'.
+ printf '%b%s %b%b%s%b %s\n' \
+ "$colory" "${3:-->}" "$colre" "${2:+$colorb}" "$1" "$colre" "$2" >&2
+}
+
+die() {
+ # Print a message and exit with '1' (error).
+ log "$1" "$2" "!>"
+ exit 1
+}
+
+trap_set() {
+ # Function to set the trap value.
+ case ${1:-cleanup} in
+ cleanup) trap pkg_clean EXIT INT ;;
+ block) trap '' INT ;;
+ unset) trap - EXIT INT ;;
+ esac
+}
+
+sepchar() (
+ # Seperate every character on the given string without resorting to external
+ # processes.
+ [ "$1" ] || return 0; str=$1; set --
+ while [ "$str" ]; do
+ str_tmp=$str
+ for i in $(_seq $(( ${#str} - 1 ))); do
+ str_tmp=${str_tmp%?}
+ done
+ set -- "$@" "$str_tmp"
+ str=${str#$str_tmp}
+ done
+ printf '%s\n' "$@"
+)
+
+_seq() (
+ # Pure shell counter meant to be used in 'for' loops.
+ i=0 buf=''
+ while [ "$(( i += 1 ))" -le "$1" ]; do
+ buf="$buf $i "
+ done
+ printf '%s' "$buf"
+)
+
+_stat() (
+ _user=; eval set -- "$(ls -ld "$1")"
+ id -u "${_user:=$3}" >/dev/null 2>&1 || _user=root
+ printf '%s' "$_user"
+)
+
+_readlinkf() (
+ # Public domain POSIX sh readlink function by Koichi Nakashima
+ [ "${1:-}" ] || return 1
+ max_symlinks=40
+ CDPATH='' # to avoid changing to an unexpected directory
+
+ target=$1
+ [ -e "${target%/}" ] || target=${1%"${1##*[!/]}"} # trim trailing slashes
+ [ -d "${target:-/}" ] && target="$target/"
+
+ cd -P . 2>/dev/null || return 1
+ while [ "$max_symlinks" -ge 0 ] && max_symlinks=$((max_symlinks - 1)); do
+ if [ ! "$target" = "${target%/*}" ]; then
+ case $target in
+ /*) cd -P "${target%/*}/" 2>/dev/null || break ;;
+ *) cd -P "./${target%/*}" 2>/dev/null || break ;;
+ esac
+ target=${target##*/}
+ fi
+
+ if [ ! -L "$target" ]; then
+ target="${PWD%/}${target:+/}${target}"
+ printf '%s\n' "${target:-/}"
+ return 0
+ fi
+
+ # `ls -dl` format: "%s %u %s %s %u %s %s -> %s\n",
+ # <file mode>, <number of links>, <owner name>, <group name>,
+ # <size>, <date and time>, <pathname of link>, <contents of link>
+ # https://pubs.opengroup.org/onlinepubs/9699919799/utilities/ls.html
+ link=$(ls -dl -- "$target" 2>/dev/null) || break
+ target=${link#*" $target -> "}
+ done
+ return 1
+)
+
+# This is the public domain getoptions shell library. It also forms a usage
+# function.
+# URL: https://github.com/ko1nksm/getoptions (v2.0.1)
+# License: Creative Commons Zero v1.0 Universal
+# shellcheck disable=2016
+getoptions() {
+ _error='' _on=1 _off='' _export='' _plus='' _mode='' _alt='' _rest=''
+ _opts='' _help='' _indent='' _init=@empty IFS=' '
+
+ for i in 0 1 2 3 4 5; do
+ eval "_$i() { echo \"$_indent\$@\"; }"
+ _indent="$_indent "
+ done
+
+ quote() {
+ q="$2'" r=''
+ while [ "$q" ]; do r="$r${q%%\'*}'\''" && q=${q#*\'}; done
+ q="'${r%????}'" && q=${q#\'\'} && q=${q%\'\'}
+ eval "$1=\${q:-\"''\"}"
+ }
+ code() {
+ [ "${1#:}" = "$1" ] && c=3 || c=4
+ eval "[ ! \${$c:+x} ] || $2 \"\$$c\""
+ }
+
+ invoke() { eval '"_$@"'; }
+ prehook() { invoke "$@"; }
+ for i in setup flag param option disp msg; do
+ eval "$i() { prehook $i \"\$@\"; }"
+ done
+
+ args() {
+ on=$_on off=$_off export=$_export init=$_init _hasarg=$1
+ while [ $# -gt 2 ] && [ "$3" != '--' ] && shift; do
+ case $2 in
+ -?) [ "$_hasarg" ] || _opts="$_opts${2#-}" ;;
+ +*) _plus=1 ;;
+ [!-+]*) eval "${2%%:*}=\${2#*:}"
+ esac
+ done
+ }
+ defvar() {
+ case $init in
+ @none) : ;;
+ @export) code "$1" _0 "export $1" ;;
+ @empty) code "$1" _0 "${export:+export }$1=''" ;;
+ @unset) code "$1" _0 "unset $1 ||:" "unset OPTARG ||:; ${1#:}" ;;
+ *)
+ case $init in @*) eval "init=\"=\${${init#@}}\""; esac
+ case $init in [!=]*) _0 "$init"; return 0; esac
+ quote init "${init#=}"
+ code "$1" _0 "${export:+export }$1=$init" "OPTARG=$init; ${1#:}"
+ esac
+ }
+ _setup() {
+ [ $# -gt 0 ] && { [ "$1" ] && _rest=$1; shift; }
+ for i; do [ "$i" = '--' ] && break; eval "_${i%%:*}=\${i#*:}"; done
+ }
+ _flag() { args : "$@"; defvar "$@"; }
+ _param() { args '' "$@"; defvar "$@"; }
+ _option() { args '' "$@"; defvar "$@"; }
+ _disp() { args : "$@"; }
+ _msg() { args : _ "$@"; }
+
+ "$@"
+ _0 "${_rest:?}=''"
+
+ args() {
+ sw='' validate='' pattern='' counter='' on=$_on off=$_off export=$_export
+ while [ $# -gt 1 ] && [ "$2" != '--' ] && shift; do
+ case $1 in
+ --\{no-\}*) sw="$sw${sw:+ | }--${1#--?no-?} | --no-${1#--?no-?}" ;;
+ [-+]? | --*) sw="$sw${sw:+ | }$1" ;;
+ *) eval "${1%%:*}=\"\${1#*:}\""
+ esac
+ done
+ }
+ setup() { :; }
+ _flag() {
+ args "$@"
+ quote on "$on" && quote off "$off"
+ [ "$counter" ] && on=1 off=-1 v="\$((\${$1:-0}+\${OPTARG:-0}))" || v=''
+ _3 "$sw)"
+ _4 '[ "${OPTARG:-}" ] && OPTARG=${OPTARG#*\=} && set -- noarg "$1" && break'
+ _4 "eval '[ \${OPTARG+x} ] &&:' && OPTARG=$on || OPTARG=$off"
+ valid "$1" "${v:-\$OPTARG}"
+ _4 ';;'
+ }
+ _param() {
+ args "$@"
+ _3 "$sw)"
+ _4 '[ $# -le 1 ] && set -- required "$1" && break'
+ _4 'OPTARG=$2'
+ valid "$1" '$OPTARG'
+ _4 'shift ;;'
+ }
+ _option() {
+ args "$@"
+ quote on "$on" && quote off "$off"
+ _3 "$sw)"
+ _4 'set -- "$1" "$@"'
+ _4 '[ ${OPTARG+x} ] && {'
+ _5 'case $1 in --no-*) set -- noarg "${1%%\=*}"; break; esac'
+ _5 '[ "${OPTARG:-}" ] && { shift; OPTARG=$2; } ||' "OPTARG=$on"
+ _4 "} || OPTARG=$off"
+ valid "$1" '$OPTARG'
+ _4 'shift ;;'
+ }
+ valid() {
+ set -- "$validate" "$pattern" "$1" "$2"
+ [ "$1" ] && _4 "$1 || { set -- ${1%% *}:\$? \"\$1\" $1; break; }"
+ [ "$2" ] && {
+ quote pattern "$2"
+ _4 "case \$OPTARG in $2) ;;"
+ _5 "*) set -- pattern:$pattern \"\$1\"; break"
+ _4 "esac"
+ }
+ code "$3" _4 "${export:+export }$3=\"$4\"" "${3#:}"
+ }
+ _disp() {
+ args "$@"
+ _3 "$sw)"
+ code "$1" _4 "echo \"\${$1}\"" "${1#:}"
+ _4 'exit 0 ;;'
+ }
+ _msg() { :; }
+
+ _0 "$2() {"
+ _1 'OPTIND=$(($#+1))'
+ _1 'while OPTARG= && [ $# -gt 0 ]; do'
+ [ "$_alt" ] && _2 'case $1 in -[!-]?*) set -- "-$@"; esac'
+ _2 'case $1 in'
+ wa() { _4 "eval '${1% *}' \${1+'\"\$@\"'}"; }
+ _3 '--?*=*) OPTARG=$1; shift'
+ wa 'set -- "${OPTARG%%\=*}" "${OPTARG#*\=}" "$@"'
+ _4 ';;'
+ _3 '--no-*) unset OPTARG ;;'
+ [ "$_alt" ] || {
+ [ "$_opts" ] && {
+ _3 "-[$_opts]?*) OPTARG=\$1; shift"
+ wa 'set -- "${OPTARG%"${OPTARG#??}"}" "${OPTARG#??}" "$@"'
+ _4 ';;'
+ }
+ _3 '-[!-]?*) OPTARG=$1; shift'
+ wa 'set -- "${OPTARG%"${OPTARG#??}"}" "-${OPTARG#??}" "$@"'
+ _4 'OPTARG= ;;'
+ }
+ [ "$_plus" ] && {
+ _3 '+??*) OPTARG=$1; shift'
+ wa 'set -- "${OPTARG%"${OPTARG#??}"}" "+${OPTARG#??}" "$@"'
+ _4 'unset OPTARG ;;'
+ _3 '+*) unset OPTARG ;;'
+ }
+ _2 'esac'
+ _2 'case $1 in'
+ "$@"
+ rest() {
+ _3 "$1"
+ _4 'while [ $# -gt 0 ]; do'
+ _5 "$_rest=\"\${$_rest}" '\"\${$((${OPTIND:-0}-$#))}\""'
+ _5 'shift'
+ _4 'done'
+ _4 'break ;;'
+ }
+ rest '--) shift'
+ _3 "[-${_plus:++}]?*)" 'set -- unknown "$1" && break ;;'
+ case $_mode in
+ +) rest '*)' ;;
+ *) _3 "*) $_rest=\"\${$_rest}" '\"\${$((${OPTIND:-0}-$#))}\""'
+ esac
+ _2 'esac'
+ _2 'shift'
+ _1 'done'
+ _1 '[ $# -eq 0 ] && { OPTIND=1; unset OPTARG; return 0; }'
+ _1 'case $1 in'
+ _2 'unknown) set -- "Unrecognized option: $2" "$@" ;;'
+ _2 'noarg) set -- "Does not allow an argument: $2" "$@" ;;'
+ _2 'required) set -- "Requires an argument: $2" "$@" ;;'
+ _2 'pattern:*) set -- "Does not match the pattern (${1#*:}): $2" "$@" ;;'
+ _2 '*) set -- "Validation error ($1): $2" "$@"'
+ _1 'esac'
+ [ "$_error" ] && _1 "$_error" '"$@" >&2 || exit $?'
+ _1 'echo "$1" >&2'
+ _1 'exit 1'
+ _0 '}'
+
+ # This throws an error on older versions of shellcheck.
+ # shellcheck disable=2086
+ [ ! "$_help" ] || eval "shift 2; getoptions_help $1 $_help" ${3+'"$@"'}
+}
+# URL: https://github.com/ko1nksm/getoptions (v2.0.1)
+# License: Creative Commons Zero v1.0 Universal
+getoptions_help() {
+ width=30 plus='' leading=' '
+
+ pad() { p=$2; while [ ${#p} -lt "$3" ]; do p="$p "; done; eval "$1=\$p"; }
+
+ args() {
+ _type=$1 var=${2%% *} sw='' label='' hidden='' _width=$width && shift 2
+ while [ $# -gt 0 ] && i=$1 && shift && [ ! "$i" = '--' ]; do
+ case $i in
+ --*) pad sw "$sw${sw:+, }" $((${plus:+4}+4)); sw="$sw$i" ;;
+ -?) sw="$sw${sw:+, }$i" ;;
+ +?) [ ! "$plus" ] || { pad sw "$sw${sw:+, }" 4; sw="$sw$i"; } ;;
+ *) eval "${i%%:*}=\${i#*:}"
+ esac
+ done
+ [ "$hidden" ] && return 0
+
+ [ "$label" ] || case $_type in
+ setup | msg) label='' _width=0 ;;
+ flag | disp) label="$sw " ;;
+ param) label="$sw $var " ;;
+ option) label="${sw}[=$var] "
+ esac
+ pad label "${label:+$leading}$label" "$_width"
+ [ ${#label} -le "$_width" ] && [ $# -gt 0 ] && label="$label$1" && shift
+ echo "$label"
+ pad label '' "$_width"
+ for i; do echo "$label$i"; done
+ }
+
+ for i in 'setup :' flag param option disp 'msg :'; do
+ eval "${i% *}() { args $i \"\$@\"; }"
+ done
+
+ echo "$2() {"
+ echo "cat<<'GETOPTIONSHERE'"
+ "$@"
+ echo "GETOPTIONSHERE"
+ echo "}"
+}
+
+global_options() {
+ msg -- '' 'Global Options:'
+ flag CPT_FORCE -f --force init:@export -- "Force operation"
+ flag CPT_PROMPT -y --no-prompt on:0 off:0 init:@export -- "Do not prompt for confirmation"
+ param CPT_ROOT --root init:@export -- "Use an alternate root directory"
+ disp :usage -h --help -- "Show this help message"
+ disp :version -v --version -- "Print version information"
+}
+
+warn() {
+ # Print a warning message
+ log "$1" "$2" "${3:-WARNING}"
+}
+
+contains() {
+ # Check if a "string list" contains a word.
+ case " $1 " in *" $2 "*) return 0; esac; return 1
+}
+
+regesc() {
+ # Escape special regular expression characters as
+ # defined in POSIX BRE. '$.*[\^'
+ printf '%s\n' "$1" |
+ sed 's|\\|\\\\|g;s|\[|\\[|g;s|\$|\\$|g;s|\.|\\.|g;s|\*|\\*|g;s|\^|\\^|g'
+}
+
+
+prompt() {
+ # If a CPT_NOPROMPT variable is set, continue.
+ # This can be useful for installation scripts and
+ # bootstrapping.
+ [ "$CPT_PROMPT" = 0 ] && return 0
+
+ # Ask the user for some input.
+ [ "$1" ] && log "$1"
+ log "Continue?: Press Enter to continue or Ctrl+C to abort here"
+
+ # POSIX 'read' has none of the "nice" options like '-n', '-p'
+ # etc etc. This is the most basic usage of 'read'.
+ # '_' is used as 'dash' errors when no variable is given to 'read'.
+ read -r _ || return 1
+}
+
+as_root() {
+ # Simple function to run a command as root using either 'sudo',
+ # 'doas' or 'su'. Hurrah for choice.
+ [ "$uid" = 0 ] || log "Using '${su:-su}' (to become ${user:=root})"
+
+ # We are exporting package manager variables, so that we still have the
+ # same repository paths / access to the same cache directories etc.
+ set -- HOME="$HOME" \
+ USER="$user" \
+ XDG_CACHE_HOME="$XDG_CACHE_HOME" \
+ CPT_CACHE="$CPT_CACHE" \
+ CPT_CHOICE="$CPT_CHOICE" \
+ CPT_COMPRESS="$CPT_COMPRESS" \
+ CPT_DEBUG="$CPT_DEBUG" \
+ CPT_FETCH="$CPT_FETCH" \
+ CPT_FORCE="$CPT_FORCE" \
+ CPT_HOOK="$CPT_HOOK" \
+ CPT_KEEPLOG="$CPT_KEEPLOG" \
+ CPT_PATH="$CPT_PATH" \
+ CPT_PID="$CPT_PID" \
+ CPT_PROMPT="$CPT_PROMPT" \
+ CPT_ROOT="$CPT_ROOT" \
+ CPT_TMPDIR="$CPT_TMPDIR" \
+ "$@"
+
+ case ${su##*/} in
+ sls|sudo|doas) "$su" -u "$user" -- env "$@" ;;
+ su) su -c "env $* <&3" "$user" 3<&0 </dev/tty ;;
+ *) die "Invalid CPT_SU value: $su" ;;
+ esac
+}
+
+pop() {
+ # Remove an item from a "string list". This allows us
+ # to remove a 'sed' call and reuse this code throughout.
+ del=$1
+ shift 2
+
+ for i do [ "$i" = "$del" ] || printf %s " $i "; done
+}
+
+run_hook() {
+ # If a fourth parameter 'root' is specified, source
+ # the hook from a predefined location to avoid privilige
+ # escalation through user scripts.
+ [ "$4" ] && CPT_HOOK=$CPT_ROOT/etc/cpt-hook
+
+ [ -f "$CPT_HOOK" ] || return 0
+
+ [ "$2" ] && log "$2" "Running $1 hook"
+
+ TYPE=${1:-null} PKG=${2:-null} DEST=${3:-null} . "$CPT_HOOK"
+}
+
+decompress() {
+ case $1 in
+ *.tar) cat ;;
+ *.bz2) bzip2 -cd ;;
+ *.xz|*.txz) xz -dcT 0 ;;
+ *.tgz|*.gz) gzip -cd ;;
+ *.zst) zstd -cd ;;
+ esac < "$1"
+}
+
+sh256() {
+ # This is a sha256sum function for outputting a standard
+ # hash digest. sha256 on BSD systems require an '-r' flag
+ # for outputting the same way with sha256sum, and still,
+ # it outputs a single space between the hash and the file
+ # whereas sha256sum outputs double spaces. It fallbacks to
+ # openssl, but that is rarely ever needed.
+ { sha256sum "$1" 2>/dev/null ||
+ sha256 -r "$1" 2>/dev/null ||
+ openssl dgst -r -sha256 "$1" ||
+ die "No sha256 program could be run." ;} |
+
+ while read -r hash _; do printf '%s %s\n' "$hash" "$1"; done
+}
+
+pkg_owner() {
+ set +f
+
+ [ "$3" ] || set -- "$1" "$2" "$sys_db"/*/manifest
+
+ pkg_owner=$(grep "$@")
+ pkg_owner=${pkg_owner%/*}
+ pkg_owner=${pkg_owner##*/}
+
+ set -f -- "$pkg_owner"; unset pkg_owner
+ [ "$1" ] && printf '%s\n' "$1"
+}
+
+pkg_isbuilt() (
+ # Check if a package is built or not.
+ read -r ver rel < "$(pkg_find "$1")/version"
+
+ set +f
+ for tarball in "$bin_dir/$1#$ver-$rel.tar."*; do
+ [ -f "$tarball" ] && return 0
+ done
+ return 1
+)
+
+pkg_lint() {
+ # Check that each mandatory file in the package entry exists.
+ log "$1" "Checking repository files"
+
+ repo_dir=$(pkg_find "$1")
+
+ cd "$repo_dir" || die "'$repo_dir' not accessible"
+ [ -f sources ] || warn "$1" "Sources file not found"
+ [ -x build ] || die "$1" "Build file not found or not executable"
+ [ -s version ] || die "$1" "Version file not found or empty"
+
+ read -r _ release 2>/dev/null < version || die "Version file not found"
+ [ "$release" ] || die "Release field not found in version file"
+
+ [ "$2" ] || [ -f checksums ] || die "$pkg" "Checksums are missing"
+}
+
+pkg_find() {
+ # Use a SEARCH_PATH variable so that we can get the sys_db into
+ # the same variable as CPT_PATH. This makes it easier when we are
+ # searching for executables instead of CPT_PATH.
+ : "${SEARCH_PATH:=$CPT_PATH:$sys_db}"
+
+ # Figure out which repository a package belongs to by
+ # searching for directories matching the package name
+ # in $CPT_PATH/*.
+ query=$1 match=$2 type=$3 IFS=:; set --
+
+ # Word splitting is intentional here.
+ # shellcheck disable=2086
+ for path in $SEARCH_PATH ; do
+ set +f
+
+ for path2 in "$path/"$query; do
+ test "${type:--d}" "$path2" && set -f -- "$@" "$path2"
+ done
+ done
+
+ IFS=$old_ifs
+
+ # A package may also not be found due to a repository not being
+ # readable by the current user. Either way, we need to die here.
+ [ "$1" ] || die "Package '$query' not in any repository"
+
+ # Show all search results if called from 'cpt search', else
+ # print only the first match.
+ [ "$match" ] && printf '%s\n' "$@" || printf '%s\n' "$1"
+}
+
+pkg_list() {
+ # List installed packages. As the format is files and
+ # directories, this just involves a simple for loop and
+ # file read.
+
+ # Change directories to the database. This allows us to
+ # avoid having to 'basename' each path. If this fails,
+ # set '$1' to mimic a failed glob which indicates that
+ # nothing is installed.
+ cd "$sys_db" 2>/dev/null || set -- "$sys_db/"\*
+
+ # Optional arguments can be passed to check for specific
+ # packages. If no arguments are passed, list all. As we
+ # loop over '$@', if there aren't any arguments we can
+ # just set the directory contents to the argument list.
+ [ "$1" ] || { set +f; set -f -- *; }
+
+ # If the 'glob' above failed, exit early as there are no
+ # packages installed.
+ [ "$1" = "$sys_db/"\* ] && return 1
+
+ # Loop over each package and print its name and version.
+ for pkg do
+ [ -d "$pkg" ] || { log "$pkg" "not installed"; return 1; }
+
+ read -r version 2>/dev/null < "$pkg/version" || version=null
+ printf '%s\n' "$pkg $version"
+ done
+}
+
+pkg_cache() {
+ read -r version release 2>/dev/null < "$(pkg_find "$1")/version"
+
+ # Initially assume that the package tarball is built with the CPT_COMPRESS
+ # value.
+ if [ -f "$bin_dir/$1#$version-$release.tar.$CPT_COMPRESS" ]; then
+ tar_file="$bin_dir/$1#$version-$release.tar.$CPT_COMPRESS"
+ else
+ set +f; set -f -- "$bin_dir/$1#$version-$release.tar."*
+ tar_file=$1
+ fi
+
+ [ -f "$tar_file" ]
+}
+
+pkg_sources() {
+ # Download any remote package sources. The existence of local
+ # files is also checked.
+ repo_dir=$(pkg_find "$1")
+
+ # Support packages without sources. Simply do nothing.
+ [ -f "$repo_dir/sources" ] || return 0
+
+ log "$1" "Downloading sources"
+
+ # Store each downloaded source in a directory named after the
+ # package it belongs to. This avoid conflicts between two packages
+ # having a source of the same name.
+ mkdir -p "$src_dir/$1" && cd "$src_dir/$1"
+
+ repo_dir=$(pkg_find "$1")
+
+ while read -r src dest || [ "$src" ]; do
+ # Remote git/hg repository or comment.
+ if [ -z "${src##\#*}" ] ||
+ [ -z "${src##git+*}" ] ||
+ [ -z "${src##hg+*}" ]
+
+ then :
+
+ # Remote source (cached).
+ elif [ -f "${src##*/}" ]; then
+ log "$1" "Found cached source '${src##*/}'"
+
+ # Remote source.
+ elif [ -z "${src##*://*}" ]; then
+ log "$1" "Downloading $src"
+
+ curl "$src" -fLo "${src##*/}" || {
+ rm -f "${src##*/}"
+ die "$1" "Failed to download $src"
+ }
+
+ # Local source.
+ elif [ -f "$repo_dir/$src" ]; then
+ log "$1" "Found local file '$src'"
+
+ else
+ die "$1" "No local file '$src'"
+ fi
+ done < "$repo_dir/sources"
+}
+
+pkg_extract() {
+ # Extract all source archives to the build directory and copy over
+ # any local repository files.
+ repo_dir=$(pkg_find "$1")
+
+ # Support packages without sources. Simply do nothing.
+ [ -f "$repo_dir/sources" ] || return 0
+
+ log "$1" "Extracting sources"
+
+ while read -r src dest || [ "$src" ]; do
+ mkdir -p "$mak_dir/$1/$dest" && cd "$mak_dir/$1/$dest"
+
+ case $src in
+ # Git repository.
+ git+*)
+ # Split the source into URL + OBJECT (branch or commit).
+ url=${src##git+} com=${url##*[@#]} com=${com#${url%[@#]*}}
+
+ log "$1" "Cloning ${url%[@#]*}"; {
+ git init
+ git remote add origin "${url%[@#]*}"
+ case "$url" in
+ # Tags are specified via '@'
+ *@*) git fetch -t --depth=1 origin "$com" || git fetch ;;
+ *) git fetch --depth=1 origin "$com" || git fetch
+ esac
+ git checkout "${com:-FETCH_HEAD}"
+ } || die "$1" "Failed to clone $src"
+ ;;
+
+ # Mercurial repository.
+ hg+*)
+ # Split the source into URL + OBJECT (branch or commit).
+ url=${src##hg+} com=${url##*[@#]} com=${com#${url%[@#]*}}
+
+ # Unfortunately, there is no shallow cloning with Mercurial.
+ log "$1" "Cloning ${url%[@#]*}"
+ hg clone -u "${com:-tip}"
+
+ ;;
+
+ # Comment or blank line.
+ \#*|'') continue ;;
+
+ # Only 'tar', 'cpio', and 'zip' archives are currently supported for
+ # extraction. Other filetypes are simply copied to '$mak_dir'
+ # which allows for manual extraction.
+ *://*.tar|*://*.tar.??|*://*.tar.???|*://*.tar.????|*://*.tgz|*://*.txz)
+
+ decompress "$src_dir/$1/${src##*/}" > .ktar
+
+ "$tar" xf .ktar || die "$1" "Couldn't extract ${src##*/}"
+
+ # We now list the contents of the tarball so we can do our
+ # version of 'strip-components'.
+ "$tar" tf .ktar |
+ while read -r file; do printf '%s\n' "${file%%/*}"; done |
+
+ # Do not repeat files.
+ uniq |
+
+ # For every directory in the base we move each file
+ # inside it to the upper directory.
+ while read -r dir ; do
+
+ # Skip if we are not dealing with a directory here.
+ # This way we don't remove files on the upper directory
+ # if a tar archive doesn't need directory stripping.
+ [ -d "${dir#.}" ] || continue
+
+ # Change into the directory in a subshell so we don't
+ # need to cd back to the upper directory.
+ (
+ cd "$dir"
+
+ # We use find because we want to move hidden files
+ # as well.
+ #
+ # Skip the file if it has the same name as the directory.
+ # We will deal with it later.
+ #
+ # Word splitting is intentional here.
+ # shellcheck disable=2046
+ find . \( ! -name . -prune \) ! -name "$dir" \
+ -exec mv -f {} .. \;
+
+ # If a file/directory with the same name as the directory
+ # exists, append a '.cptbak' to it and move it to the
+ # upper directory.
+ ! [ -e "$dir" ] || mv "$dir" "../${dir}.cptbak"
+ )
+ rmdir "$dir"
+
+ # If a backup file exists, move it into the original location.
+ ! [ -e "${dir}.cptbak" ] || mv "${dir}.cptbak" "$dir"
+ done
+
+ # Clean up the temporary tarball.
+ rm -f .ktar
+ ;;
+
+ *://*.cpio|*://*.cpio.??|*://*.cpio.???|*://*.cpio.????)
+ decompress "$src_dir/$1/${src##*/}" | cpio -i
+
+ ;;
+
+ *://*.zip)
+ unzip "$src_dir/$1/${src##*/}" ||
+ die "$1" "Couldn't extract ${src##*/}"
+
+ ;;
+
+ *)
+ # Local file.
+ if [ -f "$repo_dir/$src" ]; then
+ cp -f "$repo_dir/$src" .
+
+ # Remote file.
+ elif [ -f "$src_dir/$1/${src##*/}" ]; then
+ cp -f "$src_dir/$1/${src##*/}" .
+
+ else
+ die "$1" "Local file $src not found"
+ fi
+ ;;
+ esac
+ done < "$repo_dir/sources"
+}
+
+pkg_depends() {
+ # Resolve all dependencies and generate an ordered list.
+ # This does a depth-first search. The deepest dependencies are
+ # listed first and then the parents in reverse order.
+ contains "$deps" "$1" || {
+ # Filter out non-explicit, aleady installed dependencies.
+ # Only filter installed if called from 'pkg_build()'.
+ [ "$pkg_build" ] && [ -z "$2" ] &&
+ (pkg_list "$1" >/dev/null) && return
+
+ while read -r dep type || [ "$dep" ]; do
+ # Skip comments and empty lines.
+ [ "${dep##\#*}" ] || continue
+ # Skip test dependencies unless $CPT_TEST is set to 1.
+ #
+ # Skip make dependencies on the 'tree' operation for child packages
+ # or when the 'first-nomake' argument is given.
+ case $type in
+ test) [ "$CPT_TEST" = 1 ] || continue ;;
+ make) [ "$2" = tree ] && [ -z "${3#first-nomake}" ] && continue
+ esac
+
+ # Recurse through the dependencies of the child packages. Forward
+ # the 'tree' operation.
+ if [ "$2" = tree ]; then
+ pkg_depends "$dep" tree
+ else
+ pkg_depends "$dep"
+ fi
+ done 2>/dev/null < "$(pkg_find "$1")/depends" ||:
+
+ # After child dependencies are added to the list,
+ # add the package which depends on them.
+ [ "$2" = explicit ] || [ "$3" ] || deps="$deps $1 "
+ }
+}
+
+pkg_order() {
+ # Order a list of packages based on dependence and
+ # take into account pre-built tarballs if this is
+ # to be called from 'cpt i'.
+ order=; redro=; deps=
+
+ for pkg do case $pkg in
+ *.tar.*) deps="$deps $pkg " ;;
+ *) pkg_depends "$pkg" raw
+ esac done
+
+ # Filter the list, only keeping explicit packages.
+ # The purpose of these two loops is to order the
+ # argument list based on dependence.
+ for pkg in $deps; do ! contains "$*" "$pkg" || {
+ order="$order $pkg "
+ redro=" $pkg $redro"
+ } done
+
+ deps=
+}
+
+pkg_strip() {
+ # Strip package binaries and libraries. This saves space on the
+ # system as well as on the tarballs we ship for installation.
+
+ # Package has stripping disabled, stop here.
+ [ -f "$mak_dir/$pkg/nostrip" ] && return
+
+ log "$1" "Stripping binaries and libraries"
+
+ find "$pkg_dir/$1" -type f | while read -r file; do
+ case $(od -A o -t c -N 18 "$file") in
+ # REL (object files (.o), static libraries (.a)).
+ *177*E*L*F*0000020\ 001\ *|*\!*\<*a*r*c*h*\>*)
+ strip -g -R .comment -R .note "$file"
+ ;;
+
+ # EXEC (static binaries).
+ # DYN (shared libraries, dynamic binaries).
+ # Shared libraries keep global symbols in a separate ELF section
+ # called '.dynsym'. '--strip-all/-s' does not touch the dynamic
+ # symbol entries which makes this safe to do.
+ *177*E*L*F*0000020\ 00[23]\ *)
+ strip -s -R .comment -R .note "$file"
+ ;;
+ esac
+ done 2>/dev/null ||:
+}
+
+pkg_fix_deps() {
+ # Dynamically look for missing runtime dependencies by checking each binary
+ # and library with either 'ldd' or 'readelf'. This catches any extra
+ # libraries and or dependencies pulled in by the package's build suite.
+ log "$1" "Checking for missing dependencies"
+
+ # Go to the directory containing the built package to
+ # simplify path building.
+ cd "$pkg_dir/$1/$pkg_db/$1"
+
+ # Make a copy of the depends file if it exists to have a
+ # reference to 'diff' against.
+ if [ -f depends ]; then
+ cp -f depends "$mak_dir/d"
+ dep_file=$mak_dir/d
+ else
+ dep_file=/dev/null
+ fi
+
+ # Generate a list of all installed manifests.
+ pkg_name=$1
+ set +f; set -f -- "$sys_db/"*/manifest
+
+ # Get a list of binaries and libraries, false files
+ # will be found, however it's faster to get 'ldd' to check
+ # them anyway than to filter them out.
+ find "$pkg_dir/$pkg_name/" -type f 2>/dev/null |
+
+ while read -r file; do
+ case ${elf_prog:-ldd} in
+ *readelf) "$elf_prog" -d "$file" 2>/dev/null ;;
+ *) ldd "$file" 2>/dev/null ;;
+ esac |
+ while read -r dep; do
+ # Skip lines containing 'ldd'.
+ [ "${dep##*ldd*}" ] || continue
+ case $dep in *NEEDED*\[*\] | *'=>'*) ;; *) continue; esac
+
+ # readelf output:
+ # 0x0000 (NEEDED) Shared library: [libc.so]
+ dep=${dep##*\[}
+ dep=${dep%%\]*}
+
+ # ldd output:
+ # libc.so => /lib/ld-musl-x86_64.so.1
+ dep=${dep#* => }
+ dep=${dep% *}
+
+ # Figure out which package owns the file. Skip file if it is owned
+ # by the current package. This also handles cases where a '*-bin'
+ # package exists on the system, so the package manager doesn't think
+ # that the package we are building depends on the *-bin version of
+ # itself, or any other renamed versions of the same software.
+ pkg_owner -l "/${dep#/}\$" "$PWD/manifest" >/dev/null && continue
+ pkg_owner -l "/${dep#/}\$" "$@" ||:
+ done ||:
+ done >> depends
+
+ # Remove duplicate entries from the new depends file.
+ # This removes duplicate lines looking *only* at the
+ # first column.
+ sort -uk1,1 -o depends depends 2>/dev/null ||:
+
+ # Display a diff of the new dependencies against the old ones.
+ diff -U 3 "$dep_file" depends 2>/dev/null ||:
+
+ # Remove the depends file if it is empty.
+ [ -s depends ] || rm -f depends
+}
+
+pkg_manifest() (
+ # Generate the package's manifest file. This is a list of each file
+ # and directory inside the package. The file is used when uninstalling
+ # packages, checking for package conflicts and for general debugging.
+ log "$1" "Generating manifest"
+
+ # This function runs as a sub-shell to avoid having to 'cd' back to the
+ # prior directory before being able to continue.
+ cd "${2:-$pkg_dir}/$1"
+
+ # find: Print all files and directories and append '/' to directories.
+ # sort: Sort the output in *reverse*. Directories appear *after* their
+ # contents.
+ # sed: Remove the first character in each line (./dir -> /dir) and
+ # remove all lines which only contain '.'.
+ find . -type d -exec printf '%s/\n' {} + -o -print |
+ sort -r | sed '/^\.\/$/d;ss.ss' > "${2:-$pkg_dir}/$1/$pkg_db/$1/manifest"
+)
+
+pkg_etcsums() (
+ # This function runs as a sub-shell to avoid having to 'cd' back to the
+ # prior directory before being able to continue.
+ cd "$pkg_dir/$1/etc" 2>/dev/null || return 0; cd ..
+
+ # Generate checksums for each configuration file in the package's
+ # /etc/ directory for use in "smart" handling of these files.
+ log "$1" "Generating etcsums"
+
+
+ find etc -type f | while read -r file; do
+ sh256 "$file"
+ done > "$pkg_dir/$1/$pkg_db/$1/etcsums"
+)
+
+pkg_tar() {
+ # Create a tarball from the built package's files.
+ # This tarball also contains the package's database entry.
+ log "$1" "Creating tarball"
+
+ # Read the version information to name the package.
+ read -r version release < "$(pkg_find "$1")/version"
+
+ # Create a tarball from the contents of the built package.
+ "$tar" cf - -C "$pkg_dir/$1" . |
+ case $CPT_COMPRESS in
+ bz2) bzip2 -z ;;
+ xz) xz -zT 0 ;;
+ gz) gzip -6 ;;
+ zst) zstd -3 ;;
+ *) gzip -6 ;; # Fallback to gzip
+ esac \
+ > "$bin_dir/$1#$version-$release.tar.$CPT_COMPRESS"
+
+ log "$1" "Successfully created tarball"
+
+ run_hook post-package "$1" "$bin_dir/$1#$version-$release.tar.$CPT_COMPRESS"
+}
+
+pkg_build() {
+ # Build packages and turn them into packaged tarballs. This function
+ # also checks checksums, downloads sources and ensure all dependencies
+ # are installed.
+ pkg_build=1
+
+ log "Resolving dependencies"
+
+ for pkg do contains "$explicit" "$pkg" || {
+ pkg_depends "$pkg" explicit
+
+ # Mark packages passed on the command-line
+ # separately from those detected as dependencies.
+ explicit="$explicit $pkg "
+ } done
+
+ [ "$pkg_update" ] || explicit_build=$explicit
+
+ # If an explicit package is a dependency of another explicit
+ # package, remove it from the explicit list as it needs to be
+ # installed as a dependency.
+ # shellcheck disable=2086
+ for pkg do
+ contains "$deps" "$pkg" && explicit=$(pop "$pkg" from $explicit)
+ done
+
+ # See [1] at top of script.
+ # shellcheck disable=2046,2086
+ set -- $deps $explicit
+
+ log "Building: $*"
+
+ # Only ask for confirmation if more than one package needs to be built.
+ [ $# -gt 1 ] || [ "$pkg_update" ] && { prompt || exit 0 ;}
+
+ log "Checking for pre-built dependencies"
+
+ for pkg do pkg_lint "$pkg"; done
+
+ # Install any pre-built dependencies if they exist in the binary
+ # directory and are up to date.
+ for pkg do ! contains "$explicit_build" "$pkg" && pkg_cache "$pkg" && {
+ log "$pkg" "Found pre-built binary, installing"
+ (CPT_FORCE=1 cpt-install "$tar_file")
+
+ # Remove the now installed package from the build list.
+ # See [1] at top of script.
+ # shellcheck disable=2046,2086
+ set -- $(pop "$pkg" from "$@")
+ } done
+
+ for pkg do pkg_sources "$pkg"; done
+
+ pkg_verify "$@"
+
+ # Finally build and create tarballs for all passed packages and
+ # dependencies.
+ for pkg do
+ log "$pkg" "Building package ($((in = in + 1))/$#)"
+
+ pkg_extract "$pkg"
+ repo_dir=$(pkg_find "$pkg")
+
+ read -r build_version _ < "$repo_dir/version"
+
+ # Copy the build file to the build directory to users to modify it
+ # temporarily at runtime.
+ cp -f "$repo_dir/build" "$mak_dir/$pkg/.build.cpt"
+
+ # Install built packages to a directory under the package name
+ # to avoid collisions with other packages.
+ mkdir -p "$pkg_dir/$pkg/$pkg_db"
+
+ # Move to the build directory.
+ cd "$mak_dir/$pkg"
+
+ log "$pkg" "Starting build"
+
+ run_hook pre-build "$pkg" "$pkg_dir/$pkg"
+
+ # Notify the user if the build script is changed during the pre-build
+ # hook.
+ diff -q "$repo_dir/build" .build.cpt ||
+ log "$pkg" "Executing the modified build file"
+
+ # Call the build script, log the output to the terminal
+ # and to a file. There's no PIPEFAIL in POSIX shelll so
+ # we must resort to tricks like killing the script ourselves.
+ { ./.build.cpt "$pkg_dir/$pkg" "$build_version" "$sys_arch" 2>&1 || {
+ log "$pkg" "Build failed"
+ log "$pkg" "Log stored to $log_dir/$pkg-$time-$pid"
+ run_hook build-fail "$pkg" "$pkg_dir/$pkg"
+ pkg_clean
+ kill 0
+ } } | tee "$log_dir/$pkg-$time-$pid"
+
+ # Run the test script if it exists and the user wants to run tests. This
+ # is turned off by default.
+ [ -x "$repo_dir/test" ] && [ "$CPT_TEST" = 1 ] && {
+ run_hook pre-test "$pkg" "$pkg_dir/$pkg"
+ log "$pkg" "Running tests"
+ "$repo_dir/test" "$pkg_dir/$pkg" "$build_version" "$sys_arch" 2>&1 || {
+ log "$pkg" "Test failed"
+ log "$pkg" "Log stored to $log_dir/$pkg-$time-$pid"
+ run_hook test-fail "$pkg" "$pkg_dir/$pkg"
+ pkg_clean
+ kill 0
+ } } | tee -a "$log_dir/$pkg-$time-$pid"
+
+ # Delete the log file if the build succeeded to prevent
+ # the directory from filling very quickly with useless logs.
+ [ "$CPT_KEEPLOG" = 1 ] || rm -f "$log_dir/$pkg-$time-$pid"
+
+ # Copy the repository files to the package directory.
+ # This acts as the database entry.
+ cp -LRf "$repo_dir" "$pkg_dir/$pkg/$pkg_db/"
+
+ # Copy the modified build file to the package directory.
+ pkg_build="$pkg_dir/$pkg/$pkg_db/$pkg/build"
+ diff -U 3 "$pkg_build" .build.cpt > "$pkg_build.diff" &&
+ rm -f "$pkg_build.diff"
+
+ # We don't want the package manager to track 'dir' pages of the info
+ # directory. We don't want every single package to create their own dir
+ # files either.
+ rm -f "$pkg_dir/$pkg/usr/share/info/dir"
+
+ # We never ever want this. Let's end the endless conflicts
+ # and remove it.
+ find "$pkg_dir/$pkg" -name charset.alias -exec rm -f {} +
+
+ # Remove libtool's '*.la' library files. This removes cross-build
+ # system conflicts that may arise. Build-systems change, libtool
+ # is getting deprecated, we don't want a package that depends on
+ # some package's '.la' files.
+ find "$pkg_dir/$pkg" -name '*.la' -exec rm -f {} +
+
+ log "$pkg" "Successfully built package"
+
+ run_hook post-build "$pkg" "$pkg_dir/$pkg"
+
+ # Create the manifest file early and make it empty.
+ # This ensures that the manifest is added to the manifest.
+ : > "$pkg_dir/$pkg/$pkg_db/$pkg/manifest"
+
+ # If the package contains '/etc', add a file called
+ # 'etcsums' to the manifest. See comment directly above.
+ [ -d "$pkg_dir/$pkg/etc" ] &&
+ : > "$pkg_dir/$pkg/$pkg_db/$pkg/etcsums"
+
+ pkg_strip "$pkg"
+ pkg_manifest "$pkg"
+ pkg_fix_deps "$pkg"
+ pkg_manifest "$pkg"
+ pkg_etcsums "$pkg"
+ pkg_tar "$pkg"
+
+ # Install only dependencies of passed packages.
+ # Skip this check if this is a package update.
+ contains "$explicit" "$pkg" && [ -z "$pkg_update" ] && continue
+
+ log "$pkg" "Needed as a dependency or has an update, installing"
+
+ (CPT_FORCE=1 cpt-install "$pkg")
+ done
+
+ # End here as this was a system update and all packages have been installed.
+ [ "$pkg_update" ] && return
+
+ log "Successfully built package(s)"
+
+ # Turn the explicit packages into a 'list'.
+ # See [1] at top of script.
+ # shellcheck disable=2046,2086
+ set -- $explicit
+
+ # Only ask for confirmation if more than one package needs to be installed.
+ [ $# -gt 1 ] && prompt "Install built packages? [$*]" && {
+ cpt-install "$@"
+ return
+ }
+
+ log "Run 'cpt i $*' to install the package(s)"
+}
+
+pkg_checksums() {
+ # Generate checksums for packages.
+ repo_dir=$(pkg_find "$1")
+
+ [ -f "$repo_dir/sources" ] || return 0
+
+ while read -r src _ || [ "$src" ]; do
+ # Comment.
+ if [ -z "${src##\#*}" ]; then
+ continue
+
+ # File is local to the package.
+ elif [ -f "$repo_dir/$src" ]; then
+ src_path=$repo_dir/${src%/*}
+
+ # File is remote and was downloaded.
+ elif [ -f "$src_dir/$1/${src##*/}" ]; then
+ src_path=$src_dir/$1
+
+ # File is a git repository.
+ elif [ -z "${src##git+*}" ]; then continue
+
+ # Die here if source for some reason, doesn't exist.
+ else
+ die "$1" "Couldn't find source '$src'"
+ fi
+
+ # An easy way to get 'sha256sum' to print with the 'basename'
+ # of files is to 'cd' to the file's directory beforehand.
+ (cd "$src_path" && sh256 "${src##*/}") ||
+ die "$1" "Failed to generate checksums"
+ done < "$repo_dir/sources"
+}
+
+pkg_verify() {
+ # Verify all package checksums. This is achieved by generating a new set of
+ # checksums and then comparing those with the old set.
+ verify_cmd="NR==FNR{a[\$1];next}/^git .*/{next}!((\$1)in a){exit 1}"
+
+ for pkg; do
+ repo_dir=$(pkg_find "$pkg")
+ [ -f "$repo_dir/sources" ] || continue
+
+ pkg_checksums "$pkg" | awk "$verify_cmd" - "$repo_dir/checksums" || {
+ log "$pkg" "Checksum mismatch"
+
+ # Instead of dying above, log it to the terminal. Also define a
+ # variable so we *can* die after all checksum files have been
+ # checked.
+ mismatch="$mismatch$pkg "
+ } done
+
+ [ -z "$mismatch" ] || die "Checksum mismatch with: ${mismatch% }"
+}
+
+pkg_conflicts() {
+ # Check to see if a package conflicts with another.
+ log "$1" "Checking for package conflicts"
+
+ # Filter the tarball's manifest and select only files
+ # and any files they resolve to on the filesystem
+ # (/bin/ls -> /usr/bin/ls).
+ while read -r file; do
+ case $file in */) continue; esac
+
+ # Use $CPT_ROOT in filename so that we follow its symlinks.
+ file=$CPT_ROOT/${file#/}
+
+ # We will only follow the symlinks of the directories, so we reserve the
+ # directory name in this 'dirname' value. If we cannot find it in the
+ # system, we don't need to make this much more complex by trying so hard
+ # to find it. Simply use the original directory name.
+ dirname="$(_readlinkf "${file%/*}" 2>/dev/null)" ||
+ dirname="${file%/*}"
+
+
+ # Combine the dirname and file values, and print them into the
+ # temporary manifest to be parsed.
+ printf '%s/%s\n' "${dirname#$CPT_ROOT}" "${file##*/}"
+
+ done < "$tar_dir/$1/$pkg_db/$1/manifest" > "$CPT_TMPDIR/$pid/manifest"
+
+ p_name=$1
+
+ # Generate a list of all installed package manifests
+ # and remove the current package from the list.
+ # shellcheck disable=2046,2086
+ set -- $(set +f; pop "$sys_db/$p_name/manifest" from "$sys_db"/*/manifest)
+
+ [ -s "$CPT_TMPDIR/$pid/manifest" ] || return 0
+
+ # In rare cases where the system only has one package installed
+ # and you are reinstalling that package, grep will try to read from
+ # standard input if we continue here.
+ #
+ # Also, if we don't have any packages installed grep will give an
+ # error. This will not cause the installation to fail, but we don't
+ # need to check for conflicts if that's the case anyway. If we have
+ # only zero packages or one package, just stop wasting time and continue
+ # with the installation.
+ [ "$1" ] && [ -f "$1" ] || return 0
+
+ # Store the list of found conflicts in a file as we will be using the
+ # information multiple times. Storing it in the cache dir allows us
+ # to be lazy as they'll be automatically removed on script end.
+ "$grep" -Fxf "$CPT_TMPDIR/$pid/manifest" -- "$@" > "$CPT_TMPDIR/$pid/conflict" ||:
+
+
+ # Enable alternatives automatically if it is safe to do so.
+ # This checks to see that the package that is about to be installed
+ # doesn't overwrite anything it shouldn't in '/var/db/cpt/installed'.
+ "$grep" -q ":/var/db/cpt/installed/" "$CPT_TMPDIR/$pid/conflict" ||
+ choice_auto=1
+
+ # Use 'grep' to list matching lines between the to
+ # be installed package's manifest and the above filtered
+ # list.
+ if [ "$CPT_CHOICE" != 0 ] && [ "$choice_auto" = 1 ]; then
+
+ # This is a novel way of offering an "alternatives" system.
+ # It is entirely dynamic and all "choices" are created and
+ # destroyed on the fly.
+ #
+ # When a conflict is found between two packages, the file
+ # is moved to a directory called "choices" and its name
+ # changed to store its parent package and its intended
+ # location.
+ #
+ # The package's manifest is then updated to reflect this
+ # new location.
+ #
+ # The 'cpt choices' command parses this directory and
+ # offers you the CHOICE of *swapping* entries in this
+ # directory for those on the filesystem.
+ #
+ # The choices command does the same thing we do here,
+ # it rewrites manifests and moves files around to make
+ # this work.
+ #
+ # Pretty nifty huh?
+ while IFS=: read -r _ con; do
+ printf '%s\n' "Found conflict $con"
+
+ # Create the "choices" directory inside of the tarball.
+ # This directory will store the conflicting file.
+ mkdir -p "$tar_dir/$p_name/${cho_dir:=var/db/cpt/choices}"
+
+ # Construct the file name of the "db" entry of the
+ # conflicting file. (pkg_name>usr>bin>ls)
+ con_name=$(printf %s "$con" | sed 's|/|>|g')
+
+ # Move the conflicting file to the choices directory
+ # and name it according to the format above.
+ mv -f "$tar_dir/$p_name/$con" \
+ "$tar_dir/$p_name/$cho_dir/$p_name$con_name" 2>/dev/null || {
+ log "File must be in ${con%/*} and not a symlink to it"
+ log "This usually occurs when a binary is installed to"
+ log "/sbin instead of /usr/bin (example)"
+ log "Before this package can be used as an alternative,"
+ log "this must be fixed in $p_name. Contact the maintainer"
+ die "by checking 'git log' or by running 'cpt-maintainer'"
+ }
+ done < "$CPT_TMPDIR/$pid/conflict"
+
+ # Rewrite the package's manifest to update its location
+ # to its new spot (and name) in the choices directory.
+ pkg_manifest "$p_name" "$tar_dir" 2>/dev/null
+
+ elif [ -s "$CPT_TMPDIR/$pid/conflict" ]; then
+ log "Package '$p_name' conflicts with another package" "" "!>"
+ log "Run 'CPT_CHOICE=1 cpt i $p_name' to add conflicts" "" "!>"
+ die "as alternatives."
+ fi
+}
+
+pkg_swap() {
+ # Swap between package alternatives.
+ pkg_list "$1" >/dev/null
+
+ alt=$(printf %s "$1$2" | sed 's|/|>|g')
+ cd "$sys_db/../choices"
+
+ [ -f "$alt" ] || [ -h "$alt" ] ||
+ die "Alternative '$1 $2' doesn't exist"
+
+ if [ -f "$2" ]; then
+ # Figure out which package owns the file we are going to swap for
+ # another package's.
+ #
+ # Print the full path to the manifest file which contains
+ # the match to our search.
+
+ pkg_owns=$(pkg_owner -lFx "$2") ||
+ die "File '$2' exists on filesystem but isn't owned"
+
+ log "Swapping '$2' from '$pkg_owns' to '$1'"
+
+ # Convert the current owner to an alternative and rewrite
+ # its manifest file to reflect this. We then resort this file
+ # so no issues arise when removing packages.
+ cp -Pf "$CPT_ROOT/$2" "$pkg_owns>${alt#*>}"
+ sed "s#^$(regesc "$2")\$#${PWD#$CPT_ROOT}/$pkg_owns>${alt#*>}#" \
+ "../installed/$pkg_owns/manifest" |
+ sort -r -o "../installed/$pkg_owns/manifest"
+ fi
+
+ # Convert the desired alternative to a real file and rewrite
+ # the manifest file to reflect this. The reverse of above.
+ mv -f "$alt" "$CPT_ROOT/$2"
+ sed "s#^${PWD#$CPT_ROOT}/$(regesc "$alt")\$#$2#" "../installed/$1/manifest" |
+ sort -r -o "../installed/$1/manifest"
+}
+
+pkg_etc() {
+ [ -d "$tar_dir/$pkg_name/etc" ] || return 0
+
+ (cd "$tar_dir/$pkg_name"
+
+ # Create all directories beforehand.
+ find etc -type d | while read -r dir; do
+ mkdir -p "$CPT_ROOT/$dir"
+ done
+
+ # Handle files in /etc/ based on a 3-way checksum check.
+ find etc ! -type d | while read -r file; do
+ { sum_new=$(sh256 "$file")
+ sum_sys=$(cd "$CPT_ROOT/"; sh256 "$file")
+ sum_old=$("$grep" "$file$" "$mak_dir/c"); } 2>/dev/null ||:
+
+ log "$pkg_name" "Doing 3-way handshake for $file"
+ printf '%s\n' "Previous: ${sum_old:-null}"
+ printf '%s\n' "System: ${sum_sys:-null}"
+ printf '%s\n' "New: ${sum_new:-null}"
+
+ # Use a case statement to easily compare three strings at
+ # the same time. Pretty nifty.
+ case ${sum_old:-null}${sum_sys:-null}${sum_new} in
+ # old = Y, sys = X, new = Y
+ "${sum_new}${sum_sys}${sum_old}")
+ log "Skipping $file"
+ continue
+ ;;
+
+ # old = X, sys = X, new = X
+ # old = X, sys = Y, new = Y
+ # old = X, sys = X, new = Y
+ "${sum_old}${sum_old}${sum_old}"|\
+ "${sum_old:-null}${sum_sys}${sum_sys}"|\
+ "${sum_sys}${sum_old}"*)
+ log "Installing $file"
+ new=
+ ;;
+
+ # All other cases.
+ *)
+ warn "$pkg_name" "saving /$file as /$file.new" "->"
+ new=.new
+ ;;
+ esac
+
+ cp -fPp "$file" "$CPT_ROOT/${file}${new}"
+ chown root:root "$CPT_ROOT/${file}${new}" 2>/dev/null
+ done) ||:
+}
+
+pkg_remove() {
+ # Remove a package and all of its files. The '/etc' directory
+ # is handled differently and configuration files are *not*
+ # overwritten.
+ pkg_list "$1" >/dev/null || return
+
+ # Make sure that nothing depends on this package.
+ [ "$CPT_FORCE" = 1 ] || {
+ log "$1" "Checking for reverse dependencies"
+
+ (cd "$sys_db"; set +f; grep -lFx "$1" -- */depends) &&
+ die "$1" "Can't remove package, others depend on it"
+ }
+ # Block being able to abort the script with 'Ctrl+C' during removal.
+ # Removes all risk of the user aborting a package removal leaving
+ # an incomplete package installed.
+ trap_set block
+
+ if [ -x "$sys_db/$1/pre-remove" ]; then
+ log "$1" "Running pre-remove script"
+ "$sys_db/$1/pre-remove" ||:
+ fi
+
+ # Create a temporary list of all directories, so we don't accidentally
+ # remove anything from packages that create empty directories for a
+ # purpose (such as baselayout).
+ manifest_list="$(set +f; pop "$sys_db/$1/manifest" from "$sys_db/"*/manifest)"
+ # shellcheck disable=2086
+ [ "$manifest_list" ] && grep -h '/$' $manifest_list | sort -ur > "$mak_dir/dirs"
+
+ run_hook pre-remove "$1" "$sys_db/$1" root
+
+ while read -r file; do
+ # The file is in '/etc' skip it. This prevents the package
+ # manager from removing user edited configuration files.
+ [ "${file##/etc/*}" ] || continue
+
+ if [ -d "$CPT_ROOT/$file" ]; then
+ "$grep" -Fxq "$file" "$mak_dir/dirs" 2>/dev/null && continue
+ rmdir "$CPT_ROOT/$file" 2>/dev/null || continue
+ else
+ rm -f "$CPT_ROOT/$file"
+ fi
+ done < "$sys_db/$1/manifest"
+
+ # Reset 'trap' to its original value. Removal is done so
+ # we no longer need to block 'Ctrl+C'.
+ trap_set cleanup
+
+ run_hook post-remove "$1" "$CPT_ROOT/" root
+
+ log "$1" "Removed successfully"
+}
+
+pkg_install() {
+ # Install a built package tarball.
+
+ # Install can also take the full path to a tarball.
+ # We don't need to check the repository if this is the case.
+ if [ -f "$1" ] && [ -z "${1%%*.tar*}" ] ; then
+ tar_file=$1
+ pkg_name=${1##*/}
+ pkg_name=${pkg_name%#*}
+
+ else
+ pkg_cache "$1" ||
+ die "package has not been built, run 'cpt b pkg'"
+
+ pkg_name=$1
+ fi
+
+ mkdir -p "$tar_dir/$pkg_name"
+ log "$pkg_name" "Extracting $tar_file"
+
+ # Extract the tarball to catch any errors before installation begins.
+ decompress "$tar_file" | "$tar" xf - -C "$tar_dir/$pkg_name"
+
+ [ -f "$tar_dir/$pkg_name/$pkg_db/$pkg_name/manifest" ] ||
+ die "'${tar_file##*/}' is not a valid CPT package"
+
+ # Ensure that the tarball's manifest is correct by checking that
+ # each file and directory inside of it actually exists.
+ [ "$CPT_FORCE" != 1 ] && log "$pkg_name" "Checking package manifest" &&
+ while read -r line; do
+ # Skip symbolic links
+ [ -h "$tar_dir/$pkg_name/$line" ] ||
+ [ -e "$tar_dir/$pkg_name/$line" ] || {
+ log "File $line missing from tarball but mentioned in manifest" "" "!>"
+ TARBALL_FAIL=1
+ }
+ done < "$tar_dir/$pkg_name/$pkg_db/$pkg_name/manifest"
+ [ "$TARBALL_FAIL" ] && {
+ log "You can still install this package by setting CPT_FORCE variable"
+ die "$pkg_name" "Missing files in manifest"
+ }
+
+ log "$pkg_name" "Checking that all dependencies are installed"
+
+ # Make sure that all run-time dependencies are installed prior to
+ # installing the package.
+ [ -f "$tar_dir/$pkg_name/$pkg_db/$pkg_name/depends" ] &&
+ [ "$CPT_FORCE" != 1 ] &&
+ while read -r dep dep_type || [ "$dep" ]; do
+ [ "${dep##\#*}" ] || continue
+ [ "$dep_type" ] || pkg_list "$dep" >/dev/null ||
+ install_dep="$install_dep'$dep', "
+ done < "$tar_dir/$pkg_name/$pkg_db/$pkg_name/depends"
+
+ [ "$install_dep" ] && die "$1" "Package requires ${install_dep%, }"
+
+ run_hook pre-install "$pkg_name" "$tar_dir/$pkg_name" root
+
+ pkg_conflicts "$pkg_name"
+
+ log "$pkg_name" "Installing package incrementally"
+
+ # Block being able to abort the script with Ctrl+C during installation.
+ # Removes all risk of the user aborting a package installation leaving
+ # an incomplete package installed.
+ trap_set block
+
+ # If the package is already installed (and this is an upgrade) make a
+ # backup of the manifest and etcsums files.
+ cp -f "$sys_db/$pkg_name/manifest" "$mak_dir/m" 2>/dev/null ||:
+ cp -f "$sys_db/$pkg_name/etcsums" "$mak_dir/c" 2>/dev/null ||:
+
+ # This is repeated multiple times. Better to make it a function.
+ pkg_rsync() {
+ rsync "--chown=$USER:$USER" --chmod=Du-s,Dg-s,Do-s \
+ -WhHKa --no-compress --exclude /etc "${1:---}" \
+ "$tar_dir/$pkg_name/" "$CPT_ROOT/"
+ }
+
+ # Install the package by using 'rsync' and overwrite any existing files
+ # (excluding '/etc/').
+ pkg_rsync --info=progress2
+ pkg_etc
+
+ # Remove any leftover files if this is an upgrade.
+ "$grep" -vFxf "$sys_db/$pkg_name/manifest" "$mak_dir/m" 2>/dev/null |
+
+ while read -r file; do
+ file=$CPT_ROOT/$file
+
+ # Skip deleting some leftover files.
+ case $file in /etc/*) continue; esac
+
+ # Remove files.
+ if [ -f "$file" ] && [ ! -L "$file" ]; then
+ rm -f "$file"
+
+ # Remove file symlinks.
+ elif [ -h "$file" ] && [ ! -d "$file" ]; then
+ unlink "$file" ||:
+
+ # Skip directory symlinks.
+ elif [ -h "$file" ] && [ -d "$file" ]; then :
+
+ # Remove directories if empty.
+ elif [ -d "$file" ]; then
+ rmdir "$file" 2>/dev/null ||:
+ fi
+ done ||:
+
+ log "$pkg_name" "Verifying installation"
+ { pkg_rsync; pkg_rsync; } ||:
+
+ # Reset 'trap' to its original value. Installation is done so
+ # we no longer need to block 'Ctrl+C'.
+ trap_set cleanup
+
+ if [ -x "$sys_db/$pkg_name/post-install" ]; then
+ log "$pkg_name" "Running post-install script"
+ "$sys_db/$pkg_name/post-install" ||:
+ fi
+
+ run_hook post-install "$pkg_name" "$sys_db/$pkg_name" root
+
+ log "$pkg_name" "Installed successfully"
+}
+
+pkg_fetch() {
+ log "Updating repositories"
+
+ run_hook pre-fetch
+
+ # Create a list of all repositories.
+ # See [1] at top of script.
+ # shellcheck disable=2046,2086
+ { IFS=:; set -- $CPT_PATH; IFS=$old_ifs ;}
+
+ # Update each repository in '$CPT_PATH'. It is assumed that
+ # each repository is 'git' tracked.
+ for repo; do
+ # Go to the root of the repository (if it exists).
+ cd "$repo"
+ cd "$(git rev-parse --show-toplevel 2>/dev/null)" 2>/dev/null ||
+ cd "$(hg root 2>/dev/null)" 2>/dev/null ||:
+
+ if [ -d .git ]; then
+
+ [ "$(git remote 2>/dev/null)" ] || {
+ log "$repo" " "
+ printf '%s\n' "No remote, skipping."
+ continue
+ }
+
+ contains "$repos" "$PWD" || {
+ repos="$repos $PWD "
+
+ # Display a tick if signing is enabled for this
+ # repository.
+ case $(git config merge.verifySignatures) in
+ true) log "$PWD" "[signed] " ;;
+ *) log "$PWD" " " ;;
+ esac
+
+ if [ -w "$PWD" ] && [ "$uid" != 0 ]; then
+ git fetch
+ git merge
+ git submodule update --remote --init -f
+
+ else
+ [ "$uid" = 0 ] || log "$PWD" "Need root to update"
+
+ # Find out the owner of the repository and spawn
+ # git as this user below.
+ #
+ # This prevents 'git' from changing the original
+ # ownership of files and directories in the rare
+ # case that the repository is owned by a 3rd user.
+ (
+ user=$(_stat "$PWD")
+
+ [ "$user" = root ] ||
+ log "Dropping permissions to $user for pull"
+
+ git_cmd="git fetch && git merge && git submodule update --remote --init -f"
+ case $su in *su) git_cmd="'$git_cmd'"; esac
+
+ # Spawn a subshell to run multiple commands as
+ # root at once. This makes things easier on users
+ # who aren't using persist/timestamps for auth
+ # caching.
+ user=$user as_root sh -c "$git_cmd"
+ )
+ fi
+ }
+ elif [ -d .hg ]; then
+
+ [ "$(hg showconfig paths 2>/dev/null)" ] || {
+ log "$repo" " "
+ printf '%s\n' "No remote, skipping."
+ continue
+ }
+
+ contains "$repos $PWD" || {
+ repos="$repos $PWD"
+
+ if [ -w "$PWD" ] && [ "$uid" != 0 ]; then
+ hg pull
+ hg update
+ else
+ [ "$uid" ] || log "$PWD" "Need root to update"
+
+ # We are going to do the same operation as above, to
+ # find the owner of the repository.
+ (
+ user=$(_stat "$PWD")
+
+ [ "$user" = root ] ||
+ log "Dropping permissions to $user for pull"
+
+ hg_cmd="hg pull && hg update"
+
+ case $su in *su) hg_cmd="'$hg_cmd'"; esac
+ user=$user as_root sh -c "$hg_cmd"
+ )
+ fi
+ }
+ elif [ -f .rsync ]; then
+ # If an .rsync_root file exists, we check that the repository root
+ # exists. If it does, we change to that directory to do the fetch.
+ # This way, we allow for partial repositories while making sure that
+ # we can fetch the repository in a single operation.
+ [ -f .rsync_root ] && {
+ read -r rsync_root < .rsync_root
+ [ -f "$rsync_root/.rsync" ] && cd "$rsync_root"
+ }
+ contains "$repos" "$PWD" || {
+ repos="$repos $PWD"
+ read -r remote < .rsync
+ if [ -w "$PWD" ] && [ "$uid" != 0 ]; then
+ rsync -acvzzC --include=core --delete "$remote/" "$PWD"
+ else
+ [ "$uid" = 0 ] || log "$PWD" "Need root to update"
+
+ # Similar to the git update, we find the owner of
+ # the repository and spawn rsync as that user.
+ (
+ user=$(_stat "$PWD")
+
+ [ "$user" = root ] ||
+ log "Dropping permissions to $user for pull"
+
+ user=$user as_root rsync -acvzzC --include=core --delete "$remote/" "$PWD"
+ )
+ fi
+ }
+ else
+ log "$repo" " "
+ printf '%s\n' "Not a remote repository, skipping."
+ fi
+ done
+
+ run_hook post-fetch
+}
+
+pkg_updates(){
+ # Check all installed packages for updates. So long as the installed
+ # version and the version in the repositories differ, it's considered
+ # an update.
+ [ "$CPT_FETCH" = 0 ] || pkg_fetch
+
+ log "Checking for new package versions"
+
+ set +f
+
+ for pkg in "$sys_db/"*; do
+ pkg_name=${pkg##*/}
+
+ # Read version and release information from the installed packages
+ # and repository.
+ read -r db_ver db_rel < "$pkg/version"
+ read -r re_ver re_rel < "$(pkg_find "$pkg_name")/version"
+
+ # Compare installed packages to repository packages.
+ [ "$db_ver-$db_rel" != "$re_ver-$re_rel" ] && {
+ printf '%s\n' "$pkg_name $db_ver-$db_rel ==> $re_ver-$re_rel"
+ outdated="$outdated$pkg_name "
+ }
+ done
+
+ set -f
+
+ # If the download option is specified only download the outdated packages
+ # and exit.
+ # shellcheck disable=2154
+ [ "$download_only" = 1 ] && {
+ log "Only sources for the packages will be acquired"
+ prompt || exit 0
+
+ for pkg in $outdated; do
+ pkg_sources "$pkg"
+ done
+
+ exit 0
+ }
+
+ contains "$outdated" cpt && {
+ log "Detected package manager update"
+ log "The package manager will be updated first"
+
+ prompt || exit 0
+
+ pkg_build cpt
+ cpt-install cpt
+
+ log "Updated the package manager"
+ log "Re-run 'cpt update' to update your system"
+
+ exit 0
+ }
+
+ [ "$outdated" ] || {
+ log "Everything is up to date"
+ return
+ }
+
+ log "Packages to update: ${outdated% }"
+
+ # Tell 'pkg_build' to always prompt before build.
+ pkg_update=1
+
+ # Build all packages requiring an update.
+ # See [1] at top of script.
+ # shellcheck disable=2046,2086
+ {
+ pkg_order $outdated
+ pkg_build $order
+ }
+
+ log "Updated all packages"
+}
+
+pkg_get_base() (
+ # Print the packages defined in the /etc/cpt-base file.
+ # If an argument is given, it prints a space seperated list instead
+ # of a list seperated by newlines.
+
+ # cpt-base is an optional file, return with success if it doesn't exist.
+ [ -f "$CPT_ROOT/etc/cpt-base" ] || return 0
+ nonl=$1; set --
+
+ # Older versions of shellcheck warns us that the variable is changing on the
+ # subshell. That is our purpose here, thank you very much.
+ # shellcheck disable=SC2030
+ while read -r pkgname _; do
+ [ "${pkgname##\#*}" ] || continue
+ set -- "$@" "$pkgname"
+ done < "$CPT_ROOT/etc/cpt-base"
+ if [ "$nonl" ]; then printf '%s ' "$@"; else printf '%s\n' "$@"; fi
+)
+
+pkg_gentree() (
+ # Generate an ordered dependency tree of a package. Useful for testing
+ # whether the generated dependency tree is enough to actually building a
+ # given package. A second argument can be given as a combination of
+ # characters (similar to 'tar(1)' keys) which will be used as an option
+ # parser. See the documentation for more information on the keys.
+ deps='' reverse='' nonl='' make_deps=first
+ for op in $(sepchar "$2"); do
+ case "$op" in
+ b) deps="$(pkg_get_base nonl)" ;;
+ x) make_deps=first-nomake ;;
+ r) reverse=1 ;;
+ n) nonl=1 ;;
+ *) return 1
+ esac
+ done
+ pkg_depends "$1" tree "${make_deps:+first}"
+ eval set -- "$deps"
+ pkg_order "$@"
+ if [ "$reverse" ]; then eval set -- "$redro"; else eval set -- "$order"; fi
+ if [ "$nonl" ]; then printf '%s ' "$@"; else printf '%s\n' "$@"; fi
+)
+
+pkg_clean() {
+ # Clean up on exit or error. This removes everything related
+ # to the build.
+ [ "$CPT_DEBUG" != 1 ] || return 0
+
+ # Block 'Ctrl+C' while cache is being cleaned.
+ trap_set block
+
+ # Remove temporary items.
+ rm -rf -- "${CPT_TMPDIR:=$cac_dir/proc}/$pid"
+}
+
+create_cache() {
+ # A temporary directory can be specified apart from the cache
+ # directory in order to build in a user specified directory.
+ # /tmp could be used in order to build on ram, useful on SSDs.
+ # The user can specify CPT_TMPDIR for this.
+ #
+ # Create the required temporary directories and set the variables
+ # which point to them.
+ mkdir -p "${tmp_dir:=${CPT_TMPDIR:=$cac_dir/proc}/$pid}"
+
+ # If an argument is given, skip the creation of other cache directories.
+ # This here makes shellcheck extremely angry, so I am globally disabling
+ # SC2119.
+ [ "$1" ] || mkdir -p "${mak_dir:=$tmp_dir/build}" \
+ "${pkg_dir:=$tmp_dir/pkg}" \
+ "${tar_dir:=$tmp_dir/export}"
+
+}
+
+# main()
+{
+ set -ef
+
+ # If a parser definition exists, let's run it ourselves. This makes sure we
+ # get the variables as soon as possible.
+ command -v parser_definition >/dev/null && {
+ eval "$(getoptions parser_definition parse "$0")"
+ parse "$@"
+ eval set -- "$REST"
+ }
+
+ # Create the cache directories for CPT and set the variables which point
+ # to them. This is seperate from temporary directories created in
+ # create_cache(). That's because we need these variables set on most
+ # occasions.
+ mkdir -p "${cac_dir:=${CPT_CACHE:=${XDG_CACHE_HOME:-$HOME/.cache}/cpt}}" \
+ "${src_dir:=$cac_dir/sources}" \
+ "${log_dir:=$cac_dir/logs}" \
+ "${bin_dir:=$cac_dir/bin}"
+
+ # Set the location to the repository and package database.
+ pkg_db=var/db/cpt/installed
+
+ # The PID of the current shell process is used to isolate directories
+ # to each specific CPT instance. This allows multiple package manager
+ # instances to be run at once. Store the value in another variable so
+ # that it doesn't change beneath us.
+ pid=${CPT_PID:-$$}
+
+ # Force the C locale to speed up things like 'grep' which disable unicode
+ # etc when this is set. We don't need unicode and a speed up is always
+ # welcome.
+ export LC_ALL=C LANG=C
+
+ # Catch errors and ensure that build files and directories are cleaned
+ # up before we die. This occurs on 'Ctrl+C' as well as success and error.
+ trap_set cleanup
+
+ # Prefer GNU grep if installed as it is much much faster than busybox's
+ # implementation. Very much worth it if you value performance over
+ # POSIX correctness (grep quoted to avoid shellcheck false-positive).
+ grep=$(command -v ggrep) || grep='grep'
+
+ # Prefer libarchive tar or GNU tar if installed as they are much
+ # much faster than busybox's implementation. Very much worth it if
+ # you value performance.
+ tar=$(command -v bsdtar || command -v gtar) || tar=tar
+
+ # Figure out which 'sudo' command to use based on the user's choice or
+ # what is available on the system.
+ su=${CPT_SU:-$(command -v sls ||
+ command -v sudo ||
+ command -v doas)} || su=su
+
+ # Store the date and time of script invocation to be used as the name
+ # of the log files the package manager creates uring builds.
+ time=$(date '+%Y-%m-%d-%H:%M')
+
+ # Use readelf for fixing dependencies if it is available, fallback to
+ # ldd. readelf shows only the actual dependencies and doesn't include
+ # the libraries required by the dependencies.
+ elf_prog=${CPT_ELF:="$(
+ command -v readelf ||
+ command -v llvm-readelf ||
+ command -v eu-readelf)"} || elf_prog=ldd
+
+ # Make note of the user's current ID to do root checks later on.
+ # This is used enough to warrant a place here.
+ uid=$(id -u)
+
+ # Save IFS, so we can restore it back to what it was before.
+ old_ifs=$IFS
+
+ # Make sure that the CPT_ROOT doesn't end with a '/'. This might
+ # break some operations.
+ [ -z "$CPT_ROOT" ] || [ "${CPT_ROOT##*/}" ] || {
+ warn "" "Your CPT_ROOT variable shouldn't end with '/'"
+ CPT_ROOT=${CPT_ROOT%/}
+ }
+
+ # Define an optional sys_arch variable in order to provide
+ # information to build files with architectural information.
+ sys_arch=$(uname -m 2>/dev/null) ||:
+
+ # Define this variable but don't create its directory structure from
+ # the get go. It will be created as needed by package installation.
+ sys_db=$CPT_ROOT/$pkg_db
+
+ # This allows for automatic setup of a CPT chroot and will
+ # do nothing on a normal system.
+ mkdir -p "$CPT_ROOT/" 2>/dev/null ||:
+
+ # Set a value for CPT_COMPRESS if it isn't set.
+ : "${CPT_COMPRESS:=gz}"
+
+ # Unless being piped or the user specifically doesn't want colors, set
+ # colors. This can of course be overriden if the user specifically want
+ # colors during piping.
+ if { [ "$CPT_COLOR" != 0 ] && [ -t 1 ] ;} || [ "$CPT_COLOR" = 1 ]; then
+ colory="\033[1;33m" colorb="\033[1;36m" colre="\033[m"
+ fi
+
+}
+
+# If the library is being called with its own name, run arguments.
+if [ "${0##*/}" = cpt-lib ]; then
+ pd() {
+ setup REST help:usage -- "usage: ${0##*/} [funcall...]"
+ global_options
+ }
+ eval "$(getoptions pd parse "$0")"
+ parse "$@"
+ eval set -- "$REST"
+ "$@"
+fi