#!/bin/sh -ef # @DOCSTRING@ # shellcheck source=/dev/null # # This is the Carbs Packaging Toolchain written for Carbs Linux. # It was originally forked from the kiss package manager by # Dylan Araps. # # Currently maintained by Cem Keylan. version() { log "Carbs Packaging Tools" @VERSION@ exit 0 } out() { # Print a message as is. printf '%s\n' "$@" } log() { # Print a message prettily. # # All messages are printed to stderr to allow the user to hide build # output which is the only thing printed to stdout. # # '${3:-->}': If the 3rd argument is missing, set prefix to '->'. # '${2:+colorb}': If the 2nd argument exists, set text style of '$1'. printf '%b%s %b%b%s%b %s\n' \ "$colory" "${3:-->}" "$colre" "${2:+$colorb}" "$1" "$colre" "$2" >&2 } die() { # Print a message and exit with '1' (error). log "$1" "$2" "!>" exit 1 } trap_set() { # Function to set the trap value. case ${1:-cleanup} in cleanup) trap pkg_clean EXIT INT ;; block) trap '' INT ;; unset) trap - EXIT INT ;; esac } _stat() ( _user=; eval set -- "$(ls -ld "$1")" id -u "${_user:=$3}" >/dev/null 2>&1 || _user=root printf '%s' "$_user" ) _readlinkf() ( # Public domain POSIX sh readlink function by Koichi Nakashima [ "${1:-}" ] || return 1 max_symlinks=40 CDPATH='' # to avoid changing to an unexpected directory target=$1 [ -e "${target%/}" ] || target=${1%"${1##*[!/]}"} # trim trailing slashes [ -d "${target:-/}" ] && target="$target/" cd -P . 2>/dev/null || return 1 while [ "$max_symlinks" -ge 0 ] && max_symlinks=$((max_symlinks - 1)); do if [ ! "$target" = "${target%/*}" ]; then case $target in /*) cd -P "${target%/*}/" 2>/dev/null || break ;; *) cd -P "./${target%/*}" 2>/dev/null || break ;; esac target=${target##*/} fi if [ ! -L "$target" ]; then target="${PWD%/}${target:+/}${target}" printf '%s\n' "${target:-/}" return 0 fi # `ls -dl` format: "%s %u %s %s %u %s %s -> %s\n", # , , , , # , , , # https://pubs.opengroup.org/onlinepubs/9699919799/utilities/ls.html link=$(ls -dl -- "$target" 2>/dev/null) || break target=${link#*" $target -> "} done return 1 ) # This is the public domain getoptions shell library. It also forms a usage # function. # URL: https://github.com/ko1nksm/getoptions (v2.0.1) # License: Creative Commons Zero v1.0 Universal # shellcheck disable=2016 getoptions() { _error='' _on=1 _off='' _export='' _plus='' _mode='' _alt='' _rest='' _opts='' _help='' _indent='' _init=@empty IFS=' ' for i in 0 1 2 3 4 5; do eval "_$i() { echo \"$_indent\$@\"; }" _indent="$_indent " done quote() { q="$2'" r='' while [ "$q" ]; do r="$r${q%%\'*}'\''" && q=${q#*\'}; done q="'${r%????}'" && q=${q#\'\'} && q=${q%\'\'} eval "$1=\${q:-\"''\"}" } code() { [ "${1#:}" = "$1" ] && c=3 || c=4 eval "[ ! \${$c:+x} ] || $2 \"\$$c\"" } invoke() { eval '"_$@"'; } prehook() { invoke "$@"; } for i in setup flag param option disp msg; do eval "$i() { prehook $i \"\$@\"; }" done args() { on=$_on off=$_off export=$_export init=$_init _hasarg=$1 while [ $# -gt 2 ] && [ "$3" != '--' ] && shift; do case $2 in -?) [ "$_hasarg" ] || _opts="$_opts${2#-}" ;; +*) _plus=1 ;; [!-+]*) eval "${2%%:*}=\${2#*:}" esac done } defvar() { case $init in @none) : ;; @export) code "$1" _0 "export $1" ;; @empty) code "$1" _0 "${export:+export }$1=''" ;; @unset) code "$1" _0 "unset $1 ||:" "unset OPTARG ||:; ${1#:}" ;; *) case $init in @*) eval "init=\"=\${${init#@}}\""; esac case $init in [!=]*) _0 "$init"; return 0; esac quote init "${init#=}" code "$1" _0 "${export:+export }$1=$init" "OPTARG=$init; ${1#:}" esac } _setup() { [ $# -gt 0 ] && { [ "$1" ] && _rest=$1; shift; } for i; do [ "$i" = '--' ] && break; eval "_${i%%:*}=\${i#*:}"; done } _flag() { args : "$@"; defvar "$@"; } _param() { args '' "$@"; defvar "$@"; } _option() { args '' "$@"; defvar "$@"; } _disp() { args : "$@"; } _msg() { args : _ "$@"; } "$@" _0 "${_rest:?}=''" args() { sw='' validate='' pattern='' counter='' on=$_on off=$_off export=$_export while [ $# -gt 1 ] && [ "$2" != '--' ] && shift; do case $1 in --\{no-\}*) sw="$sw${sw:+ | }--${1#--?no-?} | --no-${1#--?no-?}" ;; [-+]? | --*) sw="$sw${sw:+ | }$1" ;; *) eval "${1%%:*}=\"\${1#*:}\"" esac done } setup() { :; } _flag() { args "$@" quote on "$on" && quote off "$off" [ "$counter" ] && on=1 off=-1 v="\$((\${$1:-0}+\${OPTARG:-0}))" || v='' _3 "$sw)" _4 '[ "${OPTARG:-}" ] && OPTARG=${OPTARG#*\=} && set -- noarg "$1" && break' _4 "eval '[ \${OPTARG+x} ] &&:' && OPTARG=$on || OPTARG=$off" valid "$1" "${v:-\$OPTARG}" _4 ';;' } _param() { args "$@" _3 "$sw)" _4 '[ $# -le 1 ] && set -- required "$1" && break' _4 'OPTARG=$2' valid "$1" '$OPTARG' _4 'shift ;;' } _option() { args "$@" quote on "$on" && quote off "$off" _3 "$sw)" _4 'set -- "$1" "$@"' _4 '[ ${OPTARG+x} ] && {' _5 'case $1 in --no-*) set -- noarg "${1%%\=*}"; break; esac' _5 '[ "${OPTARG:-}" ] && { shift; OPTARG=$2; } ||' "OPTARG=$on" _4 "} || OPTARG=$off" valid "$1" '$OPTARG' _4 'shift ;;' } valid() { set -- "$validate" "$pattern" "$1" "$2" [ "$1" ] && _4 "$1 || { set -- ${1%% *}:\$? \"\$1\" $1; break; }" [ "$2" ] && { quote pattern "$2" _4 "case \$OPTARG in $2) ;;" _5 "*) set -- pattern:$pattern \"\$1\"; break" _4 "esac" } code "$3" _4 "${export:+export }$3=\"$4\"" "${3#:}" } _disp() { args "$@" _3 "$sw)" code "$1" _4 "echo \"\${$1}\"" "${1#:}" _4 'exit 0 ;;' } _msg() { :; } _0 "$2() {" _1 'OPTIND=$(($#+1))' _1 'while OPTARG= && [ $# -gt 0 ]; do' [ "$_alt" ] && _2 'case $1 in -[!-]?*) set -- "-$@"; esac' _2 'case $1 in' wa() { _4 "eval '${1% *}' \${1+'\"\$@\"'}"; } _3 '--?*=*) OPTARG=$1; shift' wa 'set -- "${OPTARG%%\=*}" "${OPTARG#*\=}" "$@"' _4 ';;' _3 '--no-*) unset OPTARG ;;' [ "$_alt" ] || { [ "$_opts" ] && { _3 "-[$_opts]?*) OPTARG=\$1; shift" wa 'set -- "${OPTARG%"${OPTARG#??}"}" "${OPTARG#??}" "$@"' _4 ';;' } _3 '-[!-]?*) OPTARG=$1; shift' wa 'set -- "${OPTARG%"${OPTARG#??}"}" "-${OPTARG#??}" "$@"' _4 'OPTARG= ;;' } [ "$_plus" ] && { _3 '+??*) OPTARG=$1; shift' wa 'set -- "${OPTARG%"${OPTARG#??}"}" "+${OPTARG#??}" "$@"' _4 'unset OPTARG ;;' _3 '+*) unset OPTARG ;;' } _2 'esac' _2 'case $1 in' "$@" rest() { _3 "$1" _4 'while [ $# -gt 0 ]; do' _5 "$_rest=\"\${$_rest}" '\"\${$((${OPTIND:-0}-$#))}\""' _5 'shift' _4 'done' _4 'break ;;' } rest '--) shift' _3 "[-${_plus:++}]?*)" 'set -- unknown "$1" && break ;;' case $_mode in +) rest '*)' ;; *) _3 "*) $_rest=\"\${$_rest}" '\"\${$((${OPTIND:-0}-$#))}\""' esac _2 'esac' _2 'shift' _1 'done' _1 '[ $# -eq 0 ] && { OPTIND=1; unset OPTARG; return 0; }' _1 'case $1 in' _2 'unknown) set -- "Unrecognized option: $2" "$@" ;;' _2 'noarg) set -- "Does not allow an argument: $2" "$@" ;;' _2 'required) set -- "Requires an argument: $2" "$@" ;;' _2 'pattern:*) set -- "Does not match the pattern (${1#*:}): $2" "$@" ;;' _2 '*) set -- "Validation error ($1): $2" "$@"' _1 'esac' [ "$_error" ] && _1 "$_error" '"$@" >&2 || exit $?' _1 'echo "$1" >&2' _1 'exit 1' _0 '}' # This throws an error on older versions of shellcheck. # shellcheck disable=2086 [ ! "$_help" ] || eval "shift 2; getoptions_help $1 $_help" ${3+'"$@"'} } # URL: https://github.com/ko1nksm/getoptions (v2.0.1) # License: Creative Commons Zero v1.0 Universal getoptions_help() { width=30 plus='' leading=' ' pad() { p=$2; while [ ${#p} -lt "$3" ]; do p="$p "; done; eval "$1=\$p"; } args() { _type=$1 var=${2%% *} sw='' label='' hidden='' _width=$width && shift 2 while [ $# -gt 0 ] && i=$1 && shift && [ ! "$i" = '--' ]; do case $i in --*) pad sw "$sw${sw:+, }" $((${plus:+4}+4)); sw="$sw$i" ;; -?) sw="$sw${sw:+, }$i" ;; +?) [ ! "$plus" ] || { pad sw "$sw${sw:+, }" 4; sw="$sw$i"; } ;; *) eval "${i%%:*}=\${i#*:}" esac done [ "$hidden" ] && return 0 [ "$label" ] || case $_type in setup | msg) label='' _width=0 ;; flag | disp) label="$sw " ;; param) label="$sw $var " ;; option) label="${sw}[=$var] " esac pad label "${label:+$leading}$label" "$_width" [ ${#label} -le "$_width" ] && [ $# -gt 0 ] && label="$label$1" && shift echo "$label" pad label '' "$_width" for i; do echo "$label$i"; done } for i in 'setup :' flag param option disp 'msg :'; do eval "${i% *}() { args $i \"\$@\"; }" done echo "$2() {" echo "cat<<'GETOPTIONSHERE'" "$@" echo "GETOPTIONSHERE" echo "}" } global_options() { msg -- '' 'Global Options:' flag CPT_FORCE -f --force init:@export -- "Force operation" flag CPT_PROMPT -y --no-prompt on:0 off:0 init:@export -- "Do not prompt for confirmation" param CPT_ROOT --root init:@export -- "Use an alternate root directory" disp :usage -h --help -- "Show this help message" disp :version -v --version -- "Print version information" } warn() { # Print a warning message log "$1" "$2" "${3:-WARNING}" } contains() { # Check if a "string list" contains a word. case " $1 " in *" $2 "*) return 0; esac; return 1 } regesc() { # Escape special regular expression characters as # defined in POSIX BRE. '$.*[\^' printf '%s\n' "$1" | sed 's|\\|\\\\|g;s|\[|\\[|g;s|\$|\\$|g;s|\.|\\.|g;s|\*|\\*|g;s|\^|\\^|g' } prompt() { # If a CPT_NOPROMPT variable is set, continue. # This can be useful for installation scripts and # bootstrapping. [ "$CPT_PROMPT" = 0 ] && return 0 # Ask the user for some input. [ "$1" ] && log "$1" log "Continue?: Press Enter to continue or Ctrl+C to abort here" # POSIX 'read' has none of the "nice" options like '-n', '-p' # etc etc. This is the most basic usage of 'read'. # '_' is used as 'dash' errors when no variable is given to 'read'. read -r _ || return 1 } as_root() { # Simple function to run a command as root using either 'sudo', # 'doas' or 'su'. Hurrah for choice. [ "$uid" = 0 ] || log "Using '${su:-su}' (to become ${user:=root})" # We are exporting package manager variables, so that we still have the # same repository paths / access to the same cache directories etc. set -- HOME="$HOME" \ USER="$user" \ XDG_CACHE_HOME="$XDG_CACHE_HOME" \ CPT_CACHE="$CPT_CACHE" \ CPT_CHOICE="$CPT_CHOICE" \ CPT_COMPRESS="$CPT_COMPRESS" \ CPT_DEBUG="$CPT_DEBUG" \ CPT_FETCH="$CPT_FETCH" \ CPT_FORCE="$CPT_FORCE" \ CPT_HOOK="$CPT_HOOK" \ CPT_KEEPLOG="$CPT_KEEPLOG" \ CPT_PATH="$CPT_PATH" \ CPT_PID="$CPT_PID" \ CPT_PROMPT="$CPT_PROMPT" \ CPT_ROOT="$CPT_ROOT" \ CPT_CROSS_TRIPLET="$CPT_CROSS_TRIPLET" \ CPT_TMPDIR="$CPT_TMPDIR" \ "$@" case ${su##*/} in sudo|doas) "$su" -u "$user" -- env "$@" ;; su) su -c "env $* <&3" "$user" 3<&0 /dev/null || sha256 -r "$1" 2>/dev/null || openssl dgst -r -sha256 "$1" || die "No sha256 program could be run." ;} | while read -r hash _; do printf '%s %s\n' "$hash" "$1"; done } pkg_owner() { set +f [ "$3" ] || set -- "$1" "$2" "$sys_db"/*/manifest pkg_owner=$(grep "$@") pkg_owner=${pkg_owner%/*} pkg_owner=${pkg_owner##*/} set -f -- "$pkg_owner"; unset pkg_owner [ "$1" ] && printf '%s\n' "$1" } pkg_isbuilt() ( # Check if a package is built or not. read -r ver rel < "$(pkg_find "$1")/version" set +f for tarball in "$bin_dir/$1#$ver-$rel.tar."*; do [ -f "$tarball" ] && return 0 done return 1 ) pkg_lint() { # Check that each mandatory file in the package entry exists. log "$1" "Checking repository files" repo_dir=$(pkg_find "$1") cd "$repo_dir" || die "'$repo_dir' not accessible" [ -f sources ] || warn "$1" "Sources file not found" [ -x build ] || die "$1" "Build file not found or not executable" [ -s version ] || die "$1" "Version file not found or empty" read -r _ release 2>/dev/null < version || die "Version file not found" [ "$release" ] || die "Release field not found in version file" [ "$2" ] || [ -f checksums ] || die "$pkg" "Checksums are missing" } pkg_find() { # Use a SEARCH_PATH variable so that we can get the sys_db into # the same variable as CPT_PATH. This makes it easier when we are # searching for executables instead of CPT_PATH. : "${SEARCH_PATH:=$CPT_PATH:$sys_db}" # Figure out which repository a package belongs to by # searching for directories matching the package name # in $CPT_PATH/*. query=$1 match=$2 type=$3 IFS=:; set -- # Word splitting is intentional here. # shellcheck disable=2086 for path in $SEARCH_PATH ; do set +f for path2 in "$path/"$query; do test "${type:--d}" "$path2" && set -f -- "$@" "$path2" done done IFS=$old_ifs # A package may also not be found due to a repository not being # readable by the current user. Either way, we need to die here. [ "$1" ] || die "Package '$query' not in any repository" # Show all search results if called from 'cpt search', else # print only the first match. [ "$match" ] && printf '%s\n' "$@" || printf '%s\n' "$1" } pkg_list() { # List installed packages. As the format is files and # directories, this just involves a simple for loop and # file read. # Change directories to the database. This allows us to # avoid having to 'basename' each path. If this fails, # set '$1' to mimic a failed glob which indicates that # nothing is installed. cd "$sys_db" 2>/dev/null || set -- "$sys_db/"\* # Optional arguments can be passed to check for specific # packages. If no arguments are passed, list all. As we # loop over '$@', if there aren't any arguments we can # just set the directory contents to the argument list. [ "$1" ] || { set +f; set -f -- *; } # If the 'glob' above failed, exit early as there are no # packages installed. [ "$1" = "$sys_db/"\* ] && return 1 # Loop over each package and print its name and version. for pkg do [ -d "$pkg" ] || { log "$pkg" "not installed"; return 1; } read -r version 2>/dev/null < "$pkg/version" || version=null printf '%s\n' "$pkg $version" done } pkg_cache() { read -r version release 2>/dev/null < "$(pkg_find "$1")/version" # Initially assume that the package tarball is built with the CPT_COMPRESS # value. if [ -f "$bin_dir/$1#$version-$release.tar.$CPT_COMPRESS" ]; then tar_file="$bin_dir/$1#$version-$release.tar.$CPT_COMPRESS" else set +f; set -f -- "$bin_dir/$1#$version-$release.tar."* tar_file=$1 fi [ -f "$tar_file" ] } pkg_sources() { # Download any remote package sources. The existence of local # files is also checked. repo_dir=$(pkg_find "$1") # Support packages without sources. Simply do nothing. [ -f "$repo_dir/sources" ] || return 0 log "$1" "Downloading sources" # Store each downloaded source in a directory named after the # package it belongs to. This avoid conflicts between two packages # having a source of the same name. mkdir -p "$src_dir/$1" && cd "$src_dir/$1" repo_dir=$(pkg_find "$1") while read -r src dest || [ "$src" ]; do # Remote git/hg repository or comment. if [ -z "${src##\#*}" ] || [ -z "${src##git+*}" ] || [ -z "${src##hg+*}" ] then : # Remote source (cached). elif [ -f "${src##*/}" ]; then log "$1" "Found cached source '${src##*/}'" # Remote source. elif [ -z "${src##*://*}" ]; then log "$1" "Downloading $src" curl "$src" -fLo "${src##*/}" || { rm -f "${src##*/}" die "$1" "Failed to download $src" } # Local source. elif [ -f "$repo_dir/$src" ]; then log "$1" "Found local file '$src'" else die "$1" "No local file '$src'" fi done < "$repo_dir/sources" } pkg_extract() { # Extract all source archives to the build directory and copy over # any local repository files. repo_dir=$(pkg_find "$1") # Support packages without sources. Simply do nothing. [ -f "$repo_dir/sources" ] || return 0 log "$1" "Extracting sources" while read -r src dest || [ "$src" ]; do mkdir -p "$mak_dir/$1/$dest" && cd "$mak_dir/$1/$dest" case $src in # Git repository. git+*) # Split the source into URL + OBJECT (branch or commit). url=${src##git+} com=${url##*[@#]} com=${com#${url%[@#]*}} log "$1" "Cloning ${url%[@#]*}"; { git init git remote add origin "${url%[@#]*}" case "$url" in # Tags are specified via '@' *@*) git fetch -t --depth=1 origin "$com" || git fetch ;; *) git fetch --depth=1 origin "$com" || git fetch esac git checkout "${com:-FETCH_HEAD}" } || die "$1" "Failed to clone $src" ;; # Mercurial repository. hg+*) # Split the source into URL + OBJECT (branch or commit). url=${src##hg+} com=${url##*[@#]} com=${com#${url%[@#]*}} # Unfortunately, there is no shallow cloning with Mercurial. log "$1" "Cloning ${url%[@#]*}" hg clone -u "${com:-tip}" ;; # Comment or blank line. \#*|'') continue ;; # Only 'tar', 'cpio', and 'zip' archives are currently supported for # extraction. Other filetypes are simply copied to '$mak_dir' # which allows for manual extraction. *://*.tar|*://*.tar.??|*://*.tar.???|*://*.tar.????|*://*.tgz|*://*.txz) decompress "$src_dir/$1/${src##*/}" > .ktar "$tar" xf .ktar || die "$1" "Couldn't extract ${src##*/}" # We now list the contents of the tarball so we can do our # version of 'strip-components'. "$tar" tf .ktar | while read -r file; do printf '%s\n' "${file%%/*}"; done | # Do not repeat files. uniq | # For every directory in the base we move each file # inside it to the upper directory. while read -r dir ; do # Skip if we are not dealing with a directory here. # This way we don't remove files on the upper directory # if a tar archive doesn't need directory stripping. [ -d "${dir#.}" ] || continue # Change into the directory in a subshell so we don't # need to cd back to the upper directory. ( cd "$dir" # We use find because we want to move hidden files # as well. # # Skip the file if it has the same name as the directory. # We will deal with it later. # # Word splitting is intentional here. # shellcheck disable=2046 find . \( ! -name . -prune \) ! -name "$dir" \ -exec mv -f {} .. \; # If a file/directory with the same name as the directory # exists, append a '.cptbak' to it and move it to the # upper directory. ! [ -e "$dir" ] || mv "$dir" "../${dir}.cptbak" ) rmdir "$dir" # If a backup file exists, move it into the original location. ! [ -e "${dir}.cptbak" ] || mv "${dir}.cptbak" "$dir" done # Clean up the temporary tarball. rm -f .ktar ;; *://*.cpio|*://*.cpio.??|*://*.cpio.???|*://*.cpio.????) decompress "$src_dir/$1/${src##*/}" | cpio -i ;; *://*.zip) unzip "$src_dir/$1/${src##*/}" || die "$1" "Couldn't extract ${src##*/}" ;; *) # Local file. if [ -f "$repo_dir/$src" ]; then cp -f "$repo_dir/$src" . # Remote file. elif [ -f "$src_dir/$1/${src##*/}" ]; then cp -f "$src_dir/$1/${src##*/}" . else die "$1" "Local file $src not found" fi ;; esac done < "$repo_dir/sources" } pkg_depends() { # Resolve all dependencies and generate an ordered list. # This does a depth-first search. The deepest dependencies are # listed first and then the parents in reverse order. contains "$deps" "$1" || { # Filter out non-explicit, aleady installed dependencies. # Only filter installed if called from 'pkg_build()'. [ "$pkg_build" ] && [ -z "$2" ] && (pkg_list "$1" >/dev/null) && return while read -r dep type || [ "$dep" ]; do # Skip test dependencies unless $CPT_TEST is set to 1. case $type in test) [ "$CPT_TEST" = 1 ] || continue; esac # Recurse through the dependencies of the child packages. [ "${dep##\#*}" ] && pkg_depends "$dep" done 2>/dev/null < "$(pkg_find "$1")/depends" ||: # After child dependencies are added to the list, # add the package which depends on them. [ "$2" = explicit ] || deps="$deps $1 " } } pkg_order() { # Order a list of packages based on dependence and # take into account pre-built tarballs if this is # to be called from 'cpt i'. order=; redro=; deps= for pkg do case $pkg in *.tar.*) deps="$deps $pkg " ;; *) pkg_depends "$pkg" raw esac done # Filter the list, only keeping explicit packages. # The purpose of these two loops is to order the # argument list based on dependence. for pkg in $deps; do ! contains "$*" "$pkg" || { order="$order $pkg " redro=" $pkg $redro" } done deps= } pkg_strip() { # Strip package binaries and libraries. This saves space on the # system as well as on the tarballs we ship for installation. # Package has stripping disabled, stop here. [ -f "$mak_dir/$pkg/nostrip" ] && return log "$1" "Stripping binaries and libraries" find "$pkg_dir/$1" -type f | while read -r file; do case $(od -A o -t c -N 18 "$file") in # REL (object files (.o), static libraries (.a)). *177*E*L*F*0000020\ 001\ *|*\!*\<*a*r*c*h*\>*) strip -g -R .comment -R .note "$file" ;; # EXEC (static binaries). # DYN (shared libraries, dynamic binaries). # Shared libraries keep global symbols in a separate ELF section # called '.dynsym'. '--strip-all/-s' does not touch the dynamic # symbol entries which makes this safe to do. *177*E*L*F*0000020\ 00[23]\ *) strip -s -R .comment -R .note "$file" ;; esac done 2>/dev/null ||: } pkg_fix_deps() { # Dynamically look for missing runtime dependencies by checking each binary # and library with either 'ldd' or 'readelf'. This catches any extra # libraries and or dependencies pulled in by the package's build suite. log "$1" "Checking for missing dependencies" # Go to the directory containing the built package to # simplify path building. cd "$pkg_dir/$1/$pkg_db/$1" # Make a copy of the depends file if it exists to have a # reference to 'diff' against. if [ -f depends ]; then cp -f depends "$mak_dir/d" dep_file=$mak_dir/d else dep_file=/dev/null fi # Generate a list of all installed manifests. pkg_name=$1 set +f; set -f -- "$sys_db/"*/manifest # Get a list of binaries and libraries, false files # will be found, however it's faster to get 'ldd' to check # them anyway than to filter them out. find "$pkg_dir/$pkg_name/" -type f 2>/dev/null | while read -r file; do case ${elf_prog:-ldd} in *readelf) "$elf_prog" -d "$file" 2>/dev/null ;; *) ldd "$file" 2>/dev/null ;; esac | while read -r dep; do # Skip lines containing 'ldd'. [ "${dep##*ldd*}" ] || continue case $dep in *NEEDED*\[*\] | *'=>'*) ;; *) continue; esac # readelf output: # 0x0000 (NEEDED) Shared library: [libc.so] dep=${dep##*\[} dep=${dep%%\]*} # ldd output: # libc.so => /lib/ld-musl-x86_64.so.1 dep=${dep#* => } dep=${dep% *} # Figure out which package owns the file. Skip file if it is owned # by the current package. This also handles cases where a '*-bin' # package exists on the system, so the package manager doesn't think # that the package we are building depends on the *-bin version of # itself, or any other renamed versions of the same software. pkg_owner -l "/${dep#/}\$" "$PWD/manifest" >/dev/null && continue pkg_owner -l "/${dep#/}\$" "$@" ||: done ||: done >> depends # Remove duplicate entries from the new depends file. # This removes duplicate lines looking *only* at the # first column. sort -uk1,1 -o depends depends 2>/dev/null ||: # Display a diff of the new dependencies against the old ones. diff -U 3 "$dep_file" depends 2>/dev/null ||: # Remove the depends file if it is empty. [ -s depends ] || rm -f depends } pkg_manifest() ( # Generate the package's manifest file. This is a list of each file # and directory inside the package. The file is used when uninstalling # packages, checking for package conflicts and for general debugging. log "$1" "Generating manifest" # This function runs as a sub-shell to avoid having to 'cd' back to the # prior directory before being able to continue. cd "${2:-$pkg_dir}/$1" # find: Print all files and directories and append '/' to directories. # sort: Sort the output in *reverse*. Directories appear *after* their # contents. # sed: Remove the first character in each line (./dir -> /dir) and # remove all lines which only contain '.'. find . -type d -exec printf '%s/\n' {} + -o -print | sort -r | sed '/^\.\/$/d;ss.ss' > "${2:-$pkg_dir}/$1/$pkg_db/$1/manifest" ) pkg_etcsums() ( # This function runs as a sub-shell to avoid having to 'cd' back to the # prior directory before being able to continue. cd "$pkg_dir/$1/etc" 2>/dev/null || return 0; cd .. # Generate checksums for each configuration file in the package's # /etc/ directory for use in "smart" handling of these files. log "$1" "Generating etcsums" find etc -type f | while read -r file; do sh256 "$file" done > "$pkg_dir/$1/$pkg_db/$1/etcsums" ) pkg_tar() { # Create a tarball from the built package's files. # This tarball also contains the package's database entry. log "$1" "Creating tarball" # Read the version information to name the package. read -r version release < "$(pkg_find "$1")/version" # Create a tarball from the contents of the built package. "$tar" cf - -C "$pkg_dir/$1" . | case $CPT_COMPRESS in bz2) bzip2 -z ;; xz) xz -zT 0 ;; gz) gzip -6 ;; zst) zstd -3 ;; *) gzip -6 ;; # Fallback to gzip esac \ > "$bin_dir/$1#$version-$release.tar.$CPT_COMPRESS" log "$1" "Successfully created tarball" run_hook post-package "$1" "$bin_dir/$1#$version-$release.tar.$CPT_COMPRESS" } pkg_build() { # Build packages and turn them into packaged tarballs. This function # also checks checksums, downloads sources and ensure all dependencies # are installed. pkg_build=1 log "Resolving dependencies" for pkg do contains "$explicit" "$pkg" || { pkg_depends "$pkg" explicit # Mark packages passed on the command-line # separately from those detected as dependencies. explicit="$explicit $pkg " } done [ "$pkg_update" ] || explicit_build=$explicit # If an explicit package is a dependency of another explicit # package, remove it from the explicit list as it needs to be # installed as a dependency. # shellcheck disable=2086 for pkg do contains "$deps" "$pkg" && explicit=$(pop "$pkg" from $explicit) done # See [1] at top of script. # shellcheck disable=2046,2086 set -- $deps $explicit log "Building: $*" # Only ask for confirmation if more than one package needs to be built. [ $# -gt 1 ] || [ "$pkg_update" ] && { prompt || exit 0 ;} log "Checking for pre-built dependencies" for pkg do pkg_lint "$pkg"; done # Install any pre-built dependencies if they exist in the binary # directory and are up to date. for pkg do ! contains "$explicit_build" "$pkg" && pkg_cache "$pkg" && { log "$pkg" "Found pre-built binary, installing" (CPT_FORCE=1 cpt-install "$tar_file") # Remove the now installed package from the build list. # See [1] at top of script. # shellcheck disable=2046,2086 set -- $(pop "$pkg" from "$@") } done for pkg do pkg_sources "$pkg"; done pkg_verify "$@" # Finally build and create tarballs for all passed packages and # dependencies. for pkg do log "$pkg" "Building package ($((in = in + 1))/$#)" pkg_extract "$pkg" repo_dir=$(pkg_find "$pkg") read -r build_version _ < "$repo_dir/version" # Copy the build file to the build directory to users to modify it # temporarily at runtime. cp -f "$repo_dir/build" "$mak_dir/$pkg/.build.cpt" # Install built packages to a directory under the package name # to avoid collisions with other packages. mkdir -p "$pkg_dir/$pkg/$pkg_db" # Move to the build directory. cd "$mak_dir/$pkg" log "$pkg" "Starting build" run_hook pre-build "$pkg" "$pkg_dir/$pkg" # Notify the user if the build script is changed during the pre-build # hook. diff -q "$repo_dir/build" .build.cpt || log "$pkg" "Executing the modified build file" # Call the build script, log the output to the terminal # and to a file. There's no PIPEFAIL in POSIX shelll so # we must resort to tricks like killing the script ourselves. { ./.build.cpt "$pkg_dir/$pkg" "$build_version" "$sys_arch" 2>&1 || { log "$pkg" "Build failed" log "$pkg" "Log stored to $log_dir/$pkg-$time-$pid" run_hook build-fail "$pkg" "$pkg_dir/$pkg" pkg_clean kill 0 } } | tee "$log_dir/$pkg-$time-$pid" # Run the test script if it exists and the user wants to run tests. This # is turned off by default. [ -x "$repo_dir/test" ] && [ "$CPT_TEST" = 1 ] && { run_hook pre-test "$pkg" "$pkg_dir/$pkg" log "$pkg" "Running tests" "$repo_dir/test" "$pkg_dir/$pkg" "$build_version" "$sys_arch" 2>&1 || { log "$pkg" "Test failed" log "$pkg" "Log stored to $log_dir/$pkg-$time-$pid" run_hook test-fail "$pkg" "$pkg_dir/$pkg" pkg_clean kill 0 } } | tee -a "$log_dir/$pkg-$time-$pid" # Delete the log file if the build succeeded to prevent # the directory from filling very quickly with useless logs. [ "$CPT_KEEPLOG" = 1 ] || rm -f "$log_dir/$pkg-$time-$pid" # Copy the repository files to the package directory. # This acts as the database entry. cp -LRf "$repo_dir" "$pkg_dir/$pkg/$pkg_db/" # Copy the modified build file to the package directory. pkg_build="$pkg_dir/$pkg/$pkg_db/$pkg/build" diff -U 3 "$pkg_build" .build.cpt > "$pkg_build.diff" && rm -f "$pkg_build.diff" # We don't want the package manager to track 'dir' pages of the info # directory. We don't want every single package to create their own dir # files either. rm -f "$pkg_dir/$pkg/usr/share/info/dir" # We never ever want this. Let's end the endless conflicts # and remove it. find "$pkg_dir/$pkg" -name charset.alias -exec rm -f {} + # Remove libtool's '*.la' library files. This removes cross-build # system conflicts that may arise. Build-systems change, libtool # is getting deprecated, we don't want a package that depends on # some package's '.la' files. find "$pkg_dir/$pkg" -name '*.la' -exec rm -f {} + log "$pkg" "Successfully built package" run_hook post-build "$pkg" "$pkg_dir/$pkg" # Create the manifest file early and make it empty. # This ensures that the manifest is added to the manifest. : > "$pkg_dir/$pkg/$pkg_db/$pkg/manifest" # If the package contains '/etc', add a file called # 'etcsums' to the manifest. See comment directly above. [ -d "$pkg_dir/$pkg/etc" ] && : > "$pkg_dir/$pkg/$pkg_db/$pkg/etcsums" pkg_strip "$pkg" pkg_manifest "$pkg" pkg_fix_deps "$pkg" pkg_manifest "$pkg" pkg_etcsums "$pkg" pkg_tar "$pkg" # Install only dependencies of passed packages. # Skip this check if this is a package update. contains "$explicit" "$pkg" && [ -z "$pkg_update" ] && continue log "$pkg" "Needed as a dependency or has an update, installing" (CPT_FORCE=1 cpt-install "$pkg") done # End here as this was a system update and all packages have been installed. [ "$pkg_update" ] && return log "Successfully built package(s)" # Turn the explicit packages into a 'list'. # See [1] at top of script. # shellcheck disable=2046,2086 set -- $explicit # Only ask for confirmation if more than one package needs to be installed. [ $# -gt 1 ] && prompt "Install built packages? [$*]" && { cpt-install "$@" return } log "Run 'cpt i $*' to install the package(s)" } pkg_checksums() { # Generate checksums for packages. repo_dir=$(pkg_find "$1") [ -f "$repo_dir/sources" ] || return 0 while read -r src _ || [ "$src" ]; do # Comment. if [ -z "${src##\#*}" ]; then continue # File is local to the package. elif [ -f "$repo_dir/$src" ]; then src_path=$repo_dir/${src%/*} # File is remote and was downloaded. elif [ -f "$src_dir/$1/${src##*/}" ]; then src_path=$src_dir/$1 # File is a git repository. elif [ -z "${src##git+*}" ]; then continue # Die here if source for some reason, doesn't exist. else die "$1" "Couldn't find source '$src'" fi # An easy way to get 'sha256sum' to print with the 'basename' # of files is to 'cd' to the file's directory beforehand. (cd "$src_path" && sh256 "${src##*/}") || die "$1" "Failed to generate checksums" done < "$repo_dir/sources" } pkg_verify() { # Verify all package checksums. This is achieved by generating a new set of # checksums and then comparing those with the old set. verify_cmd="NR==FNR{a[\$1];next}/^git .*/{next}!((\$1)in a){exit 1}" for pkg; do repo_dir=$(pkg_find "$pkg") [ -f "$repo_dir/sources" ] || continue pkg_checksums "$pkg" | awk "$verify_cmd" - "$repo_dir/checksums" || { log "$pkg" "Checksum mismatch" # Instead of dying above, log it to the terminal. Also define a # variable so we *can* die after all checksum files have been # checked. mismatch="$mismatch$pkg " } done [ -z "$mismatch" ] || die "Checksum mismatch with: ${mismatch% }" } pkg_conflicts() { # Check to see if a package conflicts with another. log "$1" "Checking for package conflicts" # Filter the tarball's manifest and select only files # and any files they resolve to on the filesystem # (/bin/ls -> /usr/bin/ls). while read -r file; do case $file in */) continue; esac # Use $CPT_ROOT in filename so that we follow its symlinks. file=$CPT_ROOT/${file#/} # We will only follow the symlinks of the directories, so we reserve the # directory name in this 'dirname' value. If we cannot find it in the # system, we don't need to make this much more complex by trying so hard # to find it. Simply use the original directory name. dirname="$(_readlinkf "${file%/*}" 2>/dev/null)" || dirname="${file%/*}" # Combine the dirname and file values, and print them into the # temporary manifest to be parsed. printf '%s/%s\n' "${dirname#$CPT_ROOT}" "${file##*/}" done < "$tar_dir/$1/$pkg_db/$1/manifest" > "$CPT_TMPDIR/$pid/manifest" p_name=$1 # Generate a list of all installed package manifests # and remove the current package from the list. # shellcheck disable=2046,2086 set -- $(set +f; pop "$sys_db/$p_name/manifest" from "$sys_db"/*/manifest) [ -s "$CPT_TMPDIR/$pid/manifest" ] || return 0 # In rare cases where the system only has one package installed # and you are reinstalling that package, grep will try to read from # standard input if we continue here. # # Also, if we don't have any packages installed grep will give an # error. This will not cause the installation to fail, but we don't # need to check for conflicts if that's the case anyway. If we have # only zero packages or one package, just stop wasting time and continue # with the installation. [ "$1" ] && [ -f "$1" ] || return 0 # Store the list of found conflicts in a file as we will be using the # information multiple times. Storing it in the cache dir allows us # to be lazy as they'll be automatically removed on script end. "$grep" -Fxf "$CPT_TMPDIR/$pid/manifest" -- "$@" > "$CPT_TMPDIR/$pid/conflict" ||: # Enable alternatives automatically if it is safe to do so. # This checks to see that the package that is about to be installed # doesn't overwrite anything it shouldn't in '/var/db/cpt/installed'. "$grep" -q ":/var/db/cpt/installed/" "$CPT_TMPDIR/$pid/conflict" || choice_auto=1 # Use 'grep' to list matching lines between the to # be installed package's manifest and the above filtered # list. if [ "$CPT_CHOICE" != 0 ] && [ "$choice_auto" = 1 ]; then # This is a novel way of offering an "alternatives" system. # It is entirely dynamic and all "choices" are created and # destroyed on the fly. # # When a conflict is found between two packages, the file # is moved to a directory called "choices" and its name # changed to store its parent package and its intended # location. # # The package's manifest is then updated to reflect this # new location. # # The 'cpt choices' command parses this directory and # offers you the CHOICE of *swapping* entries in this # directory for those on the filesystem. # # The choices command does the same thing we do here, # it rewrites manifests and moves files around to make # this work. # # Pretty nifty huh? while IFS=: read -r _ con; do printf '%s\n' "Found conflict $con" # Create the "choices" directory inside of the tarball. # This directory will store the conflicting file. mkdir -p "$tar_dir/$p_name/${cho_dir:=var/db/cpt/choices}" # Construct the file name of the "db" entry of the # conflicting file. (pkg_name>usr>bin>ls) con_name=$(printf %s "$con" | sed 's|/|>|g') # Move the conflicting file to the choices directory # and name it according to the format above. mv -f "$tar_dir/$p_name/$con" \ "$tar_dir/$p_name/$cho_dir/$p_name$con_name" 2>/dev/null || { log "File must be in ${con%/*} and not a symlink to it" log "This usually occurs when a binary is installed to" log "/sbin instead of /usr/bin (example)" log "Before this package can be used as an alternative," log "this must be fixed in $p_name. Contact the maintainer" die "by checking 'git log' or by running 'cpt-maintainer'" } done < "$CPT_TMPDIR/$pid/conflict" # Rewrite the package's manifest to update its location # to its new spot (and name) in the choices directory. pkg_manifest "$p_name" "$tar_dir" 2>/dev/null elif [ -s "$CPT_TMPDIR/$pid/conflict" ]; then log "Package '$p_name' conflicts with another package" "" "!>" log "Run 'CPT_CHOICE=1 cpt i $p_name' to add conflicts" "" "!>" die "as alternatives." fi } pkg_swap() { # Swap between package alternatives. pkg_list "$1" >/dev/null alt=$(printf %s "$1$2" | sed 's|/|>|g') cd "$sys_db/../choices" [ -f "$alt" ] || [ -h "$alt" ] || die "Alternative '$1 $2' doesn't exist" if [ -f "$2" ]; then # Figure out which package owns the file we are going to swap for # another package's. # # Print the full path to the manifest file which contains # the match to our search. pkg_owns=$(pkg_owner -lFx "$2") || die "File '$2' exists on filesystem but isn't owned" log "Swapping '$2' from '$pkg_owns' to '$1'" # Convert the current owner to an alternative and rewrite # its manifest file to reflect this. We then resort this file # so no issues arise when removing packages. cp -Pf "$CPT_ROOT/$2" "$pkg_owns>${alt#*>}" sed "s#^$(regesc "$2")#${PWD#$CPT_ROOT}/$pkg_owns>${alt#*>}#" \ "../installed/$pkg_owns/manifest" | sort -r -o "../installed/$pkg_owns/manifest" fi # Convert the desired alternative to a real file and rewrite # the manifest file to reflect this. The reverse of above. mv -f "$alt" "$CPT_ROOT/$2" sed "s#^${PWD#$CPT_ROOT}/$(regesc "$alt")#$2#" "../installed/$1/manifest" | sort -r -o "../installed/$1/manifest" } pkg_etc() { [ -d "$tar_dir/$pkg_name/etc" ] || return 0 (cd "$tar_dir/$pkg_name" # Create all directories beforehand. find etc -type d | while read -r dir; do mkdir -p "$CPT_ROOT/$dir" done # Handle files in /etc/ based on a 3-way checksum check. find etc ! -type d | while read -r file; do { sum_new=$(sh256 "$file") sum_sys=$(cd "$CPT_ROOT/"; sh256 "$file") sum_old=$("$grep" "$file$" "$mak_dir/c"); } 2>/dev/null ||: log "$pkg_name" "Doing 3-way handshake for $file" printf '%s\n' "Previous: ${sum_old:-null}" printf '%s\n' "System: ${sum_sys:-null}" printf '%s\n' "New: ${sum_new:-null}" # Use a case statement to easily compare three strings at # the same time. Pretty nifty. case ${sum_old:-null}${sum_sys:-null}${sum_new} in # old = Y, sys = X, new = Y "${sum_new}${sum_sys}${sum_old}") log "Skipping $file" continue ;; # old = X, sys = X, new = X # old = X, sys = Y, new = Y # old = X, sys = X, new = Y "${sum_old}${sum_old}${sum_old}"|\ "${sum_old:-null}${sum_sys}${sum_sys}"|\ "${sum_sys}${sum_old}"*) log "Installing $file" new= ;; # All other cases. *) warn "$pkg_name" "saving /$file as /$file.new" "->" new=.new ;; esac cp -fPp "$file" "$CPT_ROOT/${file}${new}" chown root:root "$CPT_ROOT/${file}${new}" 2>/dev/null done) ||: } pkg_remove() { # Remove a package and all of its files. The '/etc' directory # is handled differently and configuration files are *not* # overwritten. pkg_list "$1" >/dev/null || return # Make sure that nothing depends on this package. [ "$CPT_FORCE" = 1 ] || { log "$1" "Checking for reverse dependencies" (cd "$sys_db"; set +f; grep -lFx "$1" -- */depends) && die "$1" "Can't remove package, others depend on it" } # Block being able to abort the script with 'Ctrl+C' during removal. # Removes all risk of the user aborting a package removal leaving # an incomplete package installed. trap_set block if [ -x "$sys_db/$1/pre-remove" ]; then log "$1" "Running pre-remove script" "$sys_db/$1/pre-remove" ||: fi # Create a temporary list of all directories, so we don't accidentally # remove anything from packages that create empty directories for a # purpose (such as baselayout). manifest_list="$(set +f; pop "$sys_db/$1/manifest" from "$sys_db/"*/manifest)" # shellcheck disable=2086 [ "$manifest_list" ] && grep -h '/$' $manifest_list | sort -ur > "$mak_dir/dirs" run_hook pre-remove "$1" "$sys_db/$1" root while read -r file; do # The file is in '/etc' skip it. This prevents the package # manager from removing user edited configuration files. [ "${file##/etc/*}" ] || continue if [ -d "$CPT_ROOT/$file" ]; then "$grep" -Fxq "$file" "$mak_dir/dirs" 2>/dev/null && continue rmdir "$CPT_ROOT/$file" 2>/dev/null || continue else rm -f "$CPT_ROOT/$file" fi done < "$sys_db/$1/manifest" # Reset 'trap' to its original value. Removal is done so # we no longer need to block 'Ctrl+C'. trap_set cleanup run_hook post-remove "$1" "$CPT_ROOT/" root log "$1" "Removed successfully" } pkg_install() { # Install a built package tarball. # Install can also take the full path to a tarball. # We don't need to check the repository if this is the case. if [ -f "$1" ] && [ -z "${1%%*.tar*}" ] ; then tar_file=$1 pkg_name=${1##*/} pkg_name=${pkg_name%#*} else pkg_cache "$1" || die "package has not been built, run 'cpt b pkg'" pkg_name=$1 fi mkdir -p "$tar_dir/$pkg_name" log "$pkg_name" "Extracting $tar_file" # Extract the tarball to catch any errors before installation begins. decompress "$tar_file" | "$tar" xf - -C "$tar_dir/$pkg_name" [ -f "$tar_dir/$pkg_name/$pkg_db/$pkg_name/manifest" ] || die "'${tar_file##*/}' is not a valid CPT package" # Ensure that the tarball's manifest is correct by checking that # each file and directory inside of it actually exists. [ "$CPT_FORCE" != 1 ] && log "$pkg_name" "Checking package manifest" && while read -r line; do # Skip symbolic links [ -h "$tar_dir/$pkg_name/$line" ] || [ -e "$tar_dir/$pkg_name/$line" ] || { log "File $line missing from tarball but mentioned in manifest" "" "!>" TARBALL_FAIL=1 } done < "$tar_dir/$pkg_name/$pkg_db/$pkg_name/manifest" [ "$TARBALL_FAIL" ] && { log "You can still install this package by setting CPT_FORCE variable" die "$pkg_name" "Missing files in manifest" } log "$pkg_name" "Checking that all dependencies are installed" # Make sure that all run-time dependencies are installed prior to # installing the package. [ -f "$tar_dir/$pkg_name/$pkg_db/$pkg_name/depends" ] && [ "$CPT_FORCE" != 1 ] && while read -r dep dep_type || [ "$dep" ]; do [ "${dep##\#*}" ] || continue [ "$dep_type" ] || pkg_list "$dep" >/dev/null || install_dep="$install_dep'$dep', " done < "$tar_dir/$pkg_name/$pkg_db/$pkg_name/depends" [ "$install_dep" ] && die "$1" "Package requires ${install_dep%, }" run_hook pre-install "$pkg_name" "$tar_dir/$pkg_name" root pkg_conflicts "$pkg_name" log "$pkg_name" "Installing package incrementally" # Block being able to abort the script with Ctrl+C during installation. # Removes all risk of the user aborting a package installation leaving # an incomplete package installed. trap_set block # If the package is already installed (and this is an upgrade) make a # backup of the manifest and etcsums files. cp -f "$sys_db/$pkg_name/manifest" "$mak_dir/m" 2>/dev/null ||: cp -f "$sys_db/$pkg_name/etcsums" "$mak_dir/c" 2>/dev/null ||: # This is repeated multiple times. Better to make it a function. pkg_rsync() { rsync "--chown=$USER:$USER" --chmod=Du-s,Dg-s,Do-s \ -WhHKa --no-compress --exclude /etc "${1:---}" \ "$tar_dir/$pkg_name/" "$CPT_ROOT/" } # Install the package by using 'rsync' and overwrite any existing files # (excluding '/etc/'). pkg_rsync --info=progress2 pkg_etc # Remove any leftover files if this is an upgrade. "$grep" -vFxf "$sys_db/$pkg_name/manifest" "$mak_dir/m" 2>/dev/null | while read -r file; do file=$CPT_ROOT/$file # Skip deleting some leftover files. case $file in /etc/*) continue; esac # Remove files. if [ -f "$file" ] && [ ! -L "$file" ]; then rm -f "$file" # Remove file symlinks. elif [ -h "$file" ] && [ ! -d "$file" ]; then unlink "$file" ||: # Skip directory symlinks. elif [ -h "$file" ] && [ -d "$file" ]; then : # Remove directories if empty. elif [ -d "$file" ]; then rmdir "$file" 2>/dev/null ||: fi done ||: log "$pkg_name" "Verifying installation" { pkg_rsync; pkg_rsync; } ||: # Reset 'trap' to its original value. Installation is done so # we no longer need to block 'Ctrl+C'. trap_set cleanup if [ -x "$sys_db/$pkg_name/post-install" ]; then log "$pkg_name" "Running post-install script" "$sys_db/$pkg_name/post-install" ||: fi run_hook post-install "$pkg_name" "$sys_db/$pkg_name" root log "$pkg_name" "Installed successfully" } pkg_fetch() { log "Updating repositories" run_hook pre-fetch # Create a list of all repositories. # See [1] at top of script. # shellcheck disable=2046,2086 { IFS=:; set -- $CPT_PATH; IFS=$old_ifs ;} # Update each repository in '$CPT_PATH'. It is assumed that # each repository is 'git' tracked. for repo; do # Go to the root of the repository (if it exists). cd "$repo" cd "$(git rev-parse --show-toplevel 2>/dev/null)" 2>/dev/null || cd "$(hg root 2>/dev/null)" 2>/dev/null ||: if [ -d .git ]; then [ "$(git remote 2>/dev/null)" ] || { log "$repo" " " printf '%s\n' "No remote, skipping." continue } contains "$repos" "$PWD" || { repos="$repos $PWD " # Display a tick if signing is enabled for this # repository. case $(git config merge.verifySignatures) in true) log "$PWD" "[signed] " ;; *) log "$PWD" " " ;; esac if [ -w "$PWD" ] && [ "$uid" != 0 ]; then git fetch git merge git submodule update --remote --init -f else [ "$uid" = 0 ] || log "$PWD" "Need root to update" # Find out the owner of the repository and spawn # git as this user below. # # This prevents 'git' from changing the original # ownership of files and directories in the rare # case that the repository is owned by a 3rd user. ( user=$(_stat "$PWD") [ "$user" = root ] || log "Dropping permissions to $user for pull" git_cmd="git fetch && git merge && git submodule update --remote --init -f" case $su in *su) git_cmd="'$git_cmd'"; esac # Spawn a subshell to run multiple commands as # root at once. This makes things easier on users # who aren't using persist/timestamps for auth # caching. user=$user as_root sh -c "$git_cmd" ) fi } elif [ -d .hg ]; then [ "$(hg showconfig paths 2>/dev/null)" ] || { log "$repo" " " printf '%s\n' "No remote, skipping." continue } contains "$repos $PWD" || { repos="$repos $PWD" if [ -w "$PWD" ] && [ "$uid" != 0 ]; then hg pull hg update else [ "$uid" ] || log "$PWD" "Need root to update" # We are going to do the same operation as above, to # find the owner of the repository. ( user=$(_stat "$PWD") [ "$user" = root ] || log "Dropping permissions to $user for pull" hg_cmd="hg pull && hg update" case $su in *su) hg_cmd="'$hg_cmd'"; esac user=$user as_root sh -c "$hg_cmd" ) fi } elif [ -f .rsync ]; then # If an .rsync_root file exists, we check that the repository root # exists. If it does, we change to that directory to do the fetch. # This way, we allow for partial repositories while making sure that # we can fetch the repository in a single operation. [ -f .rsync_root ] && { read -r rsync_root < .rsync_root [ -f "$rsync_root/.rsync" ] && cd "$rsync_root" } contains "$repos" "$PWD" || { repos="$repos $PWD" read -r remote < .rsync if [ -w "$PWD" ] && [ "$uid" != 0 ]; then rsync -acvzzC --include=core --delete "$remote/" "$PWD" else [ "$uid" = 0 ] || log "$PWD" "Need root to update" # Similar to the git update, we find the owner of # the repository and spawn rsync as that user. ( user=$(_stat "$PWD") [ "$user" = root ] || log "Dropping permissions to $user for pull" user=$user as_root rsync -acvzzC --include=core --delete "$remote/" "$PWD" ) fi } else log "$repo" " " printf '%s\n' "Not a remote repository, skipping." fi done run_hook post-fetch } pkg_updates(){ # Check all installed packages for updates. So long as the installed # version and the version in the repositories differ, it's considered # an update. [ "$CPT_FETCH" = 0 ] || pkg_fetch log "Checking for new package versions" set +f for pkg in "$sys_db/"*; do pkg_name=${pkg##*/} # Read version and release information from the installed packages # and repository. read -r db_ver db_rel < "$pkg/version" read -r re_ver re_rel < "$(pkg_find "$pkg_name")/version" # Compare installed packages to repository packages. [ "$db_ver-$db_rel" != "$re_ver-$re_rel" ] && { printf '%s\n' "$pkg_name $db_ver-$db_rel ==> $re_ver-$re_rel" outdated="$outdated$pkg_name " } done set -f # If the download option is specified only download the outdated packages # and exit. # shellcheck disable=2154 [ "$download_only" = 1 ] && { log "Only sources for the packages will be acquired" prompt || exit 0 for pkg in $outdated; do pkg_sources "$pkg" done exit 0 } contains "$outdated" cpt && { log "Detected package manager update" log "The package manager will be updated first" prompt || exit 0 pkg_build cpt cpt-install cpt log "Updated the package manager" log "Re-run 'cpt update' to update your system" exit 0 } [ "$outdated" ] || { log "Everything is up to date" return } log "Packages to update: ${outdated% }" # Tell 'pkg_build' to always prompt before build. pkg_update=1 # Build all packages requiring an update. # See [1] at top of script. # shellcheck disable=2046,2086 { pkg_order $outdated pkg_build $order } log "Updated all packages" } pkg_clean() { # Clean up on exit or error. This removes everything related # to the build. [ "$CPT_DEBUG" != 1 ] || return 0 # Block 'Ctrl+C' while cache is being cleaned. trap_set block # Remove temporary items. rm -rf -- "${CPT_TMPDIR:=$cac_dir/proc}/$pid" } create_cache() { # A temporary directory can be specified apart from the cache # directory in order to build in a user specified directory. # /tmp could be used in order to build on ram, useful on SSDs. # The user can specify CPT_TMPDIR for this. # # Create the required temporary directories and set the variables # which point to them. mkdir -p "${tmp_dir:=${CPT_TMPDIR:=$cac_dir/proc}/$pid}" # If an argument is given, skip the creation of other cache directories. # This here makes shellcheck extremely angry, so I am globally disabling # SC2119. [ "$1" ] || mkdir -p "${mak_dir:=$tmp_dir/build}" \ "${pkg_dir:=$tmp_dir/pkg}" \ "${tar_dir:=$tmp_dir/export}" } # main() { set -ef # If a parser definition exists, let's run it ourselves. This makes sure we # get the variables as soon as possible. command -v parser_definition >/dev/null && { eval "$(getoptions parser_definition parse "$0")" parse "$@" eval set -- "$REST" } # Create the cache directories for CPT and set the variables which point # to them. This is seperate from temporary directories created in # create_cache(). That's because we need these variables set on most # occasions. mkdir -p "${cac_dir:=${CPT_CACHE:=${XDG_CACHE_HOME:-$HOME/.cache}/cpt}}" \ "${src_dir:=$cac_dir/sources}" \ "${log_dir:=$cac_dir/logs}" \ "${bin_dir:=$cac_dir/$CPT_CROSS_TRIPLET${CPT_CROSS_TRIPLET:+-}bin}" # Set the location to the repository and package database. pkg_db=var/db/cpt/installed # The PID of the current shell process is used to isolate directories # to each specific CPT instance. This allows multiple package manager # instances to be run at once. Store the value in another variable so # that it doesn't change beneath us. pid=${CPT_PID:-$$} # Force the C locale to speed up things like 'grep' which disable unicode # etc when this is set. We don't need unicode and a speed up is always # welcome. export LC_ALL=C LANG=C # Catch errors and ensure that build files and directories are cleaned # up before we die. This occurs on 'Ctrl+C' as well as success and error. trap_set cleanup # Prefer GNU grep if installed as it is much much faster than busybox's # implementation. Very much worth it if you value performance over # POSIX correctness (grep quoted to avoid shellcheck false-positive). grep=$(command -v ggrep) || grep='grep' # Prefer libarchive tar or GNU tar if installed as they are much # much faster than busybox's implementation. Very much worth it if # you value performance. tar=$(command -v bsdtar || command -v gtar) || tar=tar # Figure out which 'sudo' command to use based on the user's choice or # what is available on the system. su=${CPT_SU:-$(command -v sudo || command -v doas)} || su=su # Store the date and time of script invocation to be used as the name # of the log files the package manager creates uring builds. time=$(date '+%Y-%m-%d-%H:%M') # Use readelf for fixing dependencies if it is available, fallback to # ldd. readelf shows only the actual dependencies and doesn't include # the libraries required by the dependencies. elf_prog=${CPT_ELF:="$( command -v readelf || command -v llvm-readelf || command -v eu-readelf)"} || elf_prog=ldd # Make note of the user's current ID to do root checks later on. # This is used enough to warrant a place here. uid=$(id -u) # Save IFS, so we can restore it back to what it was before. old_ifs=$IFS # Make sure that the CPT_ROOT doesn't end with a '/'. This might # break some operations. [ -z "$CPT_ROOT" ] || [ "${CPT_ROOT##*/}" ] || { warn "" "Your CPT_ROOT variable shouldn't end with '/'" CPT_ROOT=${CPT_ROOT%/} } # Define an optional sys_arch variable in order to provide # information to build files with architectural information. sys_arch=$(uname -m 2>/dev/null) ||: # Define this variable but don't create its directory structure from # the get go. It will be created as needed by package installation. sys_db=$CPT_ROOT/$pkg_db # This allows for automatic setup of a CPT chroot and will # do nothing on a normal system. mkdir -p "$CPT_ROOT/" 2>/dev/null ||: # Set a value for CPT_COMPRESS if it isn't set. : "${CPT_COMPRESS:=gz}" # Unless being piped or the user specifically doesn't want colors, set # colors. This can of course be overriden if the user specifically want # colors during piping. if { [ "$CPT_COLOR" != 0 ] && [ -t 1 ] ;} || [ "$CPT_COLOR" = 1 ]; then colory="\033[1;33m" colorb="\033[1;36m" colre="\033[m" fi # Set default package directories so that they can be parsed from package # helpers # TODO Document : "${CPT_PKG_PREFIX:=/usr}" \ "${CPT_PKG_ETCDIR:=/etc}" \ "${CPT_PKG_LIBDIR:=$CPT_PKG_PREFIX/lib}" \ "${CPT_PKG_BINDIR:=$CPT_PKG_PREFIX/bin}" \ "${CPT_PKG_SBINDIR:=$CPT_PKG_PREFIX/bin}" \ "${CPT_PKG_MANDIR:=/usr/share/man}" \ "${CPT_PKG_INFODIR:=/usr/share/info}" \ "${CPT_PKG_LOCALSTATEDIR:=/var}" } # If the library is being called with its own name, run arguments. if [ "${0##*/}" = cpt-lib ]; then pd() { setup REST help:usage -- "usage: ${0##*/} [funcall...]" global_options } eval "$(getoptions pd parse "$0")" parse "$@" eval set -- "$REST" "$@" fi