Compare commits
22 Commits
2025.0526.
...
main
Author | SHA1 | Date | |
---|---|---|---|
![]() |
9d4e5f76ce | ||
![]() |
366f5c2d0e | ||
![]() |
0b0f3df59c | ||
![]() |
f48302c05e | ||
![]() |
7f341699c1 | ||
![]() |
18c53acd71 | ||
![]() |
eb632c010c | ||
![]() |
964e8598b1 | ||
![]() |
a5cf9313e9 | ||
![]() |
ab73a47751 | ||
![]() |
1da7dc7951 | ||
![]() |
49d61f0da0 | ||
![]() |
27c0abcb9f | ||
![]() |
483ee4e3ef | ||
![]() |
f7294e01e4 | ||
![]() |
c836b26657 | ||
![]() |
7bf624589f | ||
![]() |
a5e339a358 | ||
![]() |
029823a6b4 | ||
![]() |
f79abd346e | ||
![]() |
940c2a12a1 | ||
![]() |
6ac651d4f0 |
@ -14,14 +14,26 @@ jobs:
|
||||
uses: actions/checkout@v4
|
||||
- name: Install build dependencies
|
||||
run: |
|
||||
cd ${{ gitea.workspace }}/source
|
||||
cd source
|
||||
./install_build_prerequisites.sh
|
||||
- name: Build
|
||||
- name: Build Native
|
||||
run: |
|
||||
cd ${{ gitea.workspace }}/source
|
||||
./multibuild.sh
|
||||
cd source
|
||||
./build_native.sh
|
||||
- name: Test
|
||||
run: |
|
||||
cd ${{ gitea.workspace }}/source/output
|
||||
./dropshell_x86_64 list
|
||||
./dropshell_x86_64 help
|
||||
cd source
|
||||
./test.sh
|
||||
- name: Build Production
|
||||
run: |
|
||||
cd source
|
||||
./build_production.sh
|
||||
- name: Test
|
||||
run: |
|
||||
cd source
|
||||
./test.sh
|
||||
- name: Publish
|
||||
run: |
|
||||
cd source
|
||||
./publish.sh
|
||||
|
||||
|
22
dropshell-install.sh
Executable file
22
dropshell-install.sh
Executable file
@ -0,0 +1,22 @@
|
||||
#!/bin/bash
|
||||
set -euo pipefail
|
||||
|
||||
# download and install dropshell
|
||||
|
||||
# 1. Determine architecture
|
||||
# -----------------------------------------------------------------------------
|
||||
|
||||
ARCH=$(uname -m)
|
||||
|
||||
TARGET_PATH="${HOME}/.local/bin/dropshell"
|
||||
[ ! -f "${TARGET_PATH}" ] || rm -f "${TARGET_PATH}"
|
||||
mkdir -p "$(dirname "${TARGET_PATH}")"
|
||||
curl -L -s -o "${TARGET_PATH}" "https://getbin.xyz/dropshell.${ARCH}" || die "Failed to download dropshell for ${ARCH}"
|
||||
chmod +x "${TARGET_PATH}"
|
||||
echo "dropshell installed successfully to $TARGET_PATH"
|
||||
echo " "
|
||||
|
||||
echo "Please:"
|
||||
echo "1. run '${TARGET_PATH} edit' to edit the configuration."
|
||||
echo "2. run '${TARGET_PATH} install' to install dropshell components on this computer."
|
||||
echo "3. run 'source ~/.bashrc' to add to your path and autocomplete for the current shell."
|
@ -1,4 +1,6 @@
|
||||
# can you make this script run in bash, but fall back to sh if bash is not installed?
|
||||
#!/bin/bash
|
||||
|
||||
# set up a remote server for use with dropshell.
|
||||
|
||||
# check if we are running as root
|
||||
if [ "$(id -u)" -ne 0 ]; then
|
49
install.sh
49
install.sh
@ -1,49 +0,0 @@
|
||||
#!/bin/bash
|
||||
set -e
|
||||
|
||||
# download and install dropshell
|
||||
|
||||
# 1. Determine architecture
|
||||
# -----------------------------------------------------------------------------
|
||||
|
||||
ARCH=$(uname -m)
|
||||
if [[ "$ARCH" == "x86_64" ]]; then
|
||||
BIN=dropshell.x86_64
|
||||
elif [[ "$ARCH" == "aarch64" || "$ARCH" == "arm64" ]]; then
|
||||
BIN=dropshell.aarch64
|
||||
else
|
||||
echo "Unsupported architecture: $ARCH" >&2
|
||||
exit 1
|
||||
fi
|
||||
|
||||
|
||||
# 2. Download the appropriate binary to a temp directory
|
||||
# -----------------------------------------------------------------------------
|
||||
TMPDIR=$(mktemp -d)
|
||||
trap 'rm -rf "$TMPDIR"' EXIT
|
||||
URL="https://gitea.jde.nz/public/dropshell/releases/download/latest/$BIN"
|
||||
echo "Downloading $BIN from $URL..."
|
||||
|
||||
TARGET_PATH="${HOME}/.local/bin/dropshell"
|
||||
|
||||
mkdir -p "${HOME}/.local/bin"
|
||||
|
||||
curl -fsSL -o "$TARGET_PATH" "$URL"
|
||||
|
||||
if [ ! -f "$TARGET_PATH" ]; then
|
||||
echo "Failed to download dropshell" >&2
|
||||
exit 1
|
||||
fi
|
||||
|
||||
chmod +x "$TARGET_PATH"
|
||||
|
||||
if [ ! -f "${HOME}/.local/bin/ds" ]; then
|
||||
ln -s "$TARGET_PATH" "${HOME}/.local/bin/ds"
|
||||
fi
|
||||
|
||||
echo "dropshell installed successfully to $TARGET_PATH"
|
||||
|
||||
echo "Please:"
|
||||
echo "1. run '${TARGET_PATH} edit' to edit the configuration."
|
||||
echo "2. run '${TARGET_PATH} install' to install dropshell components on this computer."
|
||||
echo "3. run 'source ~/.bashrc' to add to your path and autocomplete for the current shell."
|
@ -46,17 +46,16 @@ set(CMAKE_MODULE_PATH ${CMAKE_MODULE_PATH} ${CMAKE_CURRENT_SOURCE_DIR}/cmake)
|
||||
file(GLOB_RECURSE SOURCES "src/*.cpp")
|
||||
file(GLOB_RECURSE HEADERS "src/*.hpp")
|
||||
|
||||
# Add custom target to run make_createagent.sh at the start of the build process
|
||||
add_custom_target(run_createagent ALL
|
||||
COMMAND ${CMAKE_COMMAND} -E echo "Running make_createagent.sh..."
|
||||
COMMAND ${CMAKE_COMMAND} -E env bash ${CMAKE_CURRENT_SOURCE_DIR}/make_createagent.sh
|
||||
# Add custom target to run cmake_prebuild.sh at the start of the build process
|
||||
add_custom_target(run_prebuild_script ALL
|
||||
COMMAND ${CMAKE_COMMAND} -E echo "Running cmake_prebuild.sh..."
|
||||
COMMAND ${CMAKE_COMMAND} -E env bash ${CMAKE_CURRENT_SOURCE_DIR}/cmake_prebuild.sh
|
||||
WORKING_DIRECTORY ${CMAKE_CURRENT_SOURCE_DIR}
|
||||
COMMENT "Running make_createagent.sh before build"
|
||||
)
|
||||
|
||||
# Add executable
|
||||
add_executable(dropshell ${SOURCES})
|
||||
add_dependencies(dropshell run_createagent)
|
||||
add_dependencies(dropshell run_prebuild_script)
|
||||
|
||||
# Mark the generated files as GENERATED so CMake knows they'll be created during build
|
||||
set_source_files_properties(
|
||||
|
@ -1,5 +1,8 @@
|
||||
#!/bin/bash
|
||||
|
||||
# install the dropshell host agent on this computer.
|
||||
# (not for remote servers)
|
||||
|
||||
SCRIPT_DIR=$(dirname "$0")
|
||||
|
||||
echo "Installing dropshell host agent on this computer..."
|
||||
@ -22,27 +25,6 @@ _check_required_env_vars() {
|
||||
done
|
||||
}
|
||||
|
||||
# Checks if Docker is installed, running, and user has permission. Returns 1 on failure.
|
||||
_check_docker_installed() {
|
||||
if ! command -v docker &> /dev/null; then
|
||||
echo "Docker is not installed"
|
||||
return 1
|
||||
fi
|
||||
|
||||
# check if docker daemon is running
|
||||
if ! docker info &> /dev/null; then
|
||||
echo "Docker daemon is not running"
|
||||
return 1
|
||||
fi
|
||||
|
||||
# check if user has permission to run docker
|
||||
if ! docker run --rm hello-world &> /dev/null; then
|
||||
echo "User does not have permission to run docker"
|
||||
return 1
|
||||
fi
|
||||
|
||||
return 0
|
||||
}
|
||||
|
||||
function install_bb64() {
|
||||
# check curl installed
|
||||
@ -77,7 +59,5 @@ set +a
|
||||
_check_required_env_vars "AGENT_LOCAL_PATH"
|
||||
echo "Installing host agent into $AGENT_LOCAL_PATH"
|
||||
|
||||
_check_docker_installed || _die "Docker is required."
|
||||
|
||||
install_bb64
|
||||
|
||||
|
@ -38,6 +38,7 @@ CURRENT_EXIT_CODE=0
|
||||
load_dotenv(){
|
||||
local file_path=$1
|
||||
if [ -f "${file_path}" ]; then
|
||||
# shellcheck source=/dev/null
|
||||
source "${file_path}"
|
||||
fi
|
||||
}
|
||||
@ -69,9 +70,9 @@ function run_command() {
|
||||
load_dotenv "${service_path}/config/.template_info.env"
|
||||
|
||||
# update the main variables.
|
||||
CONFIG_PATH="${service_path}/config"
|
||||
SERVICE="${SERVICE_NAME}"
|
||||
DOCKER_CLI_HINTS=false
|
||||
export CONFIG_PATH="${service_path}/config"
|
||||
export SERVICE="${SERVICE_NAME}"
|
||||
export DOCKER_CLI_HINTS=false
|
||||
|
||||
set +a
|
||||
|
||||
|
@ -21,16 +21,13 @@ fi
|
||||
_check_required_env_vars "AGENT_PATH"
|
||||
|
||||
function install_bb64() {
|
||||
curl -fsSL "https://gitea.jde.nz/public/bb64/releases/download/latest/install.sh" | bash -s -- "$AGENT_PATH" "$(id -u $USER):$(id -g $USER)"
|
||||
|
||||
# test result code from curl
|
||||
if [ $? -ne 0 ]; then
|
||||
if ! curl -fsSL "https://gitea.jde.nz/public/bb64/releases/download/latest/install.sh" | \
|
||||
bash -s -- "$AGENT_PATH" "$(id -u "$USER"):$(id -g "$USER")"; then
|
||||
_die "Failed to install bb64. Curl returned non-zero exit code."
|
||||
fi
|
||||
|
||||
# test if bb64 is installed
|
||||
VER=$("$AGENT_PATH/bb64" -v)
|
||||
if [ $? -ne 0 ]; then
|
||||
if ! VER=$("$AGENT_PATH/bb64" -v); then
|
||||
_die "bb64 did not install correctly."
|
||||
fi
|
||||
|
||||
@ -53,6 +50,12 @@ if ! command -v docker &> /dev/null; then
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# check rsync installation
|
||||
if ! command -v rsync &> /dev/null; then
|
||||
echo "Rsync is not installed. Rsync is required for agent installation."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
|
||||
#-------------------------------------------------------------------------
|
||||
|
||||
|
@ -41,18 +41,18 @@ _create_and_start_container() {
|
||||
local run_cmd="$1"
|
||||
local container_name="$2"
|
||||
|
||||
if _is_container_exists $container_name; then
|
||||
_is_container_running $container_name && return 0
|
||||
_start_container $container_name
|
||||
if _is_container_exists "$container_name"; then
|
||||
_is_container_running "$container_name" && return 0
|
||||
_start_container "$container_name"
|
||||
else
|
||||
$run_cmd
|
||||
fi
|
||||
|
||||
if ! _is_container_running $container_name; then
|
||||
if ! _is_container_running "$container_name"; then
|
||||
_die "Container ${container_name} failed to start"
|
||||
fi
|
||||
|
||||
ID=$(_get_container_id $container_name)
|
||||
ID=$(_get_container_id "$container_name")
|
||||
echo "Container ${container_name} is running with ID ${ID}"
|
||||
}
|
||||
|
||||
@ -93,6 +93,7 @@ _check_docker_installed() {
|
||||
|
||||
# Checks if a container (any state) exists. Returns 1 if not found.
|
||||
_is_container_exists() {
|
||||
[ -n "${1:-}" ] || { echo "_is_container_exists: Container name is empty" >&2; return 1; }
|
||||
if ! docker ps -a --format "{{.Names}}" | grep -q "^$1$"; then
|
||||
return 1
|
||||
fi
|
||||
@ -101,6 +102,7 @@ _is_container_exists() {
|
||||
|
||||
# Checks if a container is currently running. Returns 1 if not running.
|
||||
_is_container_running() {
|
||||
[ -n "${1:-}" ] || { echo "_is_container_running: Container name is empty" >&2; return 1; }
|
||||
if ! docker ps --format "{{.Names}}" | grep -q "^$1$"; then
|
||||
return 1
|
||||
fi
|
||||
@ -119,39 +121,39 @@ _get_container_status() {
|
||||
|
||||
# Starts an existing, stopped container.
|
||||
_start_container() {
|
||||
_is_container_exists $1 || return 1
|
||||
_is_container_running $1 && return 0
|
||||
docker start $1
|
||||
_is_container_exists "$1" || return 1
|
||||
_is_container_running "$1" && return 0
|
||||
docker start "$1"
|
||||
}
|
||||
|
||||
# Stops a running container.
|
||||
_stop_container() {
|
||||
_is_container_running $1 || return 0;
|
||||
docker stop $1
|
||||
_is_container_running "$1" || return 0;
|
||||
docker stop "$1"
|
||||
}
|
||||
|
||||
# Stops (if needed) and removes a container.
|
||||
_remove_container() {
|
||||
_stop_container $1
|
||||
_is_container_exists $1 || return 0;
|
||||
docker rm $1
|
||||
_stop_container "$1"
|
||||
_is_container_exists "$1" || return 0;
|
||||
docker rm "$1"
|
||||
}
|
||||
|
||||
# Prints the logs for a container.
|
||||
_get_container_logs() {
|
||||
if ! _is_container_exists $1; then
|
||||
if ! _is_container_exists "$1"; then
|
||||
echo "Container $1 does not exist"
|
||||
return 1
|
||||
fi
|
||||
|
||||
docker logs $1
|
||||
docker logs "$1"
|
||||
}
|
||||
|
||||
# Checks if listed environment variables are set; calls _die() if any are missing.
|
||||
_check_required_env_vars() {
|
||||
local required_vars=("$@")
|
||||
for var in "${required_vars[@]}"; do
|
||||
if [ -z "${!var}" ]; then
|
||||
if [ -z "${!var:-}" ]; then
|
||||
_die "Required environment variable $var is not set"
|
||||
fi
|
||||
done
|
||||
|
@ -12,26 +12,26 @@ _autocommandrun_volume() {
|
||||
|
||||
case "$command" in
|
||||
create)
|
||||
if docker volume ls | grep -q ${volume_name}; then
|
||||
if docker volume ls | grep -q "${volume_name}"; then
|
||||
echo "Volume ${volume_name} already exists - leaving unchanged"
|
||||
return
|
||||
fi
|
||||
echo "Creating volume ${volume_name}"
|
||||
docker volume create ${volume_name}
|
||||
docker volume create "${volume_name}"
|
||||
;;
|
||||
destroy)
|
||||
echo "Destroying volume ${volume_name}"
|
||||
docker volume rm ${volume_name}
|
||||
docker volume rm "${volume_name}"
|
||||
;;
|
||||
backup)
|
||||
echo "Backing up volume ${volume_name}"
|
||||
docker run --rm -v ${volume_name}:/volume -v ${backup_folder}:/backup debian bash -c "tar -czvf /backup/backup.tgz -C /volume . && chown -R $MYID:$MYGRP /backup"
|
||||
docker run --rm -v "${volume_name}":/volume -v "${backup_folder}":/backup debian bash -c "tar -czvf /backup/backup.tgz -C /volume . && chown -R $MYID:$MYGRP /backup"
|
||||
;;
|
||||
restore)
|
||||
echo "Restoring volume ${volume_name}"
|
||||
docker volume rm ${volume_name}
|
||||
docker volume create ${volume_name}
|
||||
docker run --rm -v ${volume_name}:/volume -v ${backup_folder}:/backup debian bash -c "tar -xzvf /backup/backup.tgz -C /volume --strip-components=1"
|
||||
docker volume rm "${volume_name}"
|
||||
docker volume create "${volume_name}"
|
||||
docker run --rm -v "${volume_name}":/volume -v "${backup_folder}":/backup debian bash -c "tar -xzvf /backup/backup.tgz -C /volume --strip-components=1"
|
||||
;;
|
||||
esac
|
||||
}
|
||||
@ -48,14 +48,16 @@ _autocommandrun_path() {
|
||||
return
|
||||
fi
|
||||
echo "Creating path ${path}"
|
||||
mkdir -p ${path}
|
||||
mkdir -p "${path}"
|
||||
;;
|
||||
destroy)
|
||||
echo "Destroying path ${path}"
|
||||
local path_parent=$(dirname ${path})
|
||||
local path_child=$(basename ${path})
|
||||
local path_parent;
|
||||
path_parent=$(dirname "${path}")
|
||||
local path_child;
|
||||
path_child=$(basename "${path}")
|
||||
if [ -d "${path_parent}/${path_child}" ]; then
|
||||
docker run --rm -v ${path_parent}:/volume debian bash -c "rm -rfv /volume/${path_child}" || echo "Failed to destroy path ${path}"
|
||||
docker run --rm -v "${path_parent}":/volume debian bash -c "rm -rfv /volume/${path_child}" || echo "Failed to destroy path ${path}"
|
||||
else
|
||||
echo "Path ${path} does not exist - nothing to destroy"
|
||||
fi
|
||||
@ -63,7 +65,7 @@ _autocommandrun_path() {
|
||||
backup)
|
||||
echo "Backing up path ${path}"
|
||||
if [ -d "${path}" ]; then
|
||||
docker run --rm -v ${path}:/path -v ${backup_folder}:/backup debian bash -c "tar -czvf /backup/backup.tgz -C /path . && chown -R $MYID:$MYGRP /backup"
|
||||
docker run --rm -v "${path}":/path -v "${backup_folder}":/backup debian bash -c "tar -czvf /backup/backup.tgz -C /path . && chown -R $MYID:$MYGRP /backup"
|
||||
else
|
||||
echo "Path ${path} does not exist - nothing to backup"
|
||||
fi
|
||||
@ -73,9 +75,9 @@ _autocommandrun_path() {
|
||||
echo "Backup file ${backup_folder}/backup.tgz does not exist - nothing to restore"
|
||||
else
|
||||
echo "Clearing existing data in path ${path}"
|
||||
docker run --rm -v ${path}:/path debian bash -c "rm -rfv /path/{*,.*}"
|
||||
docker run --rm -v "${path}":/path debian bash -c "rm -rfv /path/{*,.*}"
|
||||
echo "Restoring path ${path} from backup file ${backup_folder}/backup.tgz"
|
||||
tar -xzvf ${backup_folder}/backup.tgz -C ${path} --strip-components=1
|
||||
tar -xzvf "${backup_folder}/backup.tgz" -C "${path}" --strip-components=1
|
||||
fi
|
||||
;;
|
||||
esac
|
||||
@ -88,31 +90,36 @@ _autocommandrun_file() {
|
||||
|
||||
case "$command" in
|
||||
create)
|
||||
filepath_parent=$(dirname ${filepath})
|
||||
filepath_child=$(basename ${filepath})
|
||||
if [ ! -d "${filepath_parent}" ]; then
|
||||
echo "Parent directory ${filepath_parent} of ${filepath_child} does not exist - creating"
|
||||
mkdir -p ${filepath_parent}
|
||||
local file_parent;
|
||||
file_parent=$(dirname "${filepath}")
|
||||
local file_name;
|
||||
file_name=$(basename "${filepath}")
|
||||
if [ ! -d "${file_parent}" ]; then
|
||||
echo "Parent directory ${file_parent} of ${file_name} does not exist - creating"
|
||||
mkdir -p "${file_parent}"
|
||||
fi
|
||||
;;
|
||||
destroy)
|
||||
rm -f ${filepath}
|
||||
rm -f "${filepath}"
|
||||
;;
|
||||
backup)
|
||||
echo "Backing up file ${filepath}"
|
||||
local file_parent=$(dirname ${filepath})
|
||||
local file_name=$(basename ${filepath})
|
||||
local file_parent;
|
||||
file_parent=$(dirname "${filepath}")
|
||||
local file_name;
|
||||
file_name=$(basename "${filepath}")
|
||||
if [ -f "${file_parent}/${file_name}" ]; then
|
||||
docker run --rm -v ${file_parent}:/volume -v ${backup_folder}:/backup debian bash -c "cp /volume/${file_name} /backup/${file_name} && chown -R $MYID:$MYGRP /backup"
|
||||
docker run --rm -v "${file_parent}":/volume -v "${backup_folder}":/backup debian bash -c "cp /volume/${file_name} /backup/${file_name} && chown -R $MYID:$MYGRP /backup"
|
||||
else
|
||||
echo "File ${filepath} does not exist - nothing to backup"
|
||||
fi
|
||||
;;
|
||||
restore)
|
||||
echo "Restoring file ${filepath}"
|
||||
local file_name=$(basename ${filepath})
|
||||
rm -f ${filepath} || die "Unable to remove existing file ${filepath}, restore failed."
|
||||
cp ${backup_folder}/${file_name} ${filepath} || die "Unable to copy file ${backup_folder}/${file_name} to ${filepath}, restore failed."
|
||||
local file_name;
|
||||
file_name=$(basename "${filepath}")
|
||||
rm -f "${filepath}" || return_die "Unable to remove existing file ${filepath}, restore failed."
|
||||
cp "${backup_folder}/${file_name}" "${filepath}" || return_die "Unable to copy file ${backup_folder}/${file_name} to ${filepath}, restore failed."
|
||||
;;
|
||||
esac
|
||||
}
|
||||
@ -153,9 +160,10 @@ _autocommandparse() {
|
||||
local value="${pair#*=}"
|
||||
|
||||
# create backup folder unique to key/value.
|
||||
local bfolder=$(echo "${key}_${value}" | tr -cd '[:alnum:]_-')
|
||||
local bfolder;
|
||||
bfolder=$(echo "${key}_${value}" | tr -cd '[:alnum:]_-')
|
||||
local targetpath="${backup_temp_path}/${bfolder}"
|
||||
mkdir -p ${targetpath}
|
||||
mkdir -p "${targetpath}"
|
||||
|
||||
# Key must be one of volume, path or file
|
||||
case "$key" in
|
||||
@ -191,7 +199,7 @@ databackup() {
|
||||
|
||||
|
||||
mkdir -p "$BACKUP_TEMP_PATH"
|
||||
echo "_autocommandparse [backup] [$BACKUP_TEMP_PATH] [$@]"
|
||||
echo "_autocommandparse [backup] [$BACKUP_TEMP_PATH]" "$@"
|
||||
_autocommandparse backup "$BACKUP_TEMP_PATH" "$@"
|
||||
|
||||
tar zcvf "$BACKUP_FILE" -C "$BACKUP_TEMP_PATH" .
|
||||
@ -201,7 +209,7 @@ datarestore() {
|
||||
_check_required_env_vars "BACKUP_FILE" "TEMP_DIR"
|
||||
BACKUP_TEMP_PATH="$TEMP_DIR/restore"
|
||||
|
||||
echo "_autocommandparse [restore] [$BACKUP_TEMP_PATH] [$@]"
|
||||
echo "_autocommandparse [restore] [$BACKUP_TEMP_PATH]" "$@"
|
||||
|
||||
mkdir -p "$BACKUP_TEMP_PATH"
|
||||
tar zxvf "$BACKUP_FILE" -C "$BACKUP_TEMP_PATH" --strip-components=1
|
||||
|
@ -1,86 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" &> /dev/null && pwd )"
|
||||
|
||||
# Exit on error
|
||||
set -e
|
||||
|
||||
# Colors for output
|
||||
RED='\033[0;31m'
|
||||
GREEN='\033[0;32m'
|
||||
YELLOW='\033[1;33m'
|
||||
NC='\033[0m' # No Color
|
||||
|
||||
JOBS=4
|
||||
# Determine number of CPU cores for parallel build
|
||||
if command -v nproc >/dev/null 2>&1; then
|
||||
JOBS=$(nproc)
|
||||
fi
|
||||
|
||||
# Function to print status messages
|
||||
print_status() {
|
||||
echo -e "${GREEN}[*] $1${NC}"
|
||||
}
|
||||
|
||||
print_error() {
|
||||
echo -e "${RED}[!] $1${NC}"
|
||||
}
|
||||
|
||||
print_warning() {
|
||||
echo -e "${YELLOW}[!] $1${NC}"
|
||||
}
|
||||
|
||||
# ensure we have latest dehydrate.
|
||||
dehydrate -u
|
||||
|
||||
# Check if build directory exists, if not create it
|
||||
if [ ! -d "build" ]; then
|
||||
print_status "Creating build directory..."
|
||||
mkdir build
|
||||
fi
|
||||
|
||||
# Enter build directory
|
||||
cd build
|
||||
|
||||
# Check if CMake is installed
|
||||
if ! command -v cmake &> /dev/null; then
|
||||
print_error "CMake is not installed. Please install CMake first."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Check if Ninja is installed
|
||||
if ! command -v ninja &> /dev/null; then
|
||||
print_error "Ninja is not installed. Please install Ninja first."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Check if ccache is installed
|
||||
if ! command -v ccache &> /dev/null; then
|
||||
print_warning "ccache is not installed. Builds will be slower without it."
|
||||
print_warning "Consider installing ccache for faster builds."
|
||||
fi
|
||||
|
||||
# Configure with CMake
|
||||
print_status "Configuring with CMake..."
|
||||
cmake .. -G Ninja -DCMAKE_BUILD_TYPE=Debug -DCMAKE_C_COMPILER_LAUNCHER=ccache -DCMAKE_CXX_COMPILER_LAUNCHER=ccache
|
||||
|
||||
# Build the project
|
||||
print_status "Building project..."
|
||||
ninja -j"$JOBS"
|
||||
|
||||
# Check if build was successful
|
||||
if [ $? -eq 0 ]; then
|
||||
print_status "Build successful!"
|
||||
else
|
||||
print_error "Build failed!"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
print_status "Auto-installing dropshell locally..."
|
||||
mkdir -p "${HOME}/.local/bin"
|
||||
cp "$SCRIPT_DIR/build/dropshell" "${HOME}/.local/bin/dropshell"
|
||||
|
||||
# Return to original directory
|
||||
cd ..
|
||||
|
||||
print_status "Build process completed!"
|
53
source/build_native.sh
Executable file
53
source/build_native.sh
Executable file
@ -0,0 +1,53 @@
|
||||
#!/bin/bash
|
||||
|
||||
SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" &> /dev/null && pwd )"
|
||||
OUTPUT_DIR=${SCRIPT_DIR}/output
|
||||
INSTALL_DIR=${HOME}/.local/bin
|
||||
mkdir -p "${OUTPUT_DIR}"
|
||||
|
||||
# Exit on error
|
||||
set -euo pipefail
|
||||
|
||||
ARCH=$(uname -m)
|
||||
if [ "$ARCH" != "x86_64" ] && [ "$ARCH" != "aarch64" ]; then
|
||||
echo "Unsupported architecture: $ARCH"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
|
||||
function build_native() {
|
||||
local BUILDDIR=${SCRIPT_DIR}/build/native
|
||||
local PREVDIR=$PWD
|
||||
local JOBS;
|
||||
JOBS=$(nproc) # Set JOBS to the number of available CPU cores
|
||||
mkdir -p "${BUILDDIR}"
|
||||
cd "${SCRIPT_DIR}" || exit 1
|
||||
|
||||
CC="${HOME}/.musl-cross/${ARCH}-linux-musl-native/bin/${ARCH}-linux-musl-gcc"
|
||||
CXX="${HOME}/.musl-cross/${ARCH}-linux-musl-native/bin/${ARCH}-linux-musl-g++"
|
||||
|
||||
|
||||
cmake -B "${BUILDDIR}" -G Ninja \
|
||||
-DCMAKE_BUILD_TYPE=Debug \
|
||||
-DCMAKE_C_COMPILER_LAUNCHER=ccache \
|
||||
-DCMAKE_CXX_COMPILER_LAUNCHER=ccache \
|
||||
-DCMAKE_LINKER=mold \
|
||||
-DCMAKE_C_COMPILER="${CC}" \
|
||||
-DCMAKE_CXX_COMPILER="${CXX}"
|
||||
|
||||
cd "${BUILDDIR}" || exit 1
|
||||
ninja -j"$JOBS"
|
||||
|
||||
#upx ${BUILDDIR}/dropshell
|
||||
cp "${BUILDDIR}/dropshell" "${OUTPUT_DIR}/dropshell.${ARCH}"
|
||||
|
||||
cd "${PREVDIR}" || exit 1
|
||||
}
|
||||
|
||||
build_native
|
||||
|
||||
|
||||
echo "Auto-installing dropshell locally..."
|
||||
mkdir -p "${INSTALL_DIR}"
|
||||
cp "${OUTPUT_DIR}/dropshell.${ARCH}" "${INSTALL_DIR}/dropshell"
|
||||
echo "Build process completed!"
|
@ -1,8 +1,12 @@
|
||||
#!/bin/bash
|
||||
set -euo pipefail
|
||||
SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" &> /dev/null && pwd )"
|
||||
|
||||
# Create output directory
|
||||
mkdir -p output
|
||||
|
||||
mkdir -p "${SCRIPT_DIR}/output"
|
||||
PREV_DIR=$(pwd)
|
||||
cd "${SCRIPT_DIR}"
|
||||
trap 'cd "${PREV_DIR}"' EXIT
|
||||
|
||||
function build_arch() {
|
||||
local arch=$1
|
||||
@ -16,14 +20,14 @@ function build_arch() {
|
||||
CC="${HOME}/.musl-cross/${arch}-linux-musl-cross/bin/${arch}-linux-musl-gcc"
|
||||
CXX="${HOME}/.musl-cross/${arch}-linux-musl-cross/bin/${arch}-linux-musl-g++"
|
||||
|
||||
BUILDDIR=build/${arch}
|
||||
mkdir -p ${BUILDDIR}
|
||||
BUILDDIR="${SCRIPT_DIR}/build/${arch}"
|
||||
mkdir -p "${BUILDDIR}"
|
||||
|
||||
cmake -B ${BUILDDIR} -G Ninja -DCMAKE_BUILD_TYPE=${CMAKE_BUILD_TYPE} -DCMAKE_C_COMPILER=${CC} -DCMAKE_CXX_COMPILER=${CXX}
|
||||
cmake --build ${BUILDDIR}
|
||||
cmake -B "${BUILDDIR}" -G Ninja -DCMAKE_BUILD_TYPE="${CMAKE_BUILD_TYPE}" -DCMAKE_C_COMPILER="${CC}" -DCMAKE_CXX_COMPILER="${CXX}"
|
||||
cmake --build "${BUILDDIR}"
|
||||
|
||||
upx ${BUILDDIR}/dropshell
|
||||
cp ${BUILDDIR}/dropshell output/dropshell.${arch}
|
||||
upx "${BUILDDIR}/dropshell"
|
||||
cp "${BUILDDIR}/dropshell" "${SCRIPT_DIR}/output/dropshell.${arch}"
|
||||
}
|
||||
|
||||
build_arch x86_64
|
@ -1,6 +1,9 @@
|
||||
#!/bin/bash
|
||||
set -e
|
||||
|
||||
# CMake pre-build script.
|
||||
# Runs before the build process.
|
||||
|
||||
# This script creates two files:
|
||||
# src/utils/createagent.hpp
|
||||
# src/utils/createagent.cpp
|
||||
@ -11,9 +14,11 @@ SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
|
||||
if ! command -v dehydrate &> /dev/null; then
|
||||
echo "dehydrate could not be found - installing"
|
||||
curl -fsSL https://gitea.jde.nz/public/dehydrate/releases/download/latest/install.sh | bash
|
||||
else
|
||||
# ensure we have latest dehydrate.
|
||||
dehydrate -u
|
||||
fi
|
||||
|
||||
mkdir -p "${SCRIPT_DIR}/src/autogen"
|
||||
|
||||
dehydrate "${SCRIPT_DIR}/agent-remote" "${SCRIPT_DIR}/src/autogen"
|
||||
dehydrate "${SCRIPT_DIR}/agent-local" "${SCRIPT_DIR}/src/autogen"
|
@ -37,11 +37,23 @@ fi
|
||||
|
||||
print_status "Detected OS: $OS $VER"
|
||||
|
||||
|
||||
#----------------------------------------------------------------------------------------------------------
|
||||
# INSTALL PREREQUISITE PACKAGES
|
||||
#----------------------------------------------------------------------------------------------------------
|
||||
|
||||
# Define packages based on distribution
|
||||
case $OS in
|
||||
"Ubuntu"|"Debian GNU/Linux")
|
||||
# Common packages for both Ubuntu and Debian
|
||||
PACKAGES="cmake make g++ devscripts debhelper build-essential upx"
|
||||
PACKAGES="bash cmake make g++ devscripts debhelper build-essential upx musl-tools wget tar ccache ninja-build"
|
||||
INSTALLCMD="apt-get install -y"
|
||||
UPDATECMD="apt-get update"
|
||||
;;
|
||||
"Alpine Linux")
|
||||
PACKAGES="bash build-base cmake git nlohmann-json wget tar curl ninja mold nodejs npm"
|
||||
INSTALLCMD="apk add --no-cache"
|
||||
UPDATECMD="apk update"
|
||||
;;
|
||||
*)
|
||||
print_error "Unsupported distribution: $OS"
|
||||
@ -51,19 +63,29 @@ esac
|
||||
|
||||
# Function to check if a package is installed
|
||||
is_package_installed() {
|
||||
if [ "$OS" = "Alpine Linux" ]; then
|
||||
# Use apk info <pkg> and check exit status
|
||||
apk info "$1" >/dev/null 2>&1
|
||||
return $?
|
||||
else
|
||||
dpkg -l "$1" 2>/dev/null | grep -q "^ii"
|
||||
fi
|
||||
}
|
||||
|
||||
# Update package lists
|
||||
print_status "Updating package lists..."
|
||||
apt-get update
|
||||
UPDATED=false
|
||||
|
||||
# Install missing packages
|
||||
print_status "Checking and installing required packages..."
|
||||
for pkg in $PACKAGES; do
|
||||
if ! is_package_installed "$pkg"; then
|
||||
if [ "$UPDATED" = false ]; then
|
||||
print_status "Updating package lists..."
|
||||
$UPDATECMD
|
||||
UPDATED=true
|
||||
fi
|
||||
|
||||
print_status "Installing $pkg..."
|
||||
apt-get install -y "$pkg"
|
||||
$INSTALLCMD "$pkg"
|
||||
if [ $? -ne 0 ]; then
|
||||
print_error "Failed to install $pkg"
|
||||
exit 1
|
||||
@ -73,82 +95,71 @@ for pkg in $PACKAGES; do
|
||||
fi
|
||||
done
|
||||
|
||||
# Verify all required tools are installed
|
||||
print_status "Verifying installation..."
|
||||
for tool in cmake make g++; do
|
||||
if ! command -v "$tool" &> /dev/null; then
|
||||
print_error "$tool is not installed properly"
|
||||
exit 1
|
||||
fi
|
||||
done
|
||||
|
||||
|
||||
|
||||
# Install other required packages
|
||||
apt install -y musl-tools wget tar ccache ninja-build
|
||||
# ----------------------------------------------------------------------------------------------------------
|
||||
# MUSL CROSS COMPILERS
|
||||
# ----------------------------------------------------------------------------------------------------------
|
||||
|
||||
# Set install directory
|
||||
if [ -n "$SUDO_USER" ] && [ "$SUDO_USER" != "root" ]; then
|
||||
USER_HOME=$(eval echo ~$SUDO_USER)
|
||||
USER_HOME=$(eval echo "~$SUDO_USER")
|
||||
else
|
||||
USER_HOME="$HOME"
|
||||
fi
|
||||
INSTALL_DIR="$USER_HOME/.musl-cross"
|
||||
mkdir -p "$INSTALL_DIR"
|
||||
|
||||
MUSL_CC_URL="https://musl.cc"
|
||||
TMPDIR=$(mktemp -d)
|
||||
trap 'rm -rf "$TMPDIR"' EXIT
|
||||
|
||||
# x86_64
|
||||
if [ ! -d "$INSTALL_DIR/x86_64-linux-musl-cross" ]; then
|
||||
echo "Downloading x86_64 musl cross toolchain..."
|
||||
wget -nc -O "$TMPDIR/x86_64-linux-musl-cross.tgz" $MUSL_CC_URL/x86_64-linux-musl-cross.tgz
|
||||
tar -C "$INSTALL_DIR" -xvf "$TMPDIR/x86_64-linux-musl-cross.tgz"
|
||||
function install_musl_cross() {
|
||||
local TOOLCHAIN="$1"
|
||||
local MUSL_CC_URL="https://musl.cc"
|
||||
if [ ! -d "$INSTALL_DIR/$TOOLCHAIN" ]; then
|
||||
echo "Downloading $TOOLCHAIN musl cross toolchain..."
|
||||
wget -nc -O "$TMPDIR/$TOOLCHAIN.tgz" "$MUSL_CC_URL/$TOOLCHAIN.tgz"
|
||||
tar -C "$INSTALL_DIR" -xvf "$TMPDIR/$TOOLCHAIN.tgz"
|
||||
fi
|
||||
}
|
||||
|
||||
# aarch64
|
||||
if [ ! -d "$INSTALL_DIR/aarch64-linux-musl-cross" ]; then
|
||||
echo "Downloading aarch64 musl cross toolchain..."
|
||||
wget -nc -O "$TMPDIR/aarch64-linux-musl-cross.tgz" $MUSL_CC_URL/aarch64-linux-musl-cross.tgz
|
||||
tar -C "$INSTALL_DIR" -xvf "$TMPDIR/aarch64-linux-musl-cross.tgz"
|
||||
function check_path() {
|
||||
if [ -n "$SUDO_USER" ] && [ "$SUDO_USER" != "root" ]; then
|
||||
local BASHRC="$USER_HOME/.bashrc"
|
||||
local TOOLCHAIN="$1"
|
||||
local MUSL_PATH="$INSTALL_DIR/$TOOLCHAIN/bin"
|
||||
if ! echo "$PATH" | grep -q "$MUSL_PATH"; then
|
||||
echo "Adding $MUSL_PATH to PATH in $BASHRC"
|
||||
PATH_LINE="export PATH=\"$MUSL_PATH:\$PATH\""
|
||||
if ! grep -Fxq "$PATH_LINE" "$BASHRC"; then
|
||||
echo "" >> "$BASHRC"
|
||||
echo "# Add musl cross compilers to PATH for dropshell" >> "$BASHRC"
|
||||
echo "$PATH_LINE" >> "$BASHRC"
|
||||
echo "Added musl cross compilers to $BASHRC"
|
||||
echo "You should run 'source ~/.bashrc' to update your PATH"
|
||||
else
|
||||
echo "You should run 'source ~/.bashrc' to update your PATH"
|
||||
fi
|
||||
fi
|
||||
fi
|
||||
}
|
||||
|
||||
# Print instructions for adding to PATH
|
||||
# cat <<EOF
|
||||
|
||||
# To use the musl cross compilers, add the following to your shell:
|
||||
# export PATH="$INSTALL_DIR/x86_64-linux-musl-cross/bin:$INSTALL_DIR/aarch64-linux-musl-cross/bin:$PATH"
|
||||
TOOLCHAIN_LIST=(
|
||||
"aarch64-linux-musl-cross"
|
||||
"x86_64-linux-musl-cross"
|
||||
"x86_64-linux-musl-native"
|
||||
)
|
||||
|
||||
# Or run:
|
||||
# export PATH="$INSTALL_DIR/x86_64-linux-musl-cross/bin:$INSTALL_DIR/aarch64-linux-musl-cross/bin:\$PATH"
|
||||
|
||||
# EOF
|
||||
for TOOLCHAIN in "${TOOLCHAIN_LIST[@]}"; do
|
||||
install_musl_cross "$TOOLCHAIN"
|
||||
check_path "$TOOLCHAIN"
|
||||
done
|
||||
|
||||
# Clean up
|
||||
rm -rf "$TMPDIR"
|
||||
|
||||
# If run with sudo, add to invoking user's ~/.bashrc
|
||||
if [ -n "$SUDO_USER" ] && [ "$SUDO_USER" != "root" ]; then
|
||||
BASHRC="$USER_HOME/.bashrc"
|
||||
EXPORT_LINE="export PATH=\"$INSTALL_DIR/x86_64-linux-musl-cross/bin:$INSTALL_DIR/aarch64-linux-musl-cross/bin:\$PATH\""
|
||||
if ! grep -Fxq "$EXPORT_LINE" "$BASHRC"; then
|
||||
echo "" >> "$BASHRC"
|
||||
echo "# Add musl cross compilers to PATH for bb64" >> "$BASHRC"
|
||||
echo "$EXPORT_LINE" >> "$BASHRC"
|
||||
echo "Added musl cross compilers to $BASHRC"
|
||||
else
|
||||
echo "musl cross compiler PATH already present in $BASHRC"
|
||||
fi
|
||||
fi
|
||||
|
||||
# check if dehydrate command is installed
|
||||
if ! command -v dehydrate &> /dev/null; then
|
||||
curl -fsSL https://gitea.jde.nz/public/dehydrate/releases/download/latest/install.sh | bash
|
||||
fi
|
||||
|
||||
dehydrate -u
|
||||
|
||||
# ----------------------------------------------------------------------------------------------------------
|
||||
# COMPLETE
|
||||
# ----------------------------------------------------------------------------------------------------------
|
||||
|
||||
print_status "All dependencies installed successfully!"
|
||||
print_status "You can now run ./build.sh to build the project"
|
@ -9,74 +9,21 @@ echo "Script directory: $SCRIPT_DIR"
|
||||
TOKEN="${GITEA_TOKEN_DEPLOY:-${GITEA_TOKEN}}"
|
||||
[ -z "$TOKEN" ] && { echo "Neither GITEA_TOKEN_DEPLOY nor GITEA_TOKEN environment variable set!" >&2; exit 1; }
|
||||
|
||||
$SCRIPT_DIR/multibuild.sh
|
||||
BUILD_DIR=$SCRIPT_DIR/build
|
||||
|
||||
OLD_PWD=$PWD
|
||||
cd $SCRIPT_DIR
|
||||
|
||||
# Check for required binaries
|
||||
REQUIRED_BINARIES=("dropshell.x86_64" "dropshell.aarch64")
|
||||
for binary in "${REQUIRED_BINARIES[@]}"; do
|
||||
if [ ! -f "output/$binary" ]; then
|
||||
echo "output/$binary not found!" >&2
|
||||
echo "Please run multibuild.sh first." >&2
|
||||
exit 1
|
||||
fi
|
||||
done
|
||||
|
||||
TAG=$("$SCRIPT_DIR/output/dropshell.x86_64" --version)
|
||||
[ -z "$TAG" ] && echo "Failed to get version from dropshell.x86_64" >&2 && exit 1
|
||||
OLD_PWD="$PWD"
|
||||
cd "$SCRIPT_DIR" || exit 1
|
||||
TEMP_DIR=$(mktemp -d)
|
||||
trap 'rm -rf "$TEMP_DIR" && cd "$OLD_PWD"' EXIT
|
||||
|
||||
ARCH=$(uname -m)
|
||||
TAG=$("$SCRIPT_DIR/output/dropshell.${ARCH}" --version)
|
||||
[ -z "$TAG" ] && echo "Failed to get version from dropshell.${ARCH}" >&2 && exit 1
|
||||
echo "Publishing dropshell version $TAG"
|
||||
|
||||
# make sure we've commited.
|
||||
git add "$SCRIPT_DIR/../" && git commit -m "dropshell release $TAG" && git push
|
||||
|
||||
|
||||
# Find repo info from .git/config
|
||||
REPO_URL=$(git config --get remote.origin.url)
|
||||
if [[ ! $REPO_URL =~ gitea ]]; then
|
||||
echo "Remote origin is not a Gitea repository: $REPO_URL" >&2
|
||||
function die() {
|
||||
echo "$@" >&2
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Extract base URL, owner, and repo
|
||||
# Example: https://gitea.example.com/username/reponame.git
|
||||
BASE_URL=$(echo "$REPO_URL" | sed -E 's#(https?://[^/]+)/.*#\1#')
|
||||
OWNER=$(echo "$REPO_URL" | sed -E 's#.*/([^/]+)/[^/]+(\.git)?$#\1#')
|
||||
REPO=$(echo "$REPO_URL" | sed -E 's#.*/([^/]+)(\.git)?$#\1#')
|
||||
|
||||
API_URL="$BASE_URL/api/v1/repos/$OWNER/$REPO"
|
||||
|
||||
# Create release
|
||||
RELEASE_DATA=$(cat <<EOF
|
||||
{
|
||||
"tag_name": "$TAG",
|
||||
"name": "$TAG",
|
||||
"body": "dropshell release $TAG",
|
||||
"draft": false,
|
||||
"prerelease": false
|
||||
}
|
||||
EOF
|
||||
)
|
||||
|
||||
# Capture stdout and stderr of the curl command
|
||||
CURL_OUTPUT=$(curl -X POST "$API_URL/releases" \
|
||||
-H "Content-Type: application/json" \
|
||||
-H "Authorization: token $TOKEN" \
|
||||
-d "$RELEASE_DATA" 2>&1)
|
||||
|
||||
# Extract the release ID from the captured output
|
||||
RELEASE_ID=$(echo "$CURL_OUTPUT" | grep -o '"id":[0-9]*' | head -1 | cut -d: -f2)
|
||||
|
||||
if [ -z "$RELEASE_ID" ]; then
|
||||
echo "Failed to create release on Gitea." >&2
|
||||
echo "Release ID returned: $RELEASE_ID" >&2
|
||||
echo "Curl Output/Error:" >&2
|
||||
echo "$CURL_OUTPUT" >&2
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Function to find file in specified locations
|
||||
find_file() {
|
||||
@ -94,25 +41,17 @@ find_file() {
|
||||
return 1
|
||||
}
|
||||
|
||||
curl -L -s -o "${TEMP_DIR}/sos" "https://getbin.xyz/sos" || die "Failed to download sos"
|
||||
chmod +x "${TEMP_DIR}/sos"
|
||||
|
||||
|
||||
# Upload binaries and install.sh
|
||||
for FILE in dropshell.x86_64 dropshell.aarch64 install.sh server_autosetup.sh; do
|
||||
for FILE in dropshell.x86_64 dropshell.aarch64 dropshell-install.sh dropshell-server-autosetup.sh; do
|
||||
# Pass the locations directly to the find_file function
|
||||
filetoupload=$(find_file "$FILE" "output" "../" ".")
|
||||
if [ -z "$filetoupload" ]; then
|
||||
echo "File $FILE not found in expected locations!" >&2
|
||||
continue
|
||||
fi
|
||||
[ -z "$filetoupload" ] && die "File $FILE not found in expected locations!"
|
||||
|
||||
# Auto-detect content type
|
||||
ctype=$(file --mime-type -b "$filetoupload")
|
||||
|
||||
curl -X POST "$API_URL/releases/$RELEASE_ID/assets?name=$FILE" \
|
||||
-H "Content-Type: $ctype" \
|
||||
-H "Authorization: token $TOKEN" \
|
||||
--data-binary @"$filetoupload"
|
||||
echo "Uploaded $FILE to release $TAG as $ctype."
|
||||
"${TEMP_DIR}/sos" upload getbin.xyz "$filetoupload" "$FILE:latest" "$FILE:TAG"
|
||||
done
|
||||
|
||||
echo "Published dropshell version $TAG to $REPO_URL (tag $TAG) with binaries."
|
||||
|
||||
cd $OLD_PWD
|
||||
echo "Published dropshell $TAG to getbin.xyz"
|
||||
|
@ -103,6 +103,11 @@ namespace dropshell
|
||||
if (server_name.empty() || template_name.empty() || service_name.empty())
|
||||
return false;
|
||||
|
||||
if (!legal_service_name(service_name)) {
|
||||
error << "Service name contains illegal characters: " << service_name << std::endl;
|
||||
return false;
|
||||
}
|
||||
|
||||
ServerConfig server_info(server_name);
|
||||
if (!server_info.is_valid())
|
||||
{
|
||||
|
@ -94,11 +94,11 @@ int edit_config()
|
||||
|
||||
std::string config_file = localfile::dropshell_json();
|
||||
if (!edit_file(config_file, false) || !std::filesystem::exists(config_file))
|
||||
return die("Failed to edit config file.");
|
||||
return return_die("Failed to edit config file.");
|
||||
|
||||
gConfig().load_config();
|
||||
if (!gConfig().is_config_set())
|
||||
return die("Failed to load and parse edited config file!");
|
||||
return return_die("Failed to load and parse edited config file!");
|
||||
|
||||
gConfig().save_config(true);
|
||||
|
||||
@ -130,6 +130,28 @@ int edit_server(const std::string &server_name)
|
||||
return 0;
|
||||
}
|
||||
|
||||
void list_directory(std::string dir, std::string msg)
|
||||
{
|
||||
bool first=true;
|
||||
std::vector<std::string> directories;
|
||||
for (const auto &file : std::filesystem::directory_iterator(dir))
|
||||
{
|
||||
if (first)
|
||||
{
|
||||
if (!msg.empty())
|
||||
info << msg << std::endl;
|
||||
first=false;
|
||||
}
|
||||
if (std::filesystem::is_directory(file.path()))
|
||||
directories.push_back(file.path());
|
||||
else
|
||||
info << " " << file.path() << std::endl;
|
||||
}
|
||||
|
||||
for (const auto &dir : directories)
|
||||
list_directory(dir, "");
|
||||
}
|
||||
|
||||
// ------------------------------------------------------------------------------------------------
|
||||
// edit service config
|
||||
// ------------------------------------------------------------------------------------------------
|
||||
@ -143,7 +165,14 @@ int edit_service_config(const std::string &server, const std::string &service)
|
||||
}
|
||||
|
||||
if (edit_file(config_file, true) && std::filesystem::exists(config_file))
|
||||
info << "To apply your changes, run:\n dropshell install " + server + " " + service << std::endl;
|
||||
info << "Successfully edited service config file at " << config_file << std::endl;
|
||||
|
||||
std::string service_dir = localpath::service(server, service);
|
||||
list_directory(service_dir, "You may wish to edit the other files in " + service_dir);
|
||||
|
||||
info << "Then to apply your changes, run:" << std::endl;
|
||||
info << " dropshell uninstall " + server + " " + service << std::endl;
|
||||
info << " dropshell install " + server + " " + service << std::endl;
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
83
source/src/commands/hash.cpp
Normal file
83
source/src/commands/hash.cpp
Normal file
@ -0,0 +1,83 @@
|
||||
#include "command_registry.hpp"
|
||||
#include "config.hpp"
|
||||
#include "utils/utils.hpp"
|
||||
#include "utils/directories.hpp"
|
||||
#include "shared_commands.hpp"
|
||||
#include "version.hpp"
|
||||
#include "hash.hpp"
|
||||
|
||||
#include <unistd.h>
|
||||
#include <cstring>
|
||||
#include <iostream>
|
||||
#include <sstream>
|
||||
#include <filesystem>
|
||||
#include <libassert/assert.hpp>
|
||||
|
||||
namespace dropshell {
|
||||
|
||||
void hash_autocomplete(const CommandContext& ctx);
|
||||
int hash_handler(const CommandContext& ctx);
|
||||
|
||||
static std::vector<std::string> hash_name_list={"hash"};
|
||||
|
||||
// Static registration
|
||||
struct HashCommandRegister {
|
||||
HashCommandRegister() {
|
||||
CommandRegistry::instance().register_command({
|
||||
hash_name_list,
|
||||
hash_handler,
|
||||
hash_autocomplete,
|
||||
false, // hidden
|
||||
false, // requires_config
|
||||
false, // requires_install
|
||||
0, // min_args (after command)
|
||||
1, // max_args (after command)
|
||||
"hash [FILE|DIRECTORY]",
|
||||
"Hash a file or directory.",
|
||||
// heredoc
|
||||
R"(
|
||||
Hash a file or directory recursively.
|
||||
)"
|
||||
});
|
||||
}
|
||||
} hash_command_register;
|
||||
|
||||
|
||||
void hash_autocomplete(const CommandContext& ctx) {
|
||||
if (ctx.args.size() == 0) {
|
||||
// list all files and directories in the current directory
|
||||
for (const auto& entry : std::filesystem::directory_iterator(".")) {
|
||||
rawout << entry.path().string() << std::endl;
|
||||
}
|
||||
}
|
||||
return;
|
||||
}
|
||||
|
||||
int hash_handler(const CommandContext& ctx) {
|
||||
std::filesystem::path path = safearg(ctx.args, 0);
|
||||
if (path.empty())
|
||||
path=std::filesystem::current_path();
|
||||
|
||||
if (!std::filesystem::exists(path))
|
||||
{
|
||||
error << "Does not exist: " << path.string() << std::endl;
|
||||
return 1;
|
||||
}
|
||||
|
||||
if (std::filesystem::is_directory(path))
|
||||
{
|
||||
// hash the directory recursively
|
||||
uint64_t hash = hash_directory_recursive(path.string());
|
||||
std::cout << hash << std::endl;
|
||||
}
|
||||
else
|
||||
{
|
||||
// hash the file
|
||||
uint64_t hash = hash_file(path.string());
|
||||
std::cout << hash << std::endl;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
} // namespace dropshell
|
@ -69,7 +69,29 @@ namespace dropshell
|
||||
std::string server = server_env.get_server_name();
|
||||
LocalServiceInfo service_info = get_service_info(server_env.get_server_name(), service);
|
||||
|
||||
if (!SIvalid(service_info) || !service_info.service_template_hash_match)
|
||||
if (!SIvalid(service_info))
|
||||
{
|
||||
error << "Failed to install - service information not valid." << std::endl;
|
||||
return false;
|
||||
}
|
||||
|
||||
if (!server_env.is_valid())
|
||||
return false; // should never hit this.
|
||||
|
||||
|
||||
std::string user = service_info.user;
|
||||
std::string remote_service_path = remotepath(server,user).service(service);
|
||||
|
||||
ASSERT(!remote_service_path.empty(), "Install_Service: Remote service path is empty for " + service + " on " + server);
|
||||
ASSERT(!user.empty(), "Install_Service: User is empty for " + service + " on " + server);
|
||||
|
||||
if (server_env.check_remote_dir_exists(remote_service_path, user))
|
||||
{ // uninstall the old service before we update the config or template!
|
||||
info << "Service " << service << " is already installed on " << server << std::endl;
|
||||
shared_commands::uninstall_service(server_env, service);
|
||||
}
|
||||
|
||||
if (!service_info.service_template_hash_match)
|
||||
{
|
||||
warning << "Service " << service << " is using an old template. Updating. " << std::endl;
|
||||
if (!merge_updated_service_template(server_env.get_server_name(), service))
|
||||
@ -88,9 +110,6 @@ namespace dropshell
|
||||
|
||||
maketitle("Installing " + service + " (" + service_info.template_name + ") on " + server);
|
||||
|
||||
if (!server_env.is_valid())
|
||||
return false; // should never hit this.
|
||||
|
||||
// Check if template exists
|
||||
template_info tinfo = gTemplateManager().get_template_info(service_info.template_name);
|
||||
if (!tinfo.is_set())
|
||||
@ -103,8 +122,6 @@ namespace dropshell
|
||||
}
|
||||
|
||||
// Create service directory
|
||||
std::string user = server_env.get_user_for_service(service);
|
||||
std::string remote_service_path = remotepath(server,user).service(service);
|
||||
std::string mkdir_cmd = "mkdir -p " + quote(remote_service_path);
|
||||
if (!execute_ssh_command(server_env.get_SSH_INFO(user), sCommand("", mkdir_cmd, {}), cMode::Silent))
|
||||
{
|
||||
@ -112,14 +129,6 @@ namespace dropshell
|
||||
return false;
|
||||
}
|
||||
|
||||
// Check if rsync is installed on remote host
|
||||
std::string check_rsync_cmd = "which rsync";
|
||||
if (!execute_ssh_command(server_env.get_SSH_INFO(user), sCommand("", check_rsync_cmd, {}), cMode::Silent))
|
||||
{
|
||||
std::cerr << "rsync is not installed on the remote host" << std::endl;
|
||||
return false;
|
||||
}
|
||||
|
||||
// Copy template files
|
||||
debug << "Copying: [LOCAL] " << tinfo.local_template_path() << std::endl
|
||||
<< std::string(8, ' ') << "[REMOTE] " << remotepath(server,user).service_template(service) << "/" << std::endl;
|
||||
@ -143,7 +152,13 @@ namespace dropshell
|
||||
// Run install script
|
||||
{
|
||||
info << "Running " << service_info.template_name << " install script on " << server << "..." << std::endl;
|
||||
server_env.run_remote_template_command(service, "install", {}, false, {});
|
||||
|
||||
shared_commands::cRemoteTempFolder remote_temp_folder(server_env, user);
|
||||
if (!server_env.run_remote_template_command(service, "install", {}, false, {{"TEMP_DIR", remote_temp_folder.path()}}))
|
||||
{
|
||||
error << "Failed to run install script on " << server << std::endl;
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
||||
// print health tick
|
||||
@ -295,7 +310,7 @@ complete -F _dropshell_completions ds
|
||||
std::filesystem::remove(exe_path.parent_path() / "dropshell.old");
|
||||
|
||||
// execute the new version
|
||||
execlp("bash", "bash", "-c", (parent_path / "dropshell").c_str(), "install", (char *)nullptr);
|
||||
execlp("bash", "bash", "-c", (exe_path.parent_path() / "dropshell").string() + "install", (char *)nullptr);
|
||||
error << "Failed to execute new version of dropshell." << std::endl;
|
||||
return -1;
|
||||
}
|
||||
|
@ -75,9 +75,9 @@ namespace dropshell
|
||||
// determine the architecture of the system
|
||||
std::string arch;
|
||||
#ifdef __aarch64__
|
||||
arch = "arm64";
|
||||
arch = "aarch64";
|
||||
#elif __x86_64__
|
||||
arch = "amd64";
|
||||
arch = "x86_64";
|
||||
#endif
|
||||
return arch;
|
||||
}
|
||||
|
@ -61,6 +61,12 @@ namespace dropshell
|
||||
return false;
|
||||
}
|
||||
|
||||
if (!legal_service_name(service))
|
||||
{
|
||||
error << "Service name contains illegal characters: " << service << std::endl;
|
||||
return false;
|
||||
}
|
||||
|
||||
LocalServiceInfo sinfo = get_service_info(server, service);
|
||||
if (!SIvalid(sinfo))
|
||||
{
|
||||
|
@ -51,6 +51,12 @@ namespace dropshell
|
||||
return false;
|
||||
}
|
||||
|
||||
if (!legal_service_name(service))
|
||||
{
|
||||
error << "Service name contains illegal characters: " << service << std::endl;
|
||||
return false;
|
||||
}
|
||||
|
||||
// run the start script.
|
||||
bool started = server_env.run_remote_template_command(service, "start", {}, false, {});
|
||||
|
||||
|
@ -51,6 +51,12 @@ namespace dropshell
|
||||
return false;
|
||||
}
|
||||
|
||||
if (!legal_service_name(service))
|
||||
{
|
||||
error << "Service name contains illegal characters: " << service << std::endl;
|
||||
return false;
|
||||
}
|
||||
|
||||
// run the stop script.
|
||||
bool stopped = server_env.run_remote_template_command(service, "stop", {}, false, {});
|
||||
|
||||
|
@ -6,8 +6,8 @@ _dropshell_completions() {
|
||||
cur="${COMP_WORDS[COMP_CWORD]}"
|
||||
|
||||
# call dropshell to get the list of possiblities for the current argument. Supply all previous arguments.
|
||||
local completions=($(dropshell autocomplete "${COMP_WORDS[@]:1:${COMP_CWORD}-1}"))
|
||||
COMPREPLY=( $(compgen -W "${completions[*]}" -- ${cur}) )
|
||||
mapfile -t completions < <(dropshell autocomplete "${COMP_WORDS[@]:1:${COMP_CWORD}-1}")
|
||||
mapfile -t COMPREPLY < <(compgen -W "${completions[*]}" -- "$cur")
|
||||
return 0
|
||||
}
|
||||
|
||||
|
@ -149,12 +149,20 @@ namespace dropshell
|
||||
|
||||
std::string get_user_for_service(const std::string &server, const std::string &service)
|
||||
{
|
||||
if (!legal_service_name(service))
|
||||
{
|
||||
error << "Service name contains illegal characters: " + service << std::endl;
|
||||
return "";
|
||||
}
|
||||
|
||||
auto services_info = get_server_services_info(server);
|
||||
auto it = std::find_if(services_info.begin(), services_info.end(),
|
||||
[&service](const LocalServiceInfo &si)
|
||||
{ return si.service_name == service; });
|
||||
if (it != services_info.end() && SIvalid(*it))
|
||||
return it->user;
|
||||
|
||||
debug << "Couldn't find user for service \"" << service << "\" on server \"" << server << "\"" << std::endl;
|
||||
return "";
|
||||
}
|
||||
|
||||
@ -345,6 +353,8 @@ namespace dropshell
|
||||
return std::nullopt;
|
||||
}
|
||||
|
||||
env_vars["HOST_NAME"] = get_SSH_HOST();
|
||||
|
||||
std::string argstr = "";
|
||||
for (const auto &arg : args)
|
||||
{
|
||||
|
@ -80,6 +80,9 @@ namespace dropshell
|
||||
if (server_name.empty() || service_name.empty())
|
||||
return LocalServiceInfo();
|
||||
|
||||
if (!legal_service_name(service_name))
|
||||
return LocalServiceInfo();
|
||||
|
||||
service.service_name = service_name;
|
||||
|
||||
service.local_service_path = localpath::service(server_name, service_name);
|
||||
@ -230,6 +233,13 @@ namespace dropshell
|
||||
warning << "Expected environment file not found: " << file << std::endl;
|
||||
};
|
||||
|
||||
|
||||
// add in some simple variables first, as others below may depend on/use these in bash.
|
||||
// if we change these, we also need to update agent/_allservicesstatus.sh
|
||||
all_env_vars["SERVER"] = server_name;
|
||||
all_env_vars["SERVICE"] = service_name;
|
||||
all_env_vars["DOCKER_CLI_HINTS"] = "false"; // turn off docker junk.
|
||||
|
||||
// Load environment files
|
||||
load_env_file(localfile::service_env(server_name, service_name));
|
||||
load_env_file(localfile::template_info_env(server_name, service_name));
|
||||
@ -243,13 +253,10 @@ namespace dropshell
|
||||
return false;
|
||||
}
|
||||
|
||||
// add in some handy variables.
|
||||
// if we change these, we also need to update agent/_allservicesstatus.sh
|
||||
// more additional, these depend on others above.
|
||||
all_env_vars["CONFIG_PATH"] = remotepath(server_name, user).service_config(service_name);
|
||||
all_env_vars["SERVER"] = server_name;
|
||||
all_env_vars["SERVICE"] = service_name;
|
||||
all_env_vars["AGENT_PATH"] = remotepath(server_name, user).agent();
|
||||
all_env_vars["DOCKER_CLI_HINTS"] = "false"; // turn off docker junk.
|
||||
|
||||
|
||||
// determine template name.
|
||||
auto it = all_env_vars.find("TEMPLATE");
|
||||
|
@ -171,6 +171,11 @@
|
||||
|
||||
bool template_manager::create_template(const std::string &template_name) const
|
||||
{
|
||||
if (!legal_service_name(template_name)) {
|
||||
error << "Template name contains illegal characters: " << template_name << std::endl;
|
||||
return false;
|
||||
}
|
||||
|
||||
// 1. Create a new directory in the user templates directory
|
||||
std::vector<std::string> local_server_definition_paths = gConfig().get_local_server_definition_paths();
|
||||
|
||||
|
@ -312,7 +312,7 @@ std::string requote(std::string str) {
|
||||
}
|
||||
|
||||
|
||||
int die(const std::string & msg) {
|
||||
int return_die(const std::string & msg) {
|
||||
error << "Fatal error:" << std::endl;
|
||||
error << msg << std::endl;
|
||||
return 1;
|
||||
@ -650,4 +650,24 @@ bool file_replace_or_add_segment(std::string filepath, std::string segment)
|
||||
return true;
|
||||
}
|
||||
|
||||
bool legal_service_name(const std::string &service_name) {
|
||||
static bool initialized = false;
|
||||
static bool legal_chars[256] = {false}; // Initialize all to false
|
||||
|
||||
// One-time initialization
|
||||
if (!initialized) {
|
||||
// Set true for valid characters
|
||||
for (unsigned char c : "0123456789"
|
||||
"ABCDEFGHIJKLMNOPQRSTUVWXYZ"
|
||||
"abcdefghijklmnopqrstuvwxyz"
|
||||
"._-") {
|
||||
legal_chars[c] = true;
|
||||
}
|
||||
initialized = true;
|
||||
}
|
||||
|
||||
return std::all_of(service_name.begin(), service_name.end(),
|
||||
[](unsigned char c) { return legal_chars[c]; });
|
||||
}
|
||||
|
||||
} // namespace dropshell
|
@ -45,7 +45,7 @@ int count_substring(const std::string &substring, const std::string &text);
|
||||
|
||||
std::string random_alphanumeric_string(int length);
|
||||
|
||||
int die(const std::string & msg);
|
||||
int return_die(const std::string & msg);
|
||||
std::string safearg(int argc, char *argv[], int index);
|
||||
std::string safearg(const std::vector<std::string> & args, int index);
|
||||
|
||||
@ -75,4 +75,7 @@ constexpr unsigned int switchhash(const char *s, int off = 0)
|
||||
return !s[off] ? 5381 : (switchhash(s, off + 1) * 33) ^ s[off];
|
||||
}
|
||||
|
||||
bool legal_service_name(const std::string & service_name);
|
||||
|
||||
|
||||
} // namespace dropshell
|
12
source/test.sh
Executable file
12
source/test.sh
Executable file
@ -0,0 +1,12 @@
|
||||
#!/bin/bash
|
||||
set -euo pipefail
|
||||
SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" &> /dev/null && pwd )"
|
||||
|
||||
ARCH=$(uname -m)
|
||||
|
||||
PREV_DIR=$(pwd)
|
||||
trap 'cd "$PREV_DIR"' EXIT
|
||||
|
||||
"$SCRIPT_DIR/output/dropshell.${ARCH}" hash "${SCRIPT_DIR}/test.sh"
|
||||
"$SCRIPT_DIR/output/dropshell.${ARCH}" help
|
||||
|
Loading…
x
Reference in New Issue
Block a user