Compare commits
43 Commits
2025.0524.
...
main
Author | SHA1 | Date | |
---|---|---|---|
![]() |
9d4e5f76ce | ||
![]() |
366f5c2d0e | ||
![]() |
0b0f3df59c | ||
![]() |
f48302c05e | ||
![]() |
7f341699c1 | ||
![]() |
18c53acd71 | ||
![]() |
eb632c010c | ||
![]() |
964e8598b1 | ||
![]() |
a5cf9313e9 | ||
![]() |
ab73a47751 | ||
![]() |
1da7dc7951 | ||
![]() |
49d61f0da0 | ||
![]() |
27c0abcb9f | ||
![]() |
483ee4e3ef | ||
![]() |
f7294e01e4 | ||
![]() |
c836b26657 | ||
![]() |
7bf624589f | ||
![]() |
a5e339a358 | ||
![]() |
029823a6b4 | ||
![]() |
f79abd346e | ||
![]() |
940c2a12a1 | ||
![]() |
6ac651d4f0 | ||
![]() |
84fd96e74e | ||
![]() |
8eb652b84e | ||
![]() |
c8c0c3289c | ||
![]() |
3dc82c682c | ||
![]() |
4035538ba5 | ||
![]() |
8ec4976cc0 | ||
![]() |
1b35f74bfe | ||
![]() |
08794e6480 | ||
![]() |
3b51a511a6 | ||
![]() |
e1be3dff8d | ||
![]() |
9375acafa9 | ||
![]() |
f45baa8362 | ||
![]() |
d3ceb3f4be | ||
![]() |
27a2d25fb2 | ||
![]() |
e7558be416 | ||
![]() |
8f06fc31ae | ||
![]() |
1502d6e3d2 | ||
![]() |
d71ba38754 | ||
![]() |
f45d9a33ed | ||
![]() |
73f85769a2 | ||
![]() |
9a141685de |
@ -12,12 +12,28 @@ jobs:
|
|||||||
sudo apt-get install -y openssh-server
|
sudo apt-get install -y openssh-server
|
||||||
- name: Check out repository code
|
- name: Check out repository code
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
- name: Build
|
- name: Install build dependencies
|
||||||
run: |
|
run: |
|
||||||
cd ${{ gitea.workspace }}/docker
|
cd source
|
||||||
./compile.sh
|
./install_build_prerequisites.sh
|
||||||
|
- name: Build Native
|
||||||
|
run: |
|
||||||
|
cd source
|
||||||
|
./build_native.sh
|
||||||
- name: Test
|
- name: Test
|
||||||
run: |
|
run: |
|
||||||
cd ${{ gitea.workspace }}/docker/output
|
cd source
|
||||||
./dropshell_x86_64 list
|
./test.sh
|
||||||
./dropshell_x86_64 help
|
- name: Build Production
|
||||||
|
run: |
|
||||||
|
cd source
|
||||||
|
./build_production.sh
|
||||||
|
- name: Test
|
||||||
|
run: |
|
||||||
|
cd source
|
||||||
|
./test.sh
|
||||||
|
- name: Publish
|
||||||
|
run: |
|
||||||
|
cd source
|
||||||
|
./publish.sh
|
||||||
|
|
||||||
|
1
.gitignore
vendored
1
.gitignore
vendored
@ -39,7 +39,6 @@ cmake_install.cmake
|
|||||||
Makefile
|
Makefile
|
||||||
|
|
||||||
# IDE specific files
|
# IDE specific files
|
||||||
.vscode/
|
|
||||||
.idea/
|
.idea/
|
||||||
*.swp
|
*.swp
|
||||||
*.swo
|
*.swo
|
||||||
|
17
.vscode/c_cpp_properties.json
vendored
Normal file
17
.vscode/c_cpp_properties.json
vendored
Normal file
@ -0,0 +1,17 @@
|
|||||||
|
{
|
||||||
|
"configurations": [
|
||||||
|
{
|
||||||
|
"name": "Linux",
|
||||||
|
"includePath": [
|
||||||
|
"${workspaceFolder}/**",
|
||||||
|
"${workspaceFolder}/source/build/src/autogen"
|
||||||
|
],
|
||||||
|
"defines": [],
|
||||||
|
"compilerPath": "/usr/bin/g++",
|
||||||
|
"cStandard": "c23",
|
||||||
|
"cppStandard": "c++23",
|
||||||
|
"intelliSenseMode": "linux-gcc-x64"
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"version": 4
|
||||||
|
}
|
103
.vscode/settings.json
vendored
Normal file
103
.vscode/settings.json
vendored
Normal file
@ -0,0 +1,103 @@
|
|||||||
|
{
|
||||||
|
"files.associations": {
|
||||||
|
"*.inja": "jinja-html",
|
||||||
|
"*.ipp": "cpp",
|
||||||
|
"random": "cpp",
|
||||||
|
"ostream": "cpp",
|
||||||
|
"cctype": "cpp",
|
||||||
|
"clocale": "cpp",
|
||||||
|
"cmath": "cpp",
|
||||||
|
"csetjmp": "cpp",
|
||||||
|
"csignal": "cpp",
|
||||||
|
"cstdarg": "cpp",
|
||||||
|
"cstddef": "cpp",
|
||||||
|
"cstdio": "cpp",
|
||||||
|
"cstdlib": "cpp",
|
||||||
|
"cstring": "cpp",
|
||||||
|
"ctime": "cpp",
|
||||||
|
"cwchar": "cpp",
|
||||||
|
"cwctype": "cpp",
|
||||||
|
"any": "cpp",
|
||||||
|
"array": "cpp",
|
||||||
|
"atomic": "cpp",
|
||||||
|
"strstream": "cpp",
|
||||||
|
"bit": "cpp",
|
||||||
|
"*.tcc": "cpp",
|
||||||
|
"bitset": "cpp",
|
||||||
|
"cfenv": "cpp",
|
||||||
|
"charconv": "cpp",
|
||||||
|
"chrono": "cpp",
|
||||||
|
"cinttypes": "cpp",
|
||||||
|
"codecvt": "cpp",
|
||||||
|
"compare": "cpp",
|
||||||
|
"complex": "cpp",
|
||||||
|
"concepts": "cpp",
|
||||||
|
"condition_variable": "cpp",
|
||||||
|
"cstdint": "cpp",
|
||||||
|
"deque": "cpp",
|
||||||
|
"forward_list": "cpp",
|
||||||
|
"list": "cpp",
|
||||||
|
"map": "cpp",
|
||||||
|
"set": "cpp",
|
||||||
|
"string": "cpp",
|
||||||
|
"unordered_map": "cpp",
|
||||||
|
"unordered_set": "cpp",
|
||||||
|
"vector": "cpp",
|
||||||
|
"exception": "cpp",
|
||||||
|
"expected": "cpp",
|
||||||
|
"algorithm": "cpp",
|
||||||
|
"functional": "cpp",
|
||||||
|
"iterator": "cpp",
|
||||||
|
"memory": "cpp",
|
||||||
|
"memory_resource": "cpp",
|
||||||
|
"numeric": "cpp",
|
||||||
|
"optional": "cpp",
|
||||||
|
"ratio": "cpp",
|
||||||
|
"regex": "cpp",
|
||||||
|
"source_location": "cpp",
|
||||||
|
"string_view": "cpp",
|
||||||
|
"system_error": "cpp",
|
||||||
|
"tuple": "cpp",
|
||||||
|
"type_traits": "cpp",
|
||||||
|
"utility": "cpp",
|
||||||
|
"fstream": "cpp",
|
||||||
|
"future": "cpp",
|
||||||
|
"initializer_list": "cpp",
|
||||||
|
"iomanip": "cpp",
|
||||||
|
"iosfwd": "cpp",
|
||||||
|
"iostream": "cpp",
|
||||||
|
"istream": "cpp",
|
||||||
|
"limits": "cpp",
|
||||||
|
"mutex": "cpp",
|
||||||
|
"new": "cpp",
|
||||||
|
"numbers": "cpp",
|
||||||
|
"ranges": "cpp",
|
||||||
|
"scoped_allocator": "cpp",
|
||||||
|
"semaphore": "cpp",
|
||||||
|
"shared_mutex": "cpp",
|
||||||
|
"span": "cpp",
|
||||||
|
"sstream": "cpp",
|
||||||
|
"stdexcept": "cpp",
|
||||||
|
"stop_token": "cpp",
|
||||||
|
"streambuf": "cpp",
|
||||||
|
"thread": "cpp",
|
||||||
|
"typeindex": "cpp",
|
||||||
|
"typeinfo": "cpp",
|
||||||
|
"valarray": "cpp",
|
||||||
|
"variant": "cpp",
|
||||||
|
"format": "cpp",
|
||||||
|
"stdfloat": "cpp",
|
||||||
|
"__nullptr": "cpp",
|
||||||
|
"__node_handle": "cpp",
|
||||||
|
"__split_buffer": "cpp",
|
||||||
|
"filesystem": "cpp",
|
||||||
|
"queue": "cpp",
|
||||||
|
"stack": "cpp",
|
||||||
|
"__bit_reference": "cpp",
|
||||||
|
"__functional_base": "cpp",
|
||||||
|
"__memory": "cpp",
|
||||||
|
"locale": "cpp",
|
||||||
|
"stacktrace": "cpp",
|
||||||
|
"__locale": "cpp"
|
||||||
|
}
|
||||||
|
}
|
36
README.md
36
README.md
@ -5,15 +5,23 @@ A system management tool for server operations, written in C++.
|
|||||||
## Installation
|
## Installation
|
||||||
|
|
||||||
```
|
```
|
||||||
curl -fsSL https://gitea.jde.nz/public/dropshell/releases/download/latest/install.sh | sudo bash
|
curl -fsSL https://gitea.jde.nz/public/dropshell/releases/download/latest/install.sh | bash
|
||||||
```
|
```
|
||||||
|
|
||||||
This installs as dropshell, with a symlink ds if the ds command does not already exist.
|
This installs as dropshell for the local user, with a symbolic link ds.
|
||||||
|
You'll need to run:
|
||||||
|
```
|
||||||
|
~/.local/bin/dropshell edit
|
||||||
|
~/.local/bin/dropshell install
|
||||||
|
source ~/.bashrc
|
||||||
|
```
|
||||||
|
to configure dropshell and install the local components.
|
||||||
|
|
||||||
## Installation of Agent
|
## Remote Server Setup
|
||||||
|
|
||||||
Install the Agent on each server you wish to manage. Supports amd64 (x86_64) and arm64 (aarch64) architectures.
|
### Initial setup
|
||||||
|
|
||||||
|
Auto setup script which creates a dropshell user, and includes installing docker if not already present:
|
||||||
```
|
```
|
||||||
curl -fsSL https://gitea.jde.nz/public/dropshell/releases/download/latest/server_autosetup.sh | sudo bash
|
curl -fsSL https://gitea.jde.nz/public/dropshell/releases/download/latest/server_autosetup.sh | sudo bash
|
||||||
```
|
```
|
||||||
@ -26,7 +34,21 @@ Manual steps:
|
|||||||
1. Test ssh'ing into the server.
|
1. Test ssh'ing into the server.
|
||||||
|
|
||||||
|
|
||||||
## Install Services
|
### Configure and Use Remote Server
|
||||||
|
|
||||||
Set up a server and install a service:
|
#### Add to local dropshell configuration, and install remote agent
|
||||||
1. `ds create-server SERVERNAME`
|
Back on the dropshell host:
|
||||||
|
1. `dropshell create-server SERVERNAME`
|
||||||
|
1. `dropshell edit SERVERNAME`
|
||||||
|
1. `dropshell install SERVERNAME`
|
||||||
|
|
||||||
|
#### Install Services
|
||||||
|
Create and install a service
|
||||||
|
1. `ds template list` -- see what templates are available to install.
|
||||||
|
1. `ds create-service SERVERNAME SERVICENAME TEMPLATE`
|
||||||
|
1. `ds edit SERVERNAME SERVICENAME`
|
||||||
|
1. Edit other config files if needed.
|
||||||
|
1. `ds install SERVERNAME SERVICENAME`
|
||||||
|
1. `ds list`
|
||||||
|
|
||||||
|
The service should now be seen to be running.
|
||||||
|
22
dropshell-install.sh
Executable file
22
dropshell-install.sh
Executable file
@ -0,0 +1,22 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -euo pipefail
|
||||||
|
|
||||||
|
# download and install dropshell
|
||||||
|
|
||||||
|
# 1. Determine architecture
|
||||||
|
# -----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
ARCH=$(uname -m)
|
||||||
|
|
||||||
|
TARGET_PATH="${HOME}/.local/bin/dropshell"
|
||||||
|
[ ! -f "${TARGET_PATH}" ] || rm -f "${TARGET_PATH}"
|
||||||
|
mkdir -p "$(dirname "${TARGET_PATH}")"
|
||||||
|
curl -L -s -o "${TARGET_PATH}" "https://getbin.xyz/dropshell.${ARCH}" || die "Failed to download dropshell for ${ARCH}"
|
||||||
|
chmod +x "${TARGET_PATH}"
|
||||||
|
echo "dropshell installed successfully to $TARGET_PATH"
|
||||||
|
echo " "
|
||||||
|
|
||||||
|
echo "Please:"
|
||||||
|
echo "1. run '${TARGET_PATH} edit' to edit the configuration."
|
||||||
|
echo "2. run '${TARGET_PATH} install' to install dropshell components on this computer."
|
||||||
|
echo "3. run 'source ~/.bashrc' to add to your path and autocomplete for the current shell."
|
@ -1,4 +1,6 @@
|
|||||||
# can you make this script run in bash, but fall back to sh if bash is not installed?
|
#!/bin/bash
|
||||||
|
|
||||||
|
# set up a remote server for use with dropshell.
|
||||||
|
|
||||||
# check if we are running as root
|
# check if we are running as root
|
||||||
if [ "$(id -u)" -ne 0 ]; then
|
if [ "$(id -u)" -ne 0 ]; then
|
50
install.sh
50
install.sh
@ -1,50 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
set -e
|
|
||||||
|
|
||||||
# download and install dropshell
|
|
||||||
|
|
||||||
# Check if running as root
|
|
||||||
if [ "$EUID" -ne 0 ]; then
|
|
||||||
echo "Please run this script as root (use sudo)"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
# 1. Determine architecture
|
|
||||||
# -----------------------------------------------------------------------------
|
|
||||||
|
|
||||||
ARCH=$(uname -m)
|
|
||||||
if [[ "$ARCH" == "x86_64" ]]; then
|
|
||||||
BIN=dropshell.amd64
|
|
||||||
elif [[ "$ARCH" == "aarch64" || "$ARCH" == "arm64" ]]; then
|
|
||||||
BIN=dropshell.arm64
|
|
||||||
else
|
|
||||||
echo "Unsupported architecture: $ARCH" >&2
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
|
|
||||||
# 2. Download the appropriate binary to a temp directory
|
|
||||||
# -----------------------------------------------------------------------------
|
|
||||||
TMPDIR=$(mktemp -d)
|
|
||||||
trap 'rm -rf "$TMPDIR"' EXIT
|
|
||||||
URL="https://gitea.jde.nz/public/dropshell/releases/download/latest/$BIN"
|
|
||||||
echo "Downloading $BIN from $URL..."
|
|
||||||
|
|
||||||
curl -fsSL -o "$TMPDIR/dropshell" "$URL"
|
|
||||||
|
|
||||||
if [ ! -f "$TMPDIR/dropshell" ]; then
|
|
||||||
echo "Failed to download dropshell" >&2
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
chmod +x "$TMPDIR/dropshell"
|
|
||||||
|
|
||||||
cp "$TMPDIR/dropshell" /usr/local/bin/dropshell
|
|
||||||
if [ -f /usr/local/bin/ds ]; then
|
|
||||||
rm -f /usr/local/bin/ds
|
|
||||||
fi
|
|
||||||
ln -s /usr/local/bin/dropshell /usr/local/bin/ds
|
|
||||||
rm -rf "$TMPDIR"
|
|
||||||
|
|
||||||
echo "dropshell installed successfully to /usr/local/bin/dropshell"
|
|
||||||
|
|
@ -1,6 +1,15 @@
|
|||||||
cmake_minimum_required(VERSION 3.10)
|
cmake_minimum_required(VERSION 3.10)
|
||||||
project(dropshell VERSION 1.0.0 LANGUAGES CXX)
|
project(dropshell VERSION 1.0.0 LANGUAGES CXX)
|
||||||
|
|
||||||
|
# Force static linking globally
|
||||||
|
set(CMAKE_EXE_LINKER_FLAGS "${CMAKE_EXE_LINKER_FLAGS} -static")
|
||||||
|
set(CMAKE_FIND_LIBRARY_SUFFIXES ".a")
|
||||||
|
set(BUILD_SHARED_LIBS OFF CACHE BOOL "Build shared libraries" FORCE)
|
||||||
|
set(CMAKE_POSITION_INDEPENDENT_CODE OFF)
|
||||||
|
set(CMAKE_C_FLAGS "${CMAKE_C_FLAGS} -static")
|
||||||
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -static")
|
||||||
|
set(ZLIB_USE_STATIC_LIBS "ON")
|
||||||
|
|
||||||
set(CMAKE_CXX_STANDARD 23)
|
set(CMAKE_CXX_STANDARD 23)
|
||||||
set(CMAKE_C_STANDARD 23)
|
set(CMAKE_C_STANDARD 23)
|
||||||
set(CMAKE_CXX_STANDARD_REQUIRED ON)
|
set(CMAKE_CXX_STANDARD_REQUIRED ON)
|
||||||
@ -37,17 +46,16 @@ set(CMAKE_MODULE_PATH ${CMAKE_MODULE_PATH} ${CMAKE_CURRENT_SOURCE_DIR}/cmake)
|
|||||||
file(GLOB_RECURSE SOURCES "src/*.cpp")
|
file(GLOB_RECURSE SOURCES "src/*.cpp")
|
||||||
file(GLOB_RECURSE HEADERS "src/*.hpp")
|
file(GLOB_RECURSE HEADERS "src/*.hpp")
|
||||||
|
|
||||||
# Add custom target to run make_createagent.sh at the start of the build process
|
# Add custom target to run cmake_prebuild.sh at the start of the build process
|
||||||
add_custom_target(run_createagent ALL
|
add_custom_target(run_prebuild_script ALL
|
||||||
COMMAND ${CMAKE_COMMAND} -E echo "Running make_createagent.sh..."
|
COMMAND ${CMAKE_COMMAND} -E echo "Running cmake_prebuild.sh..."
|
||||||
COMMAND ${CMAKE_COMMAND} -E env bash ${CMAKE_CURRENT_SOURCE_DIR}/make_createagent.sh
|
COMMAND ${CMAKE_COMMAND} -E env bash ${CMAKE_CURRENT_SOURCE_DIR}/cmake_prebuild.sh
|
||||||
WORKING_DIRECTORY ${CMAKE_CURRENT_SOURCE_DIR}
|
WORKING_DIRECTORY ${CMAKE_CURRENT_SOURCE_DIR}
|
||||||
COMMENT "Running make_createagent.sh before build"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
# Add executable
|
# Add executable
|
||||||
add_executable(dropshell ${SOURCES})
|
add_executable(dropshell ${SOURCES})
|
||||||
add_dependencies(dropshell run_createagent)
|
add_dependencies(dropshell run_prebuild_script)
|
||||||
|
|
||||||
# Mark the generated files as GENERATED so CMake knows they'll be created during build
|
# Mark the generated files as GENERATED so CMake knows they'll be created during build
|
||||||
set_source_files_properties(
|
set_source_files_properties(
|
||||||
@ -89,7 +97,8 @@ FetchContent_Declare(
|
|||||||
GIT_TAG v2.1.5
|
GIT_TAG v2.1.5
|
||||||
)
|
)
|
||||||
FetchContent_MakeAvailable(libassert)
|
FetchContent_MakeAvailable(libassert)
|
||||||
include(FetchContent)
|
|
||||||
|
# Add cpptrace
|
||||||
FetchContent_Declare(
|
FetchContent_Declare(
|
||||||
cpptrace
|
cpptrace
|
||||||
GIT_REPOSITORY https://github.com/jeremy-rifkin/cpptrace.git
|
GIT_REPOSITORY https://github.com/jeremy-rifkin/cpptrace.git
|
||||||
@ -97,81 +106,28 @@ FetchContent_Declare(
|
|||||||
)
|
)
|
||||||
FetchContent_MakeAvailable(cpptrace)
|
FetchContent_MakeAvailable(cpptrace)
|
||||||
|
|
||||||
|
# Add nlohmann/json
|
||||||
|
FetchContent_Declare(
|
||||||
|
nlohmann_json
|
||||||
|
GIT_REPOSITORY https://github.com/nlohmann/json.git
|
||||||
|
GIT_TAG v3.11.3
|
||||||
|
)
|
||||||
|
FetchContent_MakeAvailable(nlohmann_json)
|
||||||
|
|
||||||
# Link libraries
|
# Link libraries
|
||||||
target_link_libraries(dropshell PRIVATE
|
target_link_libraries(dropshell PRIVATE
|
||||||
libassert::assert
|
libassert::assert
|
||||||
cpptrace::cpptrace
|
cpptrace::cpptrace
|
||||||
|
nlohmann_json::nlohmann_json
|
||||||
|
)
|
||||||
|
|
||||||
|
# Set static linking flags
|
||||||
|
set_target_properties(dropshell PROPERTIES
|
||||||
|
LINK_FLAGS "-static"
|
||||||
)
|
)
|
||||||
|
|
||||||
# Install targets
|
# Install targets
|
||||||
install(TARGETS dropshell
|
install(TARGETS dropshell
|
||||||
RUNTIME DESTINATION bin
|
RUNTIME DESTINATION $ENV{HOME}/.local/bin
|
||||||
)
|
)
|
||||||
|
|
||||||
# Create symbolic link 'ds' pointing to 'dropshell'
|
|
||||||
install(CODE "
|
|
||||||
message(STATUS \"Checking if 'ds' command already exists...\")
|
|
||||||
execute_process(
|
|
||||||
COMMAND which ds
|
|
||||||
RESULT_VARIABLE DS_NOT_EXISTS
|
|
||||||
OUTPUT_QUIET
|
|
||||||
ERROR_QUIET
|
|
||||||
)
|
|
||||||
if(DS_NOT_EXISTS)
|
|
||||||
message(STATUS \"Command 'ds' does not exist. Creating symlink.\")
|
|
||||||
execute_process(
|
|
||||||
COMMAND ${CMAKE_COMMAND} -E create_symlink
|
|
||||||
\${CMAKE_INSTALL_PREFIX}/bin/dropshell
|
|
||||||
\${CMAKE_INSTALL_PREFIX}/bin/ds
|
|
||||||
)
|
|
||||||
else()
|
|
||||||
message(STATUS \"Command 'ds' already exists. Skipping symlink creation.\")
|
|
||||||
endif()
|
|
||||||
")
|
|
||||||
|
|
||||||
# Install completion script
|
|
||||||
install(FILES src/dropshell-completion.bash
|
|
||||||
DESTINATION /etc/bash_completion.d
|
|
||||||
RENAME dropshell
|
|
||||||
)
|
|
||||||
|
|
||||||
# Create a symlink for the completion script to work with 'ds' command
|
|
||||||
install(CODE "
|
|
||||||
# First check if 'ds' command exists after our installation
|
|
||||||
execute_process(
|
|
||||||
COMMAND which ds
|
|
||||||
RESULT_VARIABLE DS_NOT_EXISTS
|
|
||||||
OUTPUT_VARIABLE DS_PATH
|
|
||||||
ERROR_QUIET
|
|
||||||
OUTPUT_STRIP_TRAILING_WHITESPACE
|
|
||||||
)
|
|
||||||
|
|
||||||
# Only proceed if 'ds' exists
|
|
||||||
if(NOT DS_NOT_EXISTS)
|
|
||||||
# Check if 'ds' is a symlink pointing to dropshell
|
|
||||||
execute_process(
|
|
||||||
COMMAND readlink -f \${DS_PATH}
|
|
||||||
RESULT_VARIABLE READLINK_FAILED
|
|
||||||
OUTPUT_VARIABLE REAL_PATH
|
|
||||||
ERROR_QUIET
|
|
||||||
OUTPUT_STRIP_TRAILING_WHITESPACE
|
|
||||||
)
|
|
||||||
|
|
||||||
# Get the path to our dropshell binary
|
|
||||||
set(DROPSHELL_PATH \${CMAKE_INSTALL_PREFIX}/bin/dropshell)
|
|
||||||
|
|
||||||
# Check if the real path is our dropshell binary
|
|
||||||
if(NOT READLINK_FAILED AND \"\${REAL_PATH}\" STREQUAL \"\${DROPSHELL_PATH}\")
|
|
||||||
message(STATUS \"Command 'ds' exists and points to dropshell. Creating completion script symlink.\")
|
|
||||||
execute_process(
|
|
||||||
COMMAND ${CMAKE_COMMAND} -E create_symlink
|
|
||||||
/etc/bash_completion.d/dropshell
|
|
||||||
/etc/bash_completion.d/ds
|
|
||||||
)
|
|
||||||
else()
|
|
||||||
message(STATUS \"Command 'ds' exists but doesn't point to dropshell. Skipping completion symlink.\")
|
|
||||||
endif()
|
|
||||||
else()
|
|
||||||
message(STATUS \"Command 'ds' not found. Skipping completion symlink.\")
|
|
||||||
endif()
|
|
||||||
")
|
|
||||||
|
65
source/Dockerfile
Normal file
65
source/Dockerfile
Normal file
@ -0,0 +1,65 @@
|
|||||||
|
FROM --platform=$BUILDPLATFORM alpine:latest AS builder
|
||||||
|
|
||||||
|
# Add build arguments for platform
|
||||||
|
ARG TARGETPLATFORM
|
||||||
|
ARG BUILDPLATFORM
|
||||||
|
|
||||||
|
# Install build dependencies
|
||||||
|
RUN apk add --no-cache \
|
||||||
|
build-base \
|
||||||
|
cmake \
|
||||||
|
git \
|
||||||
|
musl-dev \
|
||||||
|
curl \
|
||||||
|
bash \
|
||||||
|
musl \
|
||||||
|
g++ \
|
||||||
|
ninja \
|
||||||
|
linux-headers
|
||||||
|
|
||||||
|
# Install cross-compilation tools for ARM64
|
||||||
|
RUN if [ "$TARGETPLATFORM" = "linux/arm64" ]; then \
|
||||||
|
apk add --no-cache \
|
||||||
|
crossbuild-essential-arm64 \
|
||||||
|
gcc-aarch64-linux-gnu \
|
||||||
|
g++-aarch64-linux-gnu; \
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Set working directory
|
||||||
|
WORKDIR /build
|
||||||
|
|
||||||
|
# Copy source files
|
||||||
|
COPY . .
|
||||||
|
|
||||||
|
# Configure and build
|
||||||
|
RUN mkdir -p build_static
|
||||||
|
|
||||||
|
# Set up cross-compilation environment for ARM64
|
||||||
|
RUN if [ "$TARGETPLATFORM" = "linux/arm64" ]; then \
|
||||||
|
export CC=aarch64-linux-gnu-gcc \
|
||||||
|
export CXX=aarch64-linux-gnu-g++ \
|
||||||
|
export CMAKE_TOOLCHAIN_FILE=/build/toolchain.cmake; \
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Create toolchain file for ARM64
|
||||||
|
RUN if [ "$TARGETPLATFORM" = "linux/arm64" ]; then \
|
||||||
|
echo "set(CMAKE_SYSTEM_NAME Linux)" > toolchain.cmake && \
|
||||||
|
echo "set(CMAKE_SYSTEM_PROCESSOR aarch64)" >> toolchain.cmake && \
|
||||||
|
echo "set(CMAKE_C_COMPILER aarch64-linux-gnu-gcc)" >> toolchain.cmake && \
|
||||||
|
echo "set(CMAKE_CXX_COMPILER aarch64-linux-gnu-g++)" >> toolchain.cmake && \
|
||||||
|
echo "set(CMAKE_FIND_ROOT_PATH_MODE_PROGRAM NEVER)" >> toolchain.cmake && \
|
||||||
|
echo "set(CMAKE_FIND_ROOT_PATH_MODE_LIBRARY ONLY)" >> toolchain.cmake && \
|
||||||
|
echo "set(CMAKE_FIND_ROOT_PATH_MODE_INCLUDE ONLY)" >> toolchain.cmake; \
|
||||||
|
fi
|
||||||
|
|
||||||
|
RUN cmake -G Ninja -B build_static -DCMAKE_BUILD_TYPE=Release \
|
||||||
|
-DCMAKE_EXE_LINKER_FLAGS="-static" \
|
||||||
|
-DCMAKE_FIND_LIBRARY_SUFFIXES=".a" \
|
||||||
|
-DBUILD_SHARED_LIBS=OFF \
|
||||||
|
${CMAKE_TOOLCHAIN_FILE:+-DCMAKE_TOOLCHAIN_FILE=$CMAKE_TOOLCHAIN_FILE}
|
||||||
|
|
||||||
|
RUN cmake --build build_static
|
||||||
|
|
||||||
|
# Final stage that only contains the binary
|
||||||
|
FROM scratch AS dropshell
|
||||||
|
COPY --from=builder /build/build_static/dropshell /dropshell
|
@ -1,5 +1,8 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
|
|
||||||
|
# install the dropshell host agent on this computer.
|
||||||
|
# (not for remote servers)
|
||||||
|
|
||||||
SCRIPT_DIR=$(dirname "$0")
|
SCRIPT_DIR=$(dirname "$0")
|
||||||
|
|
||||||
echo "Installing dropshell host agent on this computer..."
|
echo "Installing dropshell host agent on this computer..."
|
||||||
@ -22,27 +25,6 @@ _check_required_env_vars() {
|
|||||||
done
|
done
|
||||||
}
|
}
|
||||||
|
|
||||||
# Checks if Docker is installed, running, and user has permission. Returns 1 on failure.
|
|
||||||
_check_docker_installed() {
|
|
||||||
if ! command -v docker &> /dev/null; then
|
|
||||||
echo "Docker is not installed"
|
|
||||||
return 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
# check if docker daemon is running
|
|
||||||
if ! docker info &> /dev/null; then
|
|
||||||
echo "Docker daemon is not running"
|
|
||||||
return 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
# check if user has permission to run docker
|
|
||||||
if ! docker run --rm hello-world &> /dev/null; then
|
|
||||||
echo "User does not have permission to run docker"
|
|
||||||
return 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
return 0
|
|
||||||
}
|
|
||||||
|
|
||||||
function install_bb64() {
|
function install_bb64() {
|
||||||
# check curl installed
|
# check curl installed
|
||||||
@ -77,7 +59,5 @@ set +a
|
|||||||
_check_required_env_vars "AGENT_LOCAL_PATH"
|
_check_required_env_vars "AGENT_LOCAL_PATH"
|
||||||
echo "Installing host agent into $AGENT_LOCAL_PATH"
|
echo "Installing host agent into $AGENT_LOCAL_PATH"
|
||||||
|
|
||||||
_check_docker_installed || _die "Docker is required."
|
|
||||||
|
|
||||||
install_bb64
|
install_bb64
|
||||||
|
|
||||||
|
@ -38,6 +38,7 @@ CURRENT_EXIT_CODE=0
|
|||||||
load_dotenv(){
|
load_dotenv(){
|
||||||
local file_path=$1
|
local file_path=$1
|
||||||
if [ -f "${file_path}" ]; then
|
if [ -f "${file_path}" ]; then
|
||||||
|
# shellcheck source=/dev/null
|
||||||
source "${file_path}"
|
source "${file_path}"
|
||||||
fi
|
fi
|
||||||
}
|
}
|
||||||
@ -69,9 +70,9 @@ function run_command() {
|
|||||||
load_dotenv "${service_path}/config/.template_info.env"
|
load_dotenv "${service_path}/config/.template_info.env"
|
||||||
|
|
||||||
# update the main variables.
|
# update the main variables.
|
||||||
CONFIG_PATH="${service_path}/config"
|
export CONFIG_PATH="${service_path}/config"
|
||||||
SERVICE="${SERVICE_NAME}"
|
export SERVICE="${SERVICE_NAME}"
|
||||||
DOCKER_CLI_HINTS=false
|
export DOCKER_CLI_HINTS=false
|
||||||
|
|
||||||
set +a
|
set +a
|
||||||
|
|
||||||
|
@ -21,16 +21,13 @@ fi
|
|||||||
_check_required_env_vars "AGENT_PATH"
|
_check_required_env_vars "AGENT_PATH"
|
||||||
|
|
||||||
function install_bb64() {
|
function install_bb64() {
|
||||||
curl -fsSL "https://gitea.jde.nz/public/bb64/releases/download/latest/install.sh" | bash -s -- "$AGENT_PATH" "$(id -u $USER):$(id -g $USER)"
|
if ! curl -fsSL "https://gitea.jde.nz/public/bb64/releases/download/latest/install.sh" | \
|
||||||
|
bash -s -- "$AGENT_PATH" "$(id -u "$USER"):$(id -g "$USER")"; then
|
||||||
# test result code from curl
|
|
||||||
if [ $? -ne 0 ]; then
|
|
||||||
_die "Failed to install bb64. Curl returned non-zero exit code."
|
_die "Failed to install bb64. Curl returned non-zero exit code."
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# test if bb64 is installed
|
# test if bb64 is installed
|
||||||
VER=$("$AGENT_PATH/bb64" -v)
|
if ! VER=$("$AGENT_PATH/bb64" -v); then
|
||||||
if [ $? -ne 0 ]; then
|
|
||||||
_die "bb64 did not install correctly."
|
_die "bb64 did not install correctly."
|
||||||
fi
|
fi
|
||||||
|
|
||||||
@ -53,6 +50,12 @@ if ! command -v docker &> /dev/null; then
|
|||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
# check rsync installation
|
||||||
|
if ! command -v rsync &> /dev/null; then
|
||||||
|
echo "Rsync is not installed. Rsync is required for agent installation."
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------
|
#-------------------------------------------------------------------------
|
||||||
|
|
||||||
|
@ -41,18 +41,18 @@ _create_and_start_container() {
|
|||||||
local run_cmd="$1"
|
local run_cmd="$1"
|
||||||
local container_name="$2"
|
local container_name="$2"
|
||||||
|
|
||||||
if _is_container_exists $container_name; then
|
if _is_container_exists "$container_name"; then
|
||||||
_is_container_running $container_name && return 0
|
_is_container_running "$container_name" && return 0
|
||||||
_start_container $container_name
|
_start_container "$container_name"
|
||||||
else
|
else
|
||||||
$run_cmd
|
$run_cmd
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if ! _is_container_running $container_name; then
|
if ! _is_container_running "$container_name"; then
|
||||||
_die "Container ${container_name} failed to start"
|
_die "Container ${container_name} failed to start"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
ID=$(_get_container_id $container_name)
|
ID=$(_get_container_id "$container_name")
|
||||||
echo "Container ${container_name} is running with ID ${ID}"
|
echo "Container ${container_name} is running with ID ${ID}"
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -93,6 +93,7 @@ _check_docker_installed() {
|
|||||||
|
|
||||||
# Checks if a container (any state) exists. Returns 1 if not found.
|
# Checks if a container (any state) exists. Returns 1 if not found.
|
||||||
_is_container_exists() {
|
_is_container_exists() {
|
||||||
|
[ -n "${1:-}" ] || { echo "_is_container_exists: Container name is empty" >&2; return 1; }
|
||||||
if ! docker ps -a --format "{{.Names}}" | grep -q "^$1$"; then
|
if ! docker ps -a --format "{{.Names}}" | grep -q "^$1$"; then
|
||||||
return 1
|
return 1
|
||||||
fi
|
fi
|
||||||
@ -101,6 +102,7 @@ _is_container_exists() {
|
|||||||
|
|
||||||
# Checks if a container is currently running. Returns 1 if not running.
|
# Checks if a container is currently running. Returns 1 if not running.
|
||||||
_is_container_running() {
|
_is_container_running() {
|
||||||
|
[ -n "${1:-}" ] || { echo "_is_container_running: Container name is empty" >&2; return 1; }
|
||||||
if ! docker ps --format "{{.Names}}" | grep -q "^$1$"; then
|
if ! docker ps --format "{{.Names}}" | grep -q "^$1$"; then
|
||||||
return 1
|
return 1
|
||||||
fi
|
fi
|
||||||
@ -119,39 +121,39 @@ _get_container_status() {
|
|||||||
|
|
||||||
# Starts an existing, stopped container.
|
# Starts an existing, stopped container.
|
||||||
_start_container() {
|
_start_container() {
|
||||||
_is_container_exists $1 || return 1
|
_is_container_exists "$1" || return 1
|
||||||
_is_container_running $1 && return 0
|
_is_container_running "$1" && return 0
|
||||||
docker start $1
|
docker start "$1"
|
||||||
}
|
}
|
||||||
|
|
||||||
# Stops a running container.
|
# Stops a running container.
|
||||||
_stop_container() {
|
_stop_container() {
|
||||||
_is_container_running $1 || return 0;
|
_is_container_running "$1" || return 0;
|
||||||
docker stop $1
|
docker stop "$1"
|
||||||
}
|
}
|
||||||
|
|
||||||
# Stops (if needed) and removes a container.
|
# Stops (if needed) and removes a container.
|
||||||
_remove_container() {
|
_remove_container() {
|
||||||
_stop_container $1
|
_stop_container "$1"
|
||||||
_is_container_exists $1 || return 0;
|
_is_container_exists "$1" || return 0;
|
||||||
docker rm $1
|
docker rm "$1"
|
||||||
}
|
}
|
||||||
|
|
||||||
# Prints the logs for a container.
|
# Prints the logs for a container.
|
||||||
_get_container_logs() {
|
_get_container_logs() {
|
||||||
if ! _is_container_exists $1; then
|
if ! _is_container_exists "$1"; then
|
||||||
echo "Container $1 does not exist"
|
echo "Container $1 does not exist"
|
||||||
return 1
|
return 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
docker logs $1
|
docker logs "$1"
|
||||||
}
|
}
|
||||||
|
|
||||||
# Checks if listed environment variables are set; calls _die() if any are missing.
|
# Checks if listed environment variables are set; calls _die() if any are missing.
|
||||||
_check_required_env_vars() {
|
_check_required_env_vars() {
|
||||||
local required_vars=("$@")
|
local required_vars=("$@")
|
||||||
for var in "${required_vars[@]}"; do
|
for var in "${required_vars[@]}"; do
|
||||||
if [ -z "${!var}" ]; then
|
if [ -z "${!var:-}" ]; then
|
||||||
_die "Required environment variable $var is not set"
|
_die "Required environment variable $var is not set"
|
||||||
fi
|
fi
|
||||||
done
|
done
|
||||||
|
@ -12,26 +12,26 @@ _autocommandrun_volume() {
|
|||||||
|
|
||||||
case "$command" in
|
case "$command" in
|
||||||
create)
|
create)
|
||||||
if docker volume ls | grep -q ${volume_name}; then
|
if docker volume ls | grep -q "${volume_name}"; then
|
||||||
echo "Volume ${volume_name} already exists - leaving unchanged"
|
echo "Volume ${volume_name} already exists - leaving unchanged"
|
||||||
return
|
return
|
||||||
fi
|
fi
|
||||||
echo "Creating volume ${volume_name}"
|
echo "Creating volume ${volume_name}"
|
||||||
docker volume create ${volume_name}
|
docker volume create "${volume_name}"
|
||||||
;;
|
;;
|
||||||
destroy)
|
destroy)
|
||||||
echo "Destroying volume ${volume_name}"
|
echo "Destroying volume ${volume_name}"
|
||||||
docker volume rm ${volume_name}
|
docker volume rm "${volume_name}"
|
||||||
;;
|
;;
|
||||||
backup)
|
backup)
|
||||||
echo "Backing up volume ${volume_name}"
|
echo "Backing up volume ${volume_name}"
|
||||||
docker run --rm -v ${volume_name}:/volume -v ${backup_folder}:/backup debian bash -c "tar -czvf /backup/backup.tgz -C /volume . && chown -R $MYID:$MYGRP /backup"
|
docker run --rm -v "${volume_name}":/volume -v "${backup_folder}":/backup debian bash -c "tar -czvf /backup/backup.tgz -C /volume . && chown -R $MYID:$MYGRP /backup"
|
||||||
;;
|
;;
|
||||||
restore)
|
restore)
|
||||||
echo "Restoring volume ${volume_name}"
|
echo "Restoring volume ${volume_name}"
|
||||||
docker volume rm ${volume_name}
|
docker volume rm "${volume_name}"
|
||||||
docker volume create ${volume_name}
|
docker volume create "${volume_name}"
|
||||||
docker run --rm -v ${volume_name}:/volume -v ${backup_folder}:/backup debian bash -c "tar -xzvf /backup/backup.tgz -C /volume --strip-components=1"
|
docker run --rm -v "${volume_name}":/volume -v "${backup_folder}":/backup debian bash -c "tar -xzvf /backup/backup.tgz -C /volume --strip-components=1"
|
||||||
;;
|
;;
|
||||||
esac
|
esac
|
||||||
}
|
}
|
||||||
@ -48,14 +48,16 @@ _autocommandrun_path() {
|
|||||||
return
|
return
|
||||||
fi
|
fi
|
||||||
echo "Creating path ${path}"
|
echo "Creating path ${path}"
|
||||||
mkdir -p ${path}
|
mkdir -p "${path}"
|
||||||
;;
|
;;
|
||||||
destroy)
|
destroy)
|
||||||
echo "Destroying path ${path}"
|
echo "Destroying path ${path}"
|
||||||
local path_parent=$(dirname ${path})
|
local path_parent;
|
||||||
local path_child=$(basename ${path})
|
path_parent=$(dirname "${path}")
|
||||||
|
local path_child;
|
||||||
|
path_child=$(basename "${path}")
|
||||||
if [ -d "${path_parent}/${path_child}" ]; then
|
if [ -d "${path_parent}/${path_child}" ]; then
|
||||||
docker run --rm -v ${path_parent}:/volume debian bash -c "rm -rfv /volume/${path_child}" || echo "Failed to destroy path ${path}"
|
docker run --rm -v "${path_parent}":/volume debian bash -c "rm -rfv /volume/${path_child}" || echo "Failed to destroy path ${path}"
|
||||||
else
|
else
|
||||||
echo "Path ${path} does not exist - nothing to destroy"
|
echo "Path ${path} does not exist - nothing to destroy"
|
||||||
fi
|
fi
|
||||||
@ -63,7 +65,7 @@ _autocommandrun_path() {
|
|||||||
backup)
|
backup)
|
||||||
echo "Backing up path ${path}"
|
echo "Backing up path ${path}"
|
||||||
if [ -d "${path}" ]; then
|
if [ -d "${path}" ]; then
|
||||||
docker run --rm -v ${path}:/path -v ${backup_folder}:/backup debian bash -c "tar -czvf /backup/backup.tgz -C /path . && chown -R $MYID:$MYGRP /backup"
|
docker run --rm -v "${path}":/path -v "${backup_folder}":/backup debian bash -c "tar -czvf /backup/backup.tgz -C /path . && chown -R $MYID:$MYGRP /backup"
|
||||||
else
|
else
|
||||||
echo "Path ${path} does not exist - nothing to backup"
|
echo "Path ${path} does not exist - nothing to backup"
|
||||||
fi
|
fi
|
||||||
@ -73,9 +75,9 @@ _autocommandrun_path() {
|
|||||||
echo "Backup file ${backup_folder}/backup.tgz does not exist - nothing to restore"
|
echo "Backup file ${backup_folder}/backup.tgz does not exist - nothing to restore"
|
||||||
else
|
else
|
||||||
echo "Clearing existing data in path ${path}"
|
echo "Clearing existing data in path ${path}"
|
||||||
docker run --rm -v ${path}:/path debian bash -c "rm -rfv /path/{*,.*}"
|
docker run --rm -v "${path}":/path debian bash -c "rm -rfv /path/{*,.*}"
|
||||||
echo "Restoring path ${path} from backup file ${backup_folder}/backup.tgz"
|
echo "Restoring path ${path} from backup file ${backup_folder}/backup.tgz"
|
||||||
tar -xzvf ${backup_folder}/backup.tgz -C ${path} --strip-components=1
|
tar -xzvf "${backup_folder}/backup.tgz" -C "${path}" --strip-components=1
|
||||||
fi
|
fi
|
||||||
;;
|
;;
|
||||||
esac
|
esac
|
||||||
@ -88,31 +90,36 @@ _autocommandrun_file() {
|
|||||||
|
|
||||||
case "$command" in
|
case "$command" in
|
||||||
create)
|
create)
|
||||||
filepath_parent=$(dirname ${filepath})
|
local file_parent;
|
||||||
filepath_child=$(basename ${filepath})
|
file_parent=$(dirname "${filepath}")
|
||||||
if [ ! -d "${filepath_parent}" ]; then
|
local file_name;
|
||||||
echo "Parent directory ${filepath_parent} of ${filepath_child} does not exist - creating"
|
file_name=$(basename "${filepath}")
|
||||||
mkdir -p ${filepath_parent}
|
if [ ! -d "${file_parent}" ]; then
|
||||||
|
echo "Parent directory ${file_parent} of ${file_name} does not exist - creating"
|
||||||
|
mkdir -p "${file_parent}"
|
||||||
fi
|
fi
|
||||||
;;
|
;;
|
||||||
destroy)
|
destroy)
|
||||||
rm -f ${filepath}
|
rm -f "${filepath}"
|
||||||
;;
|
;;
|
||||||
backup)
|
backup)
|
||||||
echo "Backing up file ${filepath}"
|
echo "Backing up file ${filepath}"
|
||||||
local file_parent=$(dirname ${filepath})
|
local file_parent;
|
||||||
local file_name=$(basename ${filepath})
|
file_parent=$(dirname "${filepath}")
|
||||||
|
local file_name;
|
||||||
|
file_name=$(basename "${filepath}")
|
||||||
if [ -f "${file_parent}/${file_name}" ]; then
|
if [ -f "${file_parent}/${file_name}" ]; then
|
||||||
docker run --rm -v ${file_parent}:/volume -v ${backup_folder}:/backup debian bash -c "cp /volume/${file_name} /backup/${file_name} && chown -R $MYID:$MYGRP /backup"
|
docker run --rm -v "${file_parent}":/volume -v "${backup_folder}":/backup debian bash -c "cp /volume/${file_name} /backup/${file_name} && chown -R $MYID:$MYGRP /backup"
|
||||||
else
|
else
|
||||||
echo "File ${filepath} does not exist - nothing to backup"
|
echo "File ${filepath} does not exist - nothing to backup"
|
||||||
fi
|
fi
|
||||||
;;
|
;;
|
||||||
restore)
|
restore)
|
||||||
echo "Restoring file ${filepath}"
|
echo "Restoring file ${filepath}"
|
||||||
local file_name=$(basename ${filepath})
|
local file_name;
|
||||||
rm -f ${filepath} || die "Unable to remove existing file ${filepath}, restore failed."
|
file_name=$(basename "${filepath}")
|
||||||
cp ${backup_folder}/${file_name} ${filepath} || die "Unable to copy file ${backup_folder}/${file_name} to ${filepath}, restore failed."
|
rm -f "${filepath}" || return_die "Unable to remove existing file ${filepath}, restore failed."
|
||||||
|
cp "${backup_folder}/${file_name}" "${filepath}" || return_die "Unable to copy file ${backup_folder}/${file_name} to ${filepath}, restore failed."
|
||||||
;;
|
;;
|
||||||
esac
|
esac
|
||||||
}
|
}
|
||||||
@ -153,9 +160,10 @@ _autocommandparse() {
|
|||||||
local value="${pair#*=}"
|
local value="${pair#*=}"
|
||||||
|
|
||||||
# create backup folder unique to key/value.
|
# create backup folder unique to key/value.
|
||||||
local bfolder=$(echo "${key}_${value}" | tr -cd '[:alnum:]_-')
|
local bfolder;
|
||||||
|
bfolder=$(echo "${key}_${value}" | tr -cd '[:alnum:]_-')
|
||||||
local targetpath="${backup_temp_path}/${bfolder}"
|
local targetpath="${backup_temp_path}/${bfolder}"
|
||||||
mkdir -p ${targetpath}
|
mkdir -p "${targetpath}"
|
||||||
|
|
||||||
# Key must be one of volume, path or file
|
# Key must be one of volume, path or file
|
||||||
case "$key" in
|
case "$key" in
|
||||||
@ -191,7 +199,7 @@ databackup() {
|
|||||||
|
|
||||||
|
|
||||||
mkdir -p "$BACKUP_TEMP_PATH"
|
mkdir -p "$BACKUP_TEMP_PATH"
|
||||||
echo "_autocommandparse [backup] [$BACKUP_TEMP_PATH] [$@]"
|
echo "_autocommandparse [backup] [$BACKUP_TEMP_PATH]" "$@"
|
||||||
_autocommandparse backup "$BACKUP_TEMP_PATH" "$@"
|
_autocommandparse backup "$BACKUP_TEMP_PATH" "$@"
|
||||||
|
|
||||||
tar zcvf "$BACKUP_FILE" -C "$BACKUP_TEMP_PATH" .
|
tar zcvf "$BACKUP_FILE" -C "$BACKUP_TEMP_PATH" .
|
||||||
@ -201,7 +209,7 @@ datarestore() {
|
|||||||
_check_required_env_vars "BACKUP_FILE" "TEMP_DIR"
|
_check_required_env_vars "BACKUP_FILE" "TEMP_DIR"
|
||||||
BACKUP_TEMP_PATH="$TEMP_DIR/restore"
|
BACKUP_TEMP_PATH="$TEMP_DIR/restore"
|
||||||
|
|
||||||
echo "_autocommandparse [restore] [$BACKUP_TEMP_PATH] [$@]"
|
echo "_autocommandparse [restore] [$BACKUP_TEMP_PATH]" "$@"
|
||||||
|
|
||||||
mkdir -p "$BACKUP_TEMP_PATH"
|
mkdir -p "$BACKUP_TEMP_PATH"
|
||||||
tar zxvf "$BACKUP_FILE" -C "$BACKUP_TEMP_PATH" --strip-components=1
|
tar zxvf "$BACKUP_FILE" -C "$BACKUP_TEMP_PATH" --strip-components=1
|
||||||
|
105
source/build.sh
105
source/build.sh
@ -1,105 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
# Exit on error
|
|
||||||
set -e
|
|
||||||
|
|
||||||
# Colors for output
|
|
||||||
RED='\033[0;31m'
|
|
||||||
GREEN='\033[0;32m'
|
|
||||||
YELLOW='\033[1;33m'
|
|
||||||
NC='\033[0m' # No Color
|
|
||||||
|
|
||||||
JOBS=4
|
|
||||||
# Determine number of CPU cores for parallel build
|
|
||||||
if command -v nproc >/dev/null 2>&1; then
|
|
||||||
JOBS=$(nproc)
|
|
||||||
fi
|
|
||||||
|
|
||||||
|
|
||||||
# Function to print status messages
|
|
||||||
print_status() {
|
|
||||||
echo -e "${GREEN}[*] $1${NC}"
|
|
||||||
}
|
|
||||||
|
|
||||||
print_error() {
|
|
||||||
echo -e "${RED}[!] $1${NC}"
|
|
||||||
}
|
|
||||||
|
|
||||||
print_warning() {
|
|
||||||
echo -e "${YELLOW}[!] $1${NC}"
|
|
||||||
}
|
|
||||||
|
|
||||||
# ensure we have latest dehydrate.
|
|
||||||
dehydrate -u
|
|
||||||
|
|
||||||
# Check if build directory exists, if not create it
|
|
||||||
if [ ! -d "build" ]; then
|
|
||||||
print_status "Creating build directory..."
|
|
||||||
mkdir build
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Enter build directory
|
|
||||||
cd build
|
|
||||||
|
|
||||||
# Check if CMake is installed
|
|
||||||
if ! command -v cmake &> /dev/null; then
|
|
||||||
print_error "CMake is not installed. Please install CMake first."
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Check if make is installed
|
|
||||||
if ! command -v make &> /dev/null; then
|
|
||||||
print_error "Make is not installed. Please install Make first."
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Check if pkg-config is installed
|
|
||||||
if ! command -v pkg-config &> /dev/null; then
|
|
||||||
print_error "pkg-config is not installed. Please install pkg-config first."
|
|
||||||
print_warning "On Ubuntu/Debian: sudo apt-get install pkg-config"
|
|
||||||
print_warning "On Fedora: sudo dnf install pkg-config"
|
|
||||||
print_warning "On Arch: sudo pacman -S pkg-config"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Check if ncurses is installed
|
|
||||||
if ! pkg-config --exists ncurses; then
|
|
||||||
print_error "ncurses is not installed. Please install ncurses first."
|
|
||||||
print_warning "On Ubuntu/Debian: sudo apt-get install libncurses-dev"
|
|
||||||
print_warning "On Fedora: sudo dnf install ncurses-devel"
|
|
||||||
print_warning "On Arch: sudo pacman -S ncurses"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Configure with CMake
|
|
||||||
print_status "Configuring with CMake..."
|
|
||||||
cmake .. -DCMAKE_BUILD_TYPE=Debug
|
|
||||||
#cmake .. -DCMAKE_BUILD_TYPE=Release
|
|
||||||
|
|
||||||
# Build the project
|
|
||||||
print_status "Building project..."
|
|
||||||
make -j"$JOBS"
|
|
||||||
|
|
||||||
# Check if build was successful
|
|
||||||
if [ $? -eq 0 ]; then
|
|
||||||
print_status "Build successful!"
|
|
||||||
print_status "Binary location: $(pwd)/dropshell"
|
|
||||||
else
|
|
||||||
print_error "Build failed!"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
print_status "Auto-installing dropshell..."
|
|
||||||
sudo make install
|
|
||||||
if [ $? -eq 0 ]; then
|
|
||||||
print_status "Installation successful!"
|
|
||||||
else
|
|
||||||
print_error "Installation failed!"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
|
|
||||||
# Return to original directory
|
|
||||||
cd ..
|
|
||||||
|
|
||||||
print_status "Build process completed!"
|
|
53
source/build_native.sh
Executable file
53
source/build_native.sh
Executable file
@ -0,0 +1,53 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
|
||||||
|
SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" &> /dev/null && pwd )"
|
||||||
|
OUTPUT_DIR=${SCRIPT_DIR}/output
|
||||||
|
INSTALL_DIR=${HOME}/.local/bin
|
||||||
|
mkdir -p "${OUTPUT_DIR}"
|
||||||
|
|
||||||
|
# Exit on error
|
||||||
|
set -euo pipefail
|
||||||
|
|
||||||
|
ARCH=$(uname -m)
|
||||||
|
if [ "$ARCH" != "x86_64" ] && [ "$ARCH" != "aarch64" ]; then
|
||||||
|
echo "Unsupported architecture: $ARCH"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
|
||||||
|
function build_native() {
|
||||||
|
local BUILDDIR=${SCRIPT_DIR}/build/native
|
||||||
|
local PREVDIR=$PWD
|
||||||
|
local JOBS;
|
||||||
|
JOBS=$(nproc) # Set JOBS to the number of available CPU cores
|
||||||
|
mkdir -p "${BUILDDIR}"
|
||||||
|
cd "${SCRIPT_DIR}" || exit 1
|
||||||
|
|
||||||
|
CC="${HOME}/.musl-cross/${ARCH}-linux-musl-native/bin/${ARCH}-linux-musl-gcc"
|
||||||
|
CXX="${HOME}/.musl-cross/${ARCH}-linux-musl-native/bin/${ARCH}-linux-musl-g++"
|
||||||
|
|
||||||
|
|
||||||
|
cmake -B "${BUILDDIR}" -G Ninja \
|
||||||
|
-DCMAKE_BUILD_TYPE=Debug \
|
||||||
|
-DCMAKE_C_COMPILER_LAUNCHER=ccache \
|
||||||
|
-DCMAKE_CXX_COMPILER_LAUNCHER=ccache \
|
||||||
|
-DCMAKE_LINKER=mold \
|
||||||
|
-DCMAKE_C_COMPILER="${CC}" \
|
||||||
|
-DCMAKE_CXX_COMPILER="${CXX}"
|
||||||
|
|
||||||
|
cd "${BUILDDIR}" || exit 1
|
||||||
|
ninja -j"$JOBS"
|
||||||
|
|
||||||
|
#upx ${BUILDDIR}/dropshell
|
||||||
|
cp "${BUILDDIR}/dropshell" "${OUTPUT_DIR}/dropshell.${ARCH}"
|
||||||
|
|
||||||
|
cd "${PREVDIR}" || exit 1
|
||||||
|
}
|
||||||
|
|
||||||
|
build_native
|
||||||
|
|
||||||
|
|
||||||
|
echo "Auto-installing dropshell locally..."
|
||||||
|
mkdir -p "${INSTALL_DIR}"
|
||||||
|
cp "${OUTPUT_DIR}/dropshell.${ARCH}" "${INSTALL_DIR}/dropshell"
|
||||||
|
echo "Build process completed!"
|
37
source/build_production.sh
Executable file
37
source/build_production.sh
Executable file
@ -0,0 +1,37 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -euo pipefail
|
||||||
|
SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" &> /dev/null && pwd )"
|
||||||
|
|
||||||
|
# Create output directory
|
||||||
|
mkdir -p "${SCRIPT_DIR}/output"
|
||||||
|
PREV_DIR=$(pwd)
|
||||||
|
cd "${SCRIPT_DIR}"
|
||||||
|
trap 'cd "${PREV_DIR}"' EXIT
|
||||||
|
|
||||||
|
function build_arch() {
|
||||||
|
local arch=$1
|
||||||
|
|
||||||
|
if [ ! -f "${HOME}/.musl-cross/${arch}-linux-musl-cross/bin/${arch}-linux-musl-c++" ]; then
|
||||||
|
echo "Musl cross compiler for ${arch} not found. Please run install_build_prerequisites.sh first."
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
CMAKE_BUILD_TYPE=Release
|
||||||
|
CC="${HOME}/.musl-cross/${arch}-linux-musl-cross/bin/${arch}-linux-musl-gcc"
|
||||||
|
CXX="${HOME}/.musl-cross/${arch}-linux-musl-cross/bin/${arch}-linux-musl-g++"
|
||||||
|
|
||||||
|
BUILDDIR="${SCRIPT_DIR}/build/${arch}"
|
||||||
|
mkdir -p "${BUILDDIR}"
|
||||||
|
|
||||||
|
cmake -B "${BUILDDIR}" -G Ninja -DCMAKE_BUILD_TYPE="${CMAKE_BUILD_TYPE}" -DCMAKE_C_COMPILER="${CC}" -DCMAKE_CXX_COMPILER="${CXX}"
|
||||||
|
cmake --build "${BUILDDIR}"
|
||||||
|
|
||||||
|
upx "${BUILDDIR}/dropshell"
|
||||||
|
cp "${BUILDDIR}/dropshell" "${SCRIPT_DIR}/output/dropshell.${arch}"
|
||||||
|
}
|
||||||
|
|
||||||
|
build_arch x86_64
|
||||||
|
build_arch aarch64
|
||||||
|
|
||||||
|
echo "Static binaries have been created:"
|
||||||
|
ls -la output
|
@ -1,20 +1,24 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
set -e
|
set -e
|
||||||
|
|
||||||
|
# CMake pre-build script.
|
||||||
|
# Runs before the build process.
|
||||||
|
|
||||||
# This script creates two files:
|
# This script creates two files:
|
||||||
# src/utils/createagent.hpp
|
# src/utils/createagent.hpp
|
||||||
# src/utils/createagent.cpp
|
# src/utils/createagent.cpp
|
||||||
#
|
#
|
||||||
|
|
||||||
SCRIPT_DIR=$(dirname "$0")
|
SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
|
||||||
|
|
||||||
# check if dehydrate is installed
|
# check if dehydrate is installed
|
||||||
if ! command -v dehydrate &> /dev/null; then
|
if ! command -v dehydrate &> /dev/null; then
|
||||||
echo "dehydrate could not be found - installing"
|
echo "dehydrate could not be found - installing"
|
||||||
curl -fsSL https://gitea.jde.nz/public/dehydrate/releases/download/latest/install.sh | bash
|
curl -fsSL https://gitea.jde.nz/public/dehydrate/releases/download/latest/install.sh | bash
|
||||||
|
else
|
||||||
|
# ensure we have latest dehydrate.
|
||||||
|
dehydrate -u
|
||||||
fi
|
fi
|
||||||
|
|
||||||
mkdir -p "${SCRIPT_DIR}/src/autogen"
|
mkdir -p "${SCRIPT_DIR}/src/autogen"
|
||||||
|
|
||||||
dehydrate "${SCRIPT_DIR}/agent-remote" "${SCRIPT_DIR}/src/autogen"
|
dehydrate "${SCRIPT_DIR}/agent-remote" "${SCRIPT_DIR}/src/autogen"
|
||||||
dehydrate "${SCRIPT_DIR}/agent-local" "${SCRIPT_DIR}/src/autogen"
|
dehydrate "${SCRIPT_DIR}/agent-local" "${SCRIPT_DIR}/src/autogen"
|
@ -37,11 +37,23 @@ fi
|
|||||||
|
|
||||||
print_status "Detected OS: $OS $VER"
|
print_status "Detected OS: $OS $VER"
|
||||||
|
|
||||||
|
|
||||||
|
#----------------------------------------------------------------------------------------------------------
|
||||||
|
# INSTALL PREREQUISITE PACKAGES
|
||||||
|
#----------------------------------------------------------------------------------------------------------
|
||||||
|
|
||||||
# Define packages based on distribution
|
# Define packages based on distribution
|
||||||
case $OS in
|
case $OS in
|
||||||
"Ubuntu"|"Debian GNU/Linux")
|
"Ubuntu"|"Debian GNU/Linux")
|
||||||
# Common packages for both Ubuntu and Debian
|
# Common packages for both Ubuntu and Debian
|
||||||
PACKAGES="cmake make g++ devscripts debhelper"
|
PACKAGES="bash cmake make g++ devscripts debhelper build-essential upx musl-tools wget tar ccache ninja-build"
|
||||||
|
INSTALLCMD="apt-get install -y"
|
||||||
|
UPDATECMD="apt-get update"
|
||||||
|
;;
|
||||||
|
"Alpine Linux")
|
||||||
|
PACKAGES="bash build-base cmake git nlohmann-json wget tar curl ninja mold nodejs npm"
|
||||||
|
INSTALLCMD="apk add --no-cache"
|
||||||
|
UPDATECMD="apk update"
|
||||||
;;
|
;;
|
||||||
*)
|
*)
|
||||||
print_error "Unsupported distribution: $OS"
|
print_error "Unsupported distribution: $OS"
|
||||||
@ -51,19 +63,29 @@ esac
|
|||||||
|
|
||||||
# Function to check if a package is installed
|
# Function to check if a package is installed
|
||||||
is_package_installed() {
|
is_package_installed() {
|
||||||
dpkg -l "$1" 2>/dev/null | grep -q "^ii"
|
if [ "$OS" = "Alpine Linux" ]; then
|
||||||
|
# Use apk info <pkg> and check exit status
|
||||||
|
apk info "$1" >/dev/null 2>&1
|
||||||
|
return $?
|
||||||
|
else
|
||||||
|
dpkg -l "$1" 2>/dev/null | grep -q "^ii"
|
||||||
|
fi
|
||||||
}
|
}
|
||||||
|
|
||||||
# Update package lists
|
UPDATED=false
|
||||||
print_status "Updating package lists..."
|
|
||||||
apt-get update
|
|
||||||
|
|
||||||
# Install missing packages
|
# Install missing packages
|
||||||
print_status "Checking and installing required packages..."
|
print_status "Checking and installing required packages..."
|
||||||
for pkg in $PACKAGES; do
|
for pkg in $PACKAGES; do
|
||||||
if ! is_package_installed "$pkg"; then
|
if ! is_package_installed "$pkg"; then
|
||||||
|
if [ "$UPDATED" = false ]; then
|
||||||
|
print_status "Updating package lists..."
|
||||||
|
$UPDATECMD
|
||||||
|
UPDATED=true
|
||||||
|
fi
|
||||||
|
|
||||||
print_status "Installing $pkg..."
|
print_status "Installing $pkg..."
|
||||||
apt-get install -y "$pkg"
|
$INSTALLCMD "$pkg"
|
||||||
if [ $? -ne 0 ]; then
|
if [ $? -ne 0 ]; then
|
||||||
print_error "Failed to install $pkg"
|
print_error "Failed to install $pkg"
|
||||||
exit 1
|
exit 1
|
||||||
@ -73,82 +95,71 @@ for pkg in $PACKAGES; do
|
|||||||
fi
|
fi
|
||||||
done
|
done
|
||||||
|
|
||||||
# Verify all required tools are installed
|
# ----------------------------------------------------------------------------------------------------------
|
||||||
print_status "Verifying installation..."
|
# MUSL CROSS COMPILERS
|
||||||
for tool in cmake make g++; do
|
# ----------------------------------------------------------------------------------------------------------
|
||||||
if ! command -v "$tool" &> /dev/null; then
|
|
||||||
print_error "$tool is not installed properly"
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
done
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
# Install other required packages
|
|
||||||
apt install -y musl-tools wget tar
|
|
||||||
|
|
||||||
# Set install directory
|
# Set install directory
|
||||||
if [ -n "$SUDO_USER" ] && [ "$SUDO_USER" != "root" ]; then
|
if [ -n "$SUDO_USER" ] && [ "$SUDO_USER" != "root" ]; then
|
||||||
USER_HOME=$(eval echo ~$SUDO_USER)
|
USER_HOME=$(eval echo "~$SUDO_USER")
|
||||||
else
|
else
|
||||||
USER_HOME="$HOME"
|
USER_HOME="$HOME"
|
||||||
fi
|
fi
|
||||||
INSTALL_DIR="$USER_HOME/.musl-cross"
|
INSTALL_DIR="$USER_HOME/.musl-cross"
|
||||||
mkdir -p "$INSTALL_DIR"
|
mkdir -p "$INSTALL_DIR"
|
||||||
|
|
||||||
MUSL_CC_URL="https://musl.cc"
|
|
||||||
TMPDIR=$(mktemp -d)
|
TMPDIR=$(mktemp -d)
|
||||||
trap 'rm -rf "$TMPDIR"' EXIT
|
trap 'rm -rf "$TMPDIR"' EXIT
|
||||||
|
|
||||||
# x86_64
|
function install_musl_cross() {
|
||||||
if [ ! -d "$INSTALL_DIR/x86_64-linux-musl-cross" ]; then
|
local TOOLCHAIN="$1"
|
||||||
echo "Downloading x86_64 musl cross toolchain..."
|
local MUSL_CC_URL="https://musl.cc"
|
||||||
wget -nc -O "$TMPDIR/x86_64-linux-musl-cross.tgz" $MUSL_CC_URL/x86_64-linux-musl-cross.tgz
|
if [ ! -d "$INSTALL_DIR/$TOOLCHAIN" ]; then
|
||||||
tar -C "$INSTALL_DIR" -xvf "$TMPDIR/x86_64-linux-musl-cross.tgz"
|
echo "Downloading $TOOLCHAIN musl cross toolchain..."
|
||||||
fi
|
wget -nc -O "$TMPDIR/$TOOLCHAIN.tgz" "$MUSL_CC_URL/$TOOLCHAIN.tgz"
|
||||||
|
tar -C "$INSTALL_DIR" -xvf "$TMPDIR/$TOOLCHAIN.tgz"
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
# aarch64
|
function check_path() {
|
||||||
if [ ! -d "$INSTALL_DIR/aarch64-linux-musl-cross" ]; then
|
if [ -n "$SUDO_USER" ] && [ "$SUDO_USER" != "root" ]; then
|
||||||
echo "Downloading aarch64 musl cross toolchain..."
|
local BASHRC="$USER_HOME/.bashrc"
|
||||||
wget -nc -O "$TMPDIR/aarch64-linux-musl-cross.tgz" $MUSL_CC_URL/aarch64-linux-musl-cross.tgz
|
local TOOLCHAIN="$1"
|
||||||
tar -C "$INSTALL_DIR" -xvf "$TMPDIR/aarch64-linux-musl-cross.tgz"
|
local MUSL_PATH="$INSTALL_DIR/$TOOLCHAIN/bin"
|
||||||
fi
|
if ! echo "$PATH" | grep -q "$MUSL_PATH"; then
|
||||||
|
echo "Adding $MUSL_PATH to PATH in $BASHRC"
|
||||||
|
PATH_LINE="export PATH=\"$MUSL_PATH:\$PATH\""
|
||||||
|
if ! grep -Fxq "$PATH_LINE" "$BASHRC"; then
|
||||||
|
echo "" >> "$BASHRC"
|
||||||
|
echo "# Add musl cross compilers to PATH for dropshell" >> "$BASHRC"
|
||||||
|
echo "$PATH_LINE" >> "$BASHRC"
|
||||||
|
echo "Added musl cross compilers to $BASHRC"
|
||||||
|
echo "You should run 'source ~/.bashrc' to update your PATH"
|
||||||
|
else
|
||||||
|
echo "You should run 'source ~/.bashrc' to update your PATH"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
# Print instructions for adding to PATH
|
|
||||||
# cat <<EOF
|
|
||||||
|
|
||||||
# To use the musl cross compilers, add the following to your shell:
|
TOOLCHAIN_LIST=(
|
||||||
# export PATH="$INSTALL_DIR/x86_64-linux-musl-cross/bin:$INSTALL_DIR/aarch64-linux-musl-cross/bin:$PATH"
|
"aarch64-linux-musl-cross"
|
||||||
|
"x86_64-linux-musl-cross"
|
||||||
|
"x86_64-linux-musl-native"
|
||||||
|
)
|
||||||
|
|
||||||
# Or run:
|
for TOOLCHAIN in "${TOOLCHAIN_LIST[@]}"; do
|
||||||
# export PATH="$INSTALL_DIR/x86_64-linux-musl-cross/bin:$INSTALL_DIR/aarch64-linux-musl-cross/bin:\$PATH"
|
install_musl_cross "$TOOLCHAIN"
|
||||||
|
check_path "$TOOLCHAIN"
|
||||||
# EOF
|
done
|
||||||
|
|
||||||
# Clean up
|
# Clean up
|
||||||
rm -rf "$TMPDIR"
|
rm -rf "$TMPDIR"
|
||||||
|
|
||||||
# If run with sudo, add to invoking user's ~/.bashrc
|
# ----------------------------------------------------------------------------------------------------------
|
||||||
if [ -n "$SUDO_USER" ] && [ "$SUDO_USER" != "root" ]; then
|
# COMPLETE
|
||||||
BASHRC="$USER_HOME/.bashrc"
|
# ----------------------------------------------------------------------------------------------------------
|
||||||
EXPORT_LINE="export PATH=\"$INSTALL_DIR/x86_64-linux-musl-cross/bin:$INSTALL_DIR/aarch64-linux-musl-cross/bin:\$PATH\""
|
|
||||||
if ! grep -Fxq "$EXPORT_LINE" "$BASHRC"; then
|
|
||||||
echo "" >> "$BASHRC"
|
|
||||||
echo "# Add musl cross compilers to PATH for bb64" >> "$BASHRC"
|
|
||||||
echo "$EXPORT_LINE" >> "$BASHRC"
|
|
||||||
echo "Added musl cross compilers to $BASHRC"
|
|
||||||
else
|
|
||||||
echo "musl cross compiler PATH already present in $BASHRC"
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
|
|
||||||
# check if dehydrate command is installed
|
|
||||||
if ! command -v dehydrate &> /dev/null; then
|
|
||||||
curl -fsSL https://gitea.jde.nz/public/dehydrate/releases/download/latest/install.sh | bash
|
|
||||||
fi
|
|
||||||
|
|
||||||
dehydrate -u
|
|
||||||
|
|
||||||
|
|
||||||
print_status "All dependencies installed successfully!"
|
print_status "All dependencies installed successfully!"
|
||||||
print_status "You can now run ./build.sh to build the project"
|
print_status "You can now run ./build.sh to build the project"
|
@ -1,59 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
# build amd64 and arm64 versions of dropshell, to:
|
|
||||||
# build/dropshell.amd64
|
|
||||||
# build/dropshell.arm64
|
|
||||||
|
|
||||||
set -e
|
|
||||||
|
|
||||||
SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
|
|
||||||
|
|
||||||
rm -f $SCRIPT_DIR/build_amd64/dropshell $SCRIPT_DIR/build_arm64/dropshell $SCRIPT_DIR/output/dropshell.amd64 $SCRIPT_DIR/output/dropshell.arm64
|
|
||||||
|
|
||||||
# Determine number of CPU cores for parallel build
|
|
||||||
if command -v nproc >/dev/null 2>&1; then
|
|
||||||
JOBS=$(nproc)
|
|
||||||
else
|
|
||||||
JOBS=4 # fallback default
|
|
||||||
fi
|
|
||||||
|
|
||||||
PREV_PWD=$PWD
|
|
||||||
cd $SCRIPT_DIR
|
|
||||||
|
|
||||||
# Build for amd64 (musl)
|
|
||||||
echo "Building for amd64 (musl)..."
|
|
||||||
cmake -B build_amd64 -DCMAKE_BUILD_TYPE=Release \
|
|
||||||
-DCMAKE_C_COMPILER=x86_64-linux-musl-gcc \
|
|
||||||
-DCMAKE_CXX_COMPILER=x86_64-linux-musl-g++ \
|
|
||||||
-DCMAKE_EXE_LINKER_FLAGS="-static" \
|
|
||||||
-DCMAKE_CXX_FLAGS="-march=x86-64" .
|
|
||||||
cmake --build build_amd64 --target dropshell --config Release -j"$JOBS"
|
|
||||||
mkdir -p output
|
|
||||||
cp build_amd64/dropshell output/dropshell.amd64
|
|
||||||
|
|
||||||
# Build for arm64 (musl)
|
|
||||||
echo "Building for arm64 (musl)..."
|
|
||||||
cmake -B build_arm64 -DCMAKE_BUILD_TYPE=Release \
|
|
||||||
-DCMAKE_C_COMPILER=aarch64-linux-musl-gcc \
|
|
||||||
-DCMAKE_CXX_COMPILER=aarch64-linux-musl-g++ \
|
|
||||||
-DCMAKE_EXE_LINKER_FLAGS="-static" \
|
|
||||||
-DCMAKE_CXX_FLAGS="-march=armv8-a" \
|
|
||||||
-DCMAKE_SYSTEM_PROCESSOR=aarch64 .
|
|
||||||
cmake --build build_arm64 --target dropshell --config Release -j"$JOBS"
|
|
||||||
mkdir -p output
|
|
||||||
cp build_arm64/dropshell output/dropshell.arm64
|
|
||||||
|
|
||||||
if [ ! -f output/dropshell.amd64 ]; then
|
|
||||||
echo "output/dropshell.amd64 not found!" >&2
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
if [ ! -f output/dropshell.arm64 ]; then
|
|
||||||
echo "output/dropshell.arm64 not found!" >&2
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
echo "Builds complete:"
|
|
||||||
ls -lh output/dropshell.*
|
|
||||||
|
|
||||||
cd $PREV_PWD
|
|
@ -6,103 +6,52 @@ SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
|
|||||||
echo "Script directory: $SCRIPT_DIR"
|
echo "Script directory: $SCRIPT_DIR"
|
||||||
|
|
||||||
# Check for GITEA_TOKEN_DEPLOY or GITEA_TOKEN
|
# Check for GITEA_TOKEN_DEPLOY or GITEA_TOKEN
|
||||||
if [ -n "$GITEA_TOKEN_DEPLOY" ]; then
|
TOKEN="${GITEA_TOKEN_DEPLOY:-${GITEA_TOKEN}}"
|
||||||
TOKEN="$GITEA_TOKEN_DEPLOY"
|
[ -z "$TOKEN" ] && { echo "Neither GITEA_TOKEN_DEPLOY nor GITEA_TOKEN environment variable set!" >&2; exit 1; }
|
||||||
elif [ -n "$GITEA_TOKEN" ]; then
|
|
||||||
TOKEN="$GITEA_TOKEN"
|
|
||||||
else
|
|
||||||
echo "GITEA_TOKEN_DEPLOY or GITEA_TOKEN environment variable not set!" >&2
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
$SCRIPT_DIR/multibuild.sh
|
OLD_PWD="$PWD"
|
||||||
BUILD_DIR=$SCRIPT_DIR/build
|
cd "$SCRIPT_DIR" || exit 1
|
||||||
|
TEMP_DIR=$(mktemp -d)
|
||||||
OLD_PWD=$PWD
|
trap 'rm -rf "$TEMP_DIR" && cd "$OLD_PWD"' EXIT
|
||||||
cd $SCRIPT_DIR
|
|
||||||
|
|
||||||
|
|
||||||
if [ ! -f "output/dropshell.amd64" ]; then
|
|
||||||
echo "output/dropshell.amd64 not found!" >&2
|
|
||||||
echo "Please run multibuild.sh first." >&2
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
if [ ! -f "output/dropshell.arm64" ]; then
|
|
||||||
echo "output/dropshell.arm64 not found!" >&2
|
|
||||||
echo "Please run multibuild.sh first." >&2
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
TAG=$("$SCRIPT_DIR/output/dropshell.amd64" --version)
|
|
||||||
[ -z "$TAG" ] && echo "Failed to get version from dropshell.amd64" >&2 && exit 1
|
|
||||||
|
|
||||||
|
ARCH=$(uname -m)
|
||||||
|
TAG=$("$SCRIPT_DIR/output/dropshell.${ARCH}" --version)
|
||||||
|
[ -z "$TAG" ] && echo "Failed to get version from dropshell.${ARCH}" >&2 && exit 1
|
||||||
echo "Publishing dropshell version $TAG"
|
echo "Publishing dropshell version $TAG"
|
||||||
|
|
||||||
# make sure we've commited.
|
|
||||||
git add "$SCRIPT_DIR/../" && git commit -m "dropshell release $TAG" && git push
|
|
||||||
|
|
||||||
|
function die() {
|
||||||
# Find repo info from .git/config
|
echo "$@" >&2
|
||||||
REPO_URL=$(git config --get remote.origin.url)
|
|
||||||
if [[ ! $REPO_URL =~ gitea ]]; then
|
|
||||||
echo "Remote origin is not a Gitea repository: $REPO_URL" >&2
|
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
|
||||||
|
|
||||||
# Extract base URL, owner, and repo
|
|
||||||
# Example: https://gitea.example.com/username/reponame.git
|
|
||||||
BASE_URL=$(echo "$REPO_URL" | sed -E 's#(https?://[^/]+)/.*#\1#')
|
|
||||||
OWNER=$(echo "$REPO_URL" | sed -E 's#.*/([^/]+)/[^/]+(\.git)?$#\1#')
|
|
||||||
REPO=$(echo "$REPO_URL" | sed -E 's#.*/([^/]+)(\.git)?$#\1#')
|
|
||||||
|
|
||||||
API_URL="$BASE_URL/api/v1/repos/$OWNER/$REPO"
|
|
||||||
|
|
||||||
# Create release
|
|
||||||
RELEASE_DATA=$(cat <<EOF
|
|
||||||
{
|
|
||||||
"tag_name": "$TAG",
|
|
||||||
"name": "$TAG",
|
|
||||||
"body": "dropshell release $TAG",
|
|
||||||
"draft": false,
|
|
||||||
"prerelease": false
|
|
||||||
}
|
}
|
||||||
EOF
|
|
||||||
)
|
|
||||||
|
|
||||||
RELEASE_ID=$(curl -s -X POST "$API_URL/releases" \
|
# Function to find file in specified locations
|
||||||
-H "Content-Type: application/json" \
|
find_file() {
|
||||||
-H "Authorization: token $TOKEN" \
|
local filename="$1"
|
||||||
-d "$RELEASE_DATA" | grep -o '"id":[0-9]*' | head -1 | cut -d: -f2)
|
shift # remove filename from args
|
||||||
|
local locations=("$@") # grab the rest of the args as locations
|
||||||
|
|
||||||
|
for loc in "${locations[@]}"; do
|
||||||
|
if [ -f "$loc/$filename" ]; then
|
||||||
|
echo "$loc/$filename"
|
||||||
|
return 0 # Found the file, return success
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
echo "" # Return empty string if not found
|
||||||
|
return 1
|
||||||
|
}
|
||||||
|
|
||||||
|
curl -L -s -o "${TEMP_DIR}/sos" "https://getbin.xyz/sos" || die "Failed to download sos"
|
||||||
|
chmod +x "${TEMP_DIR}/sos"
|
||||||
|
|
||||||
if [ -z "$RELEASE_ID" ]; then
|
|
||||||
echo "Failed to create release on Gitea." >&2
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Upload binaries and install.sh
|
# Upload binaries and install.sh
|
||||||
for FILE in dropshell.amd64 dropshell.arm64 install.sh server_autosetup.sh; do
|
for FILE in dropshell.x86_64 dropshell.aarch64 dropshell-install.sh dropshell-server-autosetup.sh; do
|
||||||
if [ -f "output/$FILE" ]; then
|
# Pass the locations directly to the find_file function
|
||||||
filetoupload="output/$FILE"
|
filetoupload=$(find_file "$FILE" "output" "../" ".")
|
||||||
elif [ -f "../$FILE" ]; then
|
[ -z "$filetoupload" ] && die "File $FILE not found in expected locations!"
|
||||||
filetoupload="../$FILE"
|
|
||||||
elif [ -f "$FILE" ]; then
|
|
||||||
filetoupload="$FILE"
|
|
||||||
else
|
|
||||||
echo "File $FILE not found!" >&2
|
|
||||||
continue
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Auto-detect content type
|
"${TEMP_DIR}/sos" upload getbin.xyz "$filetoupload" "$FILE:latest" "$FILE:TAG"
|
||||||
ctype=$(file --mime-type -b "$filetoupload")
|
|
||||||
|
|
||||||
curl -s -X POST "$API_URL/releases/$RELEASE_ID/assets?name=$FILE" \
|
|
||||||
-H "Content-Type: $ctype" \
|
|
||||||
-H "Authorization: token $TOKEN" \
|
|
||||||
--data-binary @"$filetoupload"
|
|
||||||
echo "Uploaded $FILE to release $TAG as $ctype."
|
|
||||||
done
|
done
|
||||||
|
|
||||||
echo "Published dropshell version $TAG to $REPO_URL (tag $TAG) with binaries."
|
echo "Published dropshell $TAG to getbin.xyz"
|
||||||
|
|
||||||
cd $OLD_PWD
|
|
||||||
|
@ -84,7 +84,7 @@ namespace dropshell
|
|||||||
remote_command_script_file,
|
remote_command_script_file,
|
||||||
remotefile(server, user).service_env(service)}, user))
|
remotefile(server, user).service_env(service)}, user))
|
||||||
{
|
{
|
||||||
error << "Error: Required service directories not found on remote server" << std::endl;
|
error << "Required service directories not found on remote server" << std::endl;
|
||||||
info << "Is the service installed?" << std::endl;
|
info << "Is the service installed?" << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
@ -103,7 +103,7 @@ namespace dropshell
|
|||||||
std::string local_backups_dir = localpath::backups();
|
std::string local_backups_dir = localpath::backups();
|
||||||
if (local_backups_dir.empty())
|
if (local_backups_dir.empty())
|
||||||
{
|
{
|
||||||
error << "Error: Local backups directory not found" << std::endl;
|
error << "Local backups directory not found" << std::endl;
|
||||||
info << "Run 'dropshell edit' to configure DropShell" << std::endl;
|
info << "Run 'dropshell edit' to configure DropShell" << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
@ -103,6 +103,11 @@ namespace dropshell
|
|||||||
if (server_name.empty() || template_name.empty() || service_name.empty())
|
if (server_name.empty() || template_name.empty() || service_name.empty())
|
||||||
return false;
|
return false;
|
||||||
|
|
||||||
|
if (!legal_service_name(service_name)) {
|
||||||
|
error << "Service name contains illegal characters: " << service_name << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
ServerConfig server_info(server_name);
|
ServerConfig server_info(server_name);
|
||||||
if (!server_info.is_valid())
|
if (!server_info.is_valid())
|
||||||
{
|
{
|
||||||
|
@ -94,11 +94,11 @@ int edit_config()
|
|||||||
|
|
||||||
std::string config_file = localfile::dropshell_json();
|
std::string config_file = localfile::dropshell_json();
|
||||||
if (!edit_file(config_file, false) || !std::filesystem::exists(config_file))
|
if (!edit_file(config_file, false) || !std::filesystem::exists(config_file))
|
||||||
return die("Error: Failed to edit config file.");
|
return return_die("Failed to edit config file.");
|
||||||
|
|
||||||
gConfig().load_config();
|
gConfig().load_config();
|
||||||
if (!gConfig().is_config_set())
|
if (!gConfig().is_config_set())
|
||||||
return die("Error: Failed to load and parse edited config file!");
|
return return_die("Failed to load and parse edited config file!");
|
||||||
|
|
||||||
gConfig().save_config(true);
|
gConfig().save_config(true);
|
||||||
|
|
||||||
@ -112,7 +112,7 @@ int edit_config()
|
|||||||
int edit_server(const std::string &server_name)
|
int edit_server(const std::string &server_name)
|
||||||
{
|
{
|
||||||
if (localpath::server(server_name).empty()) {
|
if (localpath::server(server_name).empty()) {
|
||||||
std::cerr << "Error: Server not found: " << server_name << std::endl;
|
error << "Server not found: " << server_name << std::endl;
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -130,6 +130,28 @@ int edit_server(const std::string &server_name)
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void list_directory(std::string dir, std::string msg)
|
||||||
|
{
|
||||||
|
bool first=true;
|
||||||
|
std::vector<std::string> directories;
|
||||||
|
for (const auto &file : std::filesystem::directory_iterator(dir))
|
||||||
|
{
|
||||||
|
if (first)
|
||||||
|
{
|
||||||
|
if (!msg.empty())
|
||||||
|
info << msg << std::endl;
|
||||||
|
first=false;
|
||||||
|
}
|
||||||
|
if (std::filesystem::is_directory(file.path()))
|
||||||
|
directories.push_back(file.path());
|
||||||
|
else
|
||||||
|
info << " " << file.path() << std::endl;
|
||||||
|
}
|
||||||
|
|
||||||
|
for (const auto &dir : directories)
|
||||||
|
list_directory(dir, "");
|
||||||
|
}
|
||||||
|
|
||||||
// ------------------------------------------------------------------------------------------------
|
// ------------------------------------------------------------------------------------------------
|
||||||
// edit service config
|
// edit service config
|
||||||
// ------------------------------------------------------------------------------------------------
|
// ------------------------------------------------------------------------------------------------
|
||||||
@ -143,7 +165,14 @@ int edit_service_config(const std::string &server, const std::string &service)
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (edit_file(config_file, true) && std::filesystem::exists(config_file))
|
if (edit_file(config_file, true) && std::filesystem::exists(config_file))
|
||||||
info << "To apply your changes, run:\n dropshell install " + server + " " + service << std::endl;
|
info << "Successfully edited service config file at " << config_file << std::endl;
|
||||||
|
|
||||||
|
std::string service_dir = localpath::service(server, service);
|
||||||
|
list_directory(service_dir, "You may wish to edit the other files in " + service_dir);
|
||||||
|
|
||||||
|
info << "Then to apply your changes, run:" << std::endl;
|
||||||
|
info << " dropshell uninstall " + server + " " + service << std::endl;
|
||||||
|
info << " dropshell install " + server + " " + service << std::endl;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
83
source/src/commands/hash.cpp
Normal file
83
source/src/commands/hash.cpp
Normal file
@ -0,0 +1,83 @@
|
|||||||
|
#include "command_registry.hpp"
|
||||||
|
#include "config.hpp"
|
||||||
|
#include "utils/utils.hpp"
|
||||||
|
#include "utils/directories.hpp"
|
||||||
|
#include "shared_commands.hpp"
|
||||||
|
#include "version.hpp"
|
||||||
|
#include "hash.hpp"
|
||||||
|
|
||||||
|
#include <unistd.h>
|
||||||
|
#include <cstring>
|
||||||
|
#include <iostream>
|
||||||
|
#include <sstream>
|
||||||
|
#include <filesystem>
|
||||||
|
#include <libassert/assert.hpp>
|
||||||
|
|
||||||
|
namespace dropshell {
|
||||||
|
|
||||||
|
void hash_autocomplete(const CommandContext& ctx);
|
||||||
|
int hash_handler(const CommandContext& ctx);
|
||||||
|
|
||||||
|
static std::vector<std::string> hash_name_list={"hash"};
|
||||||
|
|
||||||
|
// Static registration
|
||||||
|
struct HashCommandRegister {
|
||||||
|
HashCommandRegister() {
|
||||||
|
CommandRegistry::instance().register_command({
|
||||||
|
hash_name_list,
|
||||||
|
hash_handler,
|
||||||
|
hash_autocomplete,
|
||||||
|
false, // hidden
|
||||||
|
false, // requires_config
|
||||||
|
false, // requires_install
|
||||||
|
0, // min_args (after command)
|
||||||
|
1, // max_args (after command)
|
||||||
|
"hash [FILE|DIRECTORY]",
|
||||||
|
"Hash a file or directory.",
|
||||||
|
// heredoc
|
||||||
|
R"(
|
||||||
|
Hash a file or directory recursively.
|
||||||
|
)"
|
||||||
|
});
|
||||||
|
}
|
||||||
|
} hash_command_register;
|
||||||
|
|
||||||
|
|
||||||
|
void hash_autocomplete(const CommandContext& ctx) {
|
||||||
|
if (ctx.args.size() == 0) {
|
||||||
|
// list all files and directories in the current directory
|
||||||
|
for (const auto& entry : std::filesystem::directory_iterator(".")) {
|
||||||
|
rawout << entry.path().string() << std::endl;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
int hash_handler(const CommandContext& ctx) {
|
||||||
|
std::filesystem::path path = safearg(ctx.args, 0);
|
||||||
|
if (path.empty())
|
||||||
|
path=std::filesystem::current_path();
|
||||||
|
|
||||||
|
if (!std::filesystem::exists(path))
|
||||||
|
{
|
||||||
|
error << "Does not exist: " << path.string() << std::endl;
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (std::filesystem::is_directory(path))
|
||||||
|
{
|
||||||
|
// hash the directory recursively
|
||||||
|
uint64_t hash = hash_directory_recursive(path.string());
|
||||||
|
std::cout << hash << std::endl;
|
||||||
|
}
|
||||||
|
else
|
||||||
|
{
|
||||||
|
// hash the file
|
||||||
|
uint64_t hash = hash_file(path.string());
|
||||||
|
std::cout << hash << std::endl;
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
} // namespace dropshell
|
@ -18,6 +18,7 @@
|
|||||||
#include <filesystem>
|
#include <filesystem>
|
||||||
#include <libassert/assert.hpp>
|
#include <libassert/assert.hpp>
|
||||||
#include "servers.hpp"
|
#include "servers.hpp"
|
||||||
|
#include <sys/stat.h>
|
||||||
|
|
||||||
namespace dropshell
|
namespace dropshell
|
||||||
{
|
{
|
||||||
@ -68,7 +69,29 @@ namespace dropshell
|
|||||||
std::string server = server_env.get_server_name();
|
std::string server = server_env.get_server_name();
|
||||||
LocalServiceInfo service_info = get_service_info(server_env.get_server_name(), service);
|
LocalServiceInfo service_info = get_service_info(server_env.get_server_name(), service);
|
||||||
|
|
||||||
if (!SIvalid(service_info) || !service_info.service_template_hash_match)
|
if (!SIvalid(service_info))
|
||||||
|
{
|
||||||
|
error << "Failed to install - service information not valid." << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!server_env.is_valid())
|
||||||
|
return false; // should never hit this.
|
||||||
|
|
||||||
|
|
||||||
|
std::string user = service_info.user;
|
||||||
|
std::string remote_service_path = remotepath(server,user).service(service);
|
||||||
|
|
||||||
|
ASSERT(!remote_service_path.empty(), "Install_Service: Remote service path is empty for " + service + " on " + server);
|
||||||
|
ASSERT(!user.empty(), "Install_Service: User is empty for " + service + " on " + server);
|
||||||
|
|
||||||
|
if (server_env.check_remote_dir_exists(remote_service_path, user))
|
||||||
|
{ // uninstall the old service before we update the config or template!
|
||||||
|
info << "Service " << service << " is already installed on " << server << std::endl;
|
||||||
|
shared_commands::uninstall_service(server_env, service);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!service_info.service_template_hash_match)
|
||||||
{
|
{
|
||||||
warning << "Service " << service << " is using an old template. Updating. " << std::endl;
|
warning << "Service " << service << " is using an old template. Updating. " << std::endl;
|
||||||
if (!merge_updated_service_template(server_env.get_server_name(), service))
|
if (!merge_updated_service_template(server_env.get_server_name(), service))
|
||||||
@ -87,9 +110,6 @@ namespace dropshell
|
|||||||
|
|
||||||
maketitle("Installing " + service + " (" + service_info.template_name + ") on " + server);
|
maketitle("Installing " + service + " (" + service_info.template_name + ") on " + server);
|
||||||
|
|
||||||
if (!server_env.is_valid())
|
|
||||||
return false; // should never hit this.
|
|
||||||
|
|
||||||
// Check if template exists
|
// Check if template exists
|
||||||
template_info tinfo = gTemplateManager().get_template_info(service_info.template_name);
|
template_info tinfo = gTemplateManager().get_template_info(service_info.template_name);
|
||||||
if (!tinfo.is_set())
|
if (!tinfo.is_set())
|
||||||
@ -102,8 +122,6 @@ namespace dropshell
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Create service directory
|
// Create service directory
|
||||||
std::string user = server_env.get_user_for_service(service);
|
|
||||||
std::string remote_service_path = remotepath(server,user).service(service);
|
|
||||||
std::string mkdir_cmd = "mkdir -p " + quote(remote_service_path);
|
std::string mkdir_cmd = "mkdir -p " + quote(remote_service_path);
|
||||||
if (!execute_ssh_command(server_env.get_SSH_INFO(user), sCommand("", mkdir_cmd, {}), cMode::Silent))
|
if (!execute_ssh_command(server_env.get_SSH_INFO(user), sCommand("", mkdir_cmd, {}), cMode::Silent))
|
||||||
{
|
{
|
||||||
@ -111,14 +129,6 @@ namespace dropshell
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check if rsync is installed on remote host
|
|
||||||
std::string check_rsync_cmd = "which rsync";
|
|
||||||
if (!execute_ssh_command(server_env.get_SSH_INFO(user), sCommand("", check_rsync_cmd, {}), cMode::Silent))
|
|
||||||
{
|
|
||||||
std::cerr << "rsync is not installed on the remote host" << std::endl;
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Copy template files
|
// Copy template files
|
||||||
debug << "Copying: [LOCAL] " << tinfo.local_template_path() << std::endl
|
debug << "Copying: [LOCAL] " << tinfo.local_template_path() << std::endl
|
||||||
<< std::string(8, ' ') << "[REMOTE] " << remotepath(server,user).service_template(service) << "/" << std::endl;
|
<< std::string(8, ' ') << "[REMOTE] " << remotepath(server,user).service_template(service) << "/" << std::endl;
|
||||||
@ -142,7 +152,13 @@ namespace dropshell
|
|||||||
// Run install script
|
// Run install script
|
||||||
{
|
{
|
||||||
info << "Running " << service_info.template_name << " install script on " << server << "..." << std::endl;
|
info << "Running " << service_info.template_name << " install script on " << server << "..." << std::endl;
|
||||||
server_env.run_remote_template_command(service, "install", {}, false, {});
|
|
||||||
|
shared_commands::cRemoteTempFolder remote_temp_folder(server_env, user);
|
||||||
|
if (!server_env.run_remote_template_command(service, "install", {}, false, {{"TEMP_DIR", remote_temp_folder.path()}}))
|
||||||
|
{
|
||||||
|
error << "Failed to run install script on " << server << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// print health tick
|
// print health tick
|
||||||
@ -176,49 +192,107 @@ namespace dropshell
|
|||||||
return trim(result);
|
return trim(result);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
int configure_autocomplete()
|
||||||
|
{
|
||||||
|
debug << "Ensuring dropshell autocomplete is registered in ~/.bashrc..." << std::endl;
|
||||||
|
|
||||||
|
std::filesystem::path bashrc = localpath::current_user_home() +"/.bashrc";
|
||||||
|
|
||||||
|
std::string autocomplete_script = R"(
|
||||||
|
#---DROPSHELL AUTOCOMPLETE START---
|
||||||
|
_dropshell_completions() {
|
||||||
|
local cur
|
||||||
|
COMPREPLY=()
|
||||||
|
cur="${COMP_WORDS[COMP_CWORD]}"
|
||||||
|
|
||||||
|
# call dropshell to get the list of possiblities for the current argument. Supply all previous arguments.
|
||||||
|
local completions=($(dropshell autocomplete "${COMP_WORDS[@]:1:${COMP_CWORD}-1}"))
|
||||||
|
COMPREPLY=( $(compgen -W "${completions[*]}" -- ${cur}) )
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
# Register the completion function
|
||||||
|
complete -F _dropshell_completions dropshell
|
||||||
|
complete -F _dropshell_completions ds
|
||||||
|
#---DROPSHELL AUTOCOMPLETE END---
|
||||||
|
)";
|
||||||
|
|
||||||
|
file_replace_or_add_segment(bashrc.string(), autocomplete_script);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
int configure_localbin()
|
||||||
|
{
|
||||||
|
debug << "Ensuring ~/.local/bin is in the ~/.bashrc path..." << std::endl;
|
||||||
|
|
||||||
|
std::filesystem::path bashrc = localpath::current_user_home() +"/.bashrc";
|
||||||
|
std::filesystem::path localbin = localpath::current_user_home() + "/.local/bin";
|
||||||
|
std::filesystem::create_directories(localbin);
|
||||||
|
// check if already in path
|
||||||
|
const char* env_p = std::getenv("PATH");
|
||||||
|
if (env_p) {
|
||||||
|
std::string path_str = env_p;
|
||||||
|
if (path_str.find(localbin.string()) == std::string::npos) {
|
||||||
|
std::string pathstr="#---DROPSHELL PATH START---\nexport PATH=\""+localbin.string()+":$PATH\"\n#---DROPSHELL PATH END---\n";
|
||||||
|
file_replace_or_add_segment(bashrc.string(), pathstr);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
int update_dropshell()
|
int update_dropshell()
|
||||||
{
|
{
|
||||||
maketitle("Updating dropshell on this computer...");
|
maketitle("Updating dropshell on this computer...");
|
||||||
|
|
||||||
|
configure_localbin();
|
||||||
|
configure_autocomplete();
|
||||||
|
|
||||||
// determine path to this executable
|
// determine path to this executable
|
||||||
std::filesystem::path dropshell_path = std::filesystem::canonical("/proc/self/exe");
|
std::filesystem::path exe_path = std::filesystem::canonical("/proc/self/exe");
|
||||||
std::filesystem::path parent_path = dropshell_path.parent_path();
|
std::filesystem::path parent_path = exe_path.parent_path();
|
||||||
|
|
||||||
// determine the architecture of the system
|
// determine the architecture of the system
|
||||||
std::string arch = shared_commands::get_arch();
|
std::string arch = shared_commands::get_arch();
|
||||||
|
|
||||||
std::string url = "https://gitea.jde.nz/public/dropshell/releases/download/latest/dropshell." + arch;
|
std::string url = "https://gitea.jde.nz/public/dropshell/releases/download/latest/dropshell." + arch;
|
||||||
|
|
||||||
// download new version, preserve permissions and ownership
|
// check that the user that owns the exe is the current user this process is running as.
|
||||||
std::string bash_script;
|
struct stat st;
|
||||||
bash_script += "docker run --rm -v " + parent_path.string() + ":/target";
|
if (stat(exe_path.c_str(), &st) != 0) {
|
||||||
bash_script += " gitea.jde.nz/public/debian-curl:latest";
|
error << "Failed to stat dropshell executable: " << strerror(errno) << std::endl;
|
||||||
bash_script += " sh -c \"";
|
|
||||||
bash_script += " curl -fsSL " + url + " -o /target/dropshell_temp &&";
|
|
||||||
bash_script += " chmod --reference=/target/dropshell /target/dropshell_temp &&";
|
|
||||||
bash_script += " chown --reference=/target/dropshell /target/dropshell_temp";
|
|
||||||
bash_script += "\"";
|
|
||||||
|
|
||||||
std::string cmd = "bash -c '" + bash_script + "'";
|
|
||||||
int rval = system(cmd.c_str());
|
|
||||||
if (rval != 0)
|
|
||||||
{
|
|
||||||
std::cerr << "Failed to download new version of dropshell." << std::endl;
|
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
uid_t current_uid = getuid();
|
||||||
|
if (st.st_uid != current_uid) {
|
||||||
|
warning << "Current user does not own the dropshell executable. Please run as the owner to update." << std::endl;
|
||||||
|
return -1;
|
||||||
|
}
|
||||||
|
|
||||||
|
shared_commands::cLocalTempFolder local_temp_folder;
|
||||||
|
std::filesystem::path temp_file = local_temp_folder.path() / "dropshell";
|
||||||
|
bool download_okay = download_file(url, temp_file);
|
||||||
|
if (!download_okay)
|
||||||
|
{
|
||||||
|
error << "Failed to download new version of dropshell." << std::endl;
|
||||||
|
return -1;
|
||||||
|
}
|
||||||
|
|
||||||
|
// make executable
|
||||||
|
chmod(temp_file.c_str(), 0755);
|
||||||
|
|
||||||
// check if the new version is the same as the old version
|
// check if the new version is the same as the old version
|
||||||
uint64_t new_hash = hash_file(parent_path / "dropshell_temp");
|
uint64_t new_hash = hash_file(temp_file);
|
||||||
uint64_t old_hash = hash_file(parent_path / "dropshell");
|
uint64_t old_hash = hash_file(exe_path);
|
||||||
if (new_hash == old_hash)
|
if (new_hash == old_hash)
|
||||||
{
|
{
|
||||||
std::cout << "Confirmed dropshell is the latest version." << std::endl;
|
info << "Confirmed dropshell is the latest version." << std::endl;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string runvercmd = (parent_path / "dropshell").string() + " version";
|
std::string runvercmd = exe_path.string() + " version";
|
||||||
std::string currentver = _exec(runvercmd.c_str());
|
std::string currentver = _exec(runvercmd.c_str());
|
||||||
runvercmd = (parent_path / "dropshell_temp").string() + " version";
|
runvercmd = temp_file.string() + " version";
|
||||||
std::string newver = _exec(runvercmd.c_str());
|
std::string newver = _exec(runvercmd.c_str());
|
||||||
|
|
||||||
if (currentver >= newver)
|
if (currentver >= newver)
|
||||||
@ -228,21 +302,15 @@ namespace dropshell
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
return 0;
|
// move the new version to the old version.
|
||||||
|
std::filesystem::rename(exe_path, exe_path.parent_path() / "dropshell.old");
|
||||||
|
std::filesystem::rename(temp_file, exe_path);
|
||||||
|
|
||||||
std::string bash_script_2 = "docker run --rm -v " + parent_path.string() + ":/target gitea.jde.nz/public/debian-curl:latest " +
|
// remove the old version.
|
||||||
"sh -c \"mv /target/dropshell_temp /target/dropshell\"";
|
std::filesystem::remove(exe_path.parent_path() / "dropshell.old");
|
||||||
rval = system(bash_script_2.c_str());
|
|
||||||
if (rval != 0)
|
|
||||||
{
|
|
||||||
error << "Failed to install new version of dropshell." << std::endl;
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
info << "Successfully updated " << dropshell_path << " to the latest " << arch << " version." << std::endl;
|
|
||||||
|
|
||||||
// execute the new version
|
// execute the new version
|
||||||
execlp("bash", "bash", "-c", (parent_path / "dropshell").c_str(), "install", (char *)nullptr);
|
execlp("bash", "bash", "-c", (exe_path.parent_path() / "dropshell").string() + "install", (char *)nullptr);
|
||||||
error << "Failed to execute new version of dropshell." << std::endl;
|
error << "Failed to execute new version of dropshell." << std::endl;
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
@ -281,7 +349,7 @@ namespace dropshell
|
|||||||
info << "Installing agent for user " << user.user << " on " << server.get_server_name() << std::endl;
|
info << "Installing agent for user " << user.user << " on " << server.get_server_name() << std::endl;
|
||||||
|
|
||||||
std::string agent_path = remotepath(server.get_server_name(),user.user).agent();
|
std::string agent_path = remotepath(server.get_server_name(),user.user).agent();
|
||||||
ASSERT(agent_path == user.dir+"/agent", "Agent path does not match user directory for "+user.user+"@" + server.get_server_name() + " : " + agent_path + " != " + user.dir);
|
ASSERT(agent_path == user.dir+"/agent", "Remote agent path does not match user directory for "+user.user+"@" + server.get_server_name() + " : " + agent_path + " != " + user.dir);
|
||||||
ASSERT(!agent_path.empty(), "Agent path is empty for " + user.user + "@" + server.get_server_name());
|
ASSERT(!agent_path.empty(), "Agent path is empty for " + user.user + "@" + server.get_server_name());
|
||||||
|
|
||||||
// now create the agent.
|
// now create the agent.
|
||||||
@ -295,7 +363,7 @@ namespace dropshell
|
|||||||
bool okay = execute_ssh_command(server.get_SSH_INFO(user.user), sCommand(agent_path, "agent-install.sh",{}), cMode::Defaults | cMode::NoBB64, nullptr);
|
bool okay = execute_ssh_command(server.get_SSH_INFO(user.user), sCommand(agent_path, "agent-install.sh",{}), cMode::Defaults | cMode::NoBB64, nullptr);
|
||||||
if (!okay)
|
if (!okay)
|
||||||
{
|
{
|
||||||
error << "ERROR: Failed to install remote agent on " << server.get_server_name() << std::endl;
|
error << "Failed to install remote agent on " << server.get_server_name() << std::endl;
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -147,7 +147,7 @@ void list_servers() {
|
|||||||
void show_server_details(const std::string& server_name) {
|
void show_server_details(const std::string& server_name) {
|
||||||
ServerConfig env(server_name);
|
ServerConfig env(server_name);
|
||||||
if (!env.is_valid()) {
|
if (!env.is_valid()) {
|
||||||
error << "Error: Invalid server environment file: " << server_name << std::endl;
|
error << "Invalid server environment file: " << server_name << std::endl;
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -60,7 +60,7 @@ namespace dropshell
|
|||||||
std::string local_backups_dir = localpath::backups();
|
std::string local_backups_dir = localpath::backups();
|
||||||
if (local_backups_dir.empty() || !std::filesystem::exists(local_backups_dir))
|
if (local_backups_dir.empty() || !std::filesystem::exists(local_backups_dir))
|
||||||
{
|
{
|
||||||
error << "Error: Local backups directory not found: " << local_backups_dir << std::endl;
|
error << "Local backups directory not found: " << local_backups_dir << std::endl;
|
||||||
return {};
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -150,19 +150,19 @@ namespace dropshell
|
|||||||
std::string local_backups_dir = localpath::backups();
|
std::string local_backups_dir = localpath::backups();
|
||||||
if (local_backups_dir.empty() || !std::filesystem::exists(local_backups_dir))
|
if (local_backups_dir.empty() || !std::filesystem::exists(local_backups_dir))
|
||||||
{
|
{
|
||||||
error << "Error: Local backups directory not found: " << local_backups_dir << std::endl;
|
error << "Local backups directory not found: " << local_backups_dir << std::endl;
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
std::string local_backup_file_path = (std::filesystem::path(local_backups_dir) / backup_details->get_filename()).string();
|
std::string local_backup_file_path = (std::filesystem::path(local_backups_dir) / backup_details->get_filename()).string();
|
||||||
if (!std::filesystem::exists(local_backup_file_path))
|
if (!std::filesystem::exists(local_backup_file_path))
|
||||||
{
|
{
|
||||||
error << "Error: Backup file not found at " << local_backup_file_path << std::endl;
|
error << "Backup file not found at " << local_backup_file_path << std::endl;
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (backup_details->get_template_name() != service_info.template_name)
|
if (backup_details->get_template_name() != service_info.template_name)
|
||||||
{
|
{
|
||||||
error << "Error: Backup template does not match service template. Can't restore." << std::endl;
|
error << "Backup template does not match service template. Can't restore." << std::endl;
|
||||||
info << "Backup template: " << backup_details->get_template_name() << std::endl;
|
info << "Backup template: " << backup_details->get_template_name() << std::endl;
|
||||||
info << "Service template: " << service_info.template_name << std::endl;
|
info << "Service template: " << service_info.template_name << std::endl;
|
||||||
return 1;
|
return 1;
|
||||||
|
@ -75,9 +75,9 @@ namespace dropshell
|
|||||||
// determine the architecture of the system
|
// determine the architecture of the system
|
||||||
std::string arch;
|
std::string arch;
|
||||||
#ifdef __aarch64__
|
#ifdef __aarch64__
|
||||||
arch = "arm64";
|
arch = "aarch64";
|
||||||
#elif __x86_64__
|
#elif __x86_64__
|
||||||
arch = "amd64";
|
arch = "x86_64";
|
||||||
#endif
|
#endif
|
||||||
return arch;
|
return arch;
|
||||||
}
|
}
|
||||||
@ -107,6 +107,22 @@ namespace dropshell
|
|||||||
return mPath;
|
return mPath;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
cLocalTempFolder::cLocalTempFolder()
|
||||||
|
{
|
||||||
|
mPath = std::filesystem::temp_directory_path() / random_alphanumeric_string(10);
|
||||||
|
std::filesystem::create_directories(mPath);
|
||||||
|
}
|
||||||
|
cLocalTempFolder::~cLocalTempFolder()
|
||||||
|
{
|
||||||
|
std::filesystem::remove_all(mPath);
|
||||||
|
}
|
||||||
|
std::filesystem::path cLocalTempFolder::path() const
|
||||||
|
{
|
||||||
|
return mPath;
|
||||||
|
}
|
||||||
|
|
||||||
// ------------------------------------------------------------------------------------------------
|
// ------------------------------------------------------------------------------------------------
|
||||||
// get_all_services_status : SHARED COMMAND
|
// get_all_services_status : SHARED COMMAND
|
||||||
// ------------------------------------------------------------------------------------------------
|
// ------------------------------------------------------------------------------------------------
|
||||||
|
@ -1,6 +1,8 @@
|
|||||||
#ifndef SHARED_COMMANDS_HPP
|
#ifndef SHARED_COMMANDS_HPP
|
||||||
#define SHARED_COMMANDS_HPP
|
#define SHARED_COMMANDS_HPP
|
||||||
|
|
||||||
|
#include <filesystem>
|
||||||
|
|
||||||
#include "servers.hpp"
|
#include "servers.hpp"
|
||||||
#include "command_registry.hpp"
|
#include "command_registry.hpp"
|
||||||
#include "servers.hpp"
|
#include "servers.hpp"
|
||||||
@ -40,6 +42,16 @@ namespace dropshell
|
|||||||
std::string mUser;
|
std::string mUser;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
class cLocalTempFolder
|
||||||
|
{
|
||||||
|
public:
|
||||||
|
cLocalTempFolder(); // create a temp folder on the local machine
|
||||||
|
~cLocalTempFolder(); // delete the temp folder on the local machine
|
||||||
|
std::filesystem::path path() const; // get the path to the temp folder on the local machine
|
||||||
|
private:
|
||||||
|
std::filesystem::path mPath;
|
||||||
|
};
|
||||||
|
|
||||||
bool rsync_tree_to_remote(
|
bool rsync_tree_to_remote(
|
||||||
const std::string &local_path,
|
const std::string &local_path,
|
||||||
const std::string &remote_path,
|
const std::string &remote_path,
|
||||||
|
@ -61,6 +61,12 @@ namespace dropshell
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (!legal_service_name(service))
|
||||||
|
{
|
||||||
|
error << "Service name contains illegal characters: " << service << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
LocalServiceInfo sinfo = get_service_info(server, service);
|
LocalServiceInfo sinfo = get_service_info(server, service);
|
||||||
if (!SIvalid(sinfo))
|
if (!SIvalid(sinfo))
|
||||||
{
|
{
|
||||||
|
@ -51,6 +51,12 @@ namespace dropshell
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (!legal_service_name(service))
|
||||||
|
{
|
||||||
|
error << "Service name contains illegal characters: " << service << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
// run the start script.
|
// run the start script.
|
||||||
bool started = server_env.run_remote_template_command(service, "start", {}, false, {});
|
bool started = server_env.run_remote_template_command(service, "start", {}, false, {});
|
||||||
|
|
||||||
@ -67,7 +73,7 @@ namespace dropshell
|
|||||||
{
|
{
|
||||||
if (ctx.args.size() < 2)
|
if (ctx.args.size() < 2)
|
||||||
{
|
{
|
||||||
std::cerr << "Error: Server name and service name are both required" << std::endl;
|
error << "Server name and service name are both required" << std::endl;
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -51,6 +51,12 @@ namespace dropshell
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (!legal_service_name(service))
|
||||||
|
{
|
||||||
|
error << "Service name contains illegal characters: " << service << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
// run the stop script.
|
// run the stop script.
|
||||||
bool stopped = server_env.run_remote_template_command(service, "stop", {}, false, {});
|
bool stopped = server_env.run_remote_template_command(service, "stop", {}, false, {});
|
||||||
|
|
||||||
|
@ -37,7 +37,7 @@ bool config::load_config() { // load json config file.
|
|||||||
}
|
}
|
||||||
catch (nlohmann::json::parse_error& ex)
|
catch (nlohmann::json::parse_error& ex)
|
||||||
{
|
{
|
||||||
std::cerr << "Error: Failed to parse config file: " << ex.what() << std::endl;
|
error << "Failed to parse config file: " << ex.what() << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -69,7 +69,7 @@ bool config::save_config(bool create_aux_directories)
|
|||||||
if (!mIsConfigSet)
|
if (!mIsConfigSet)
|
||||||
{
|
{
|
||||||
std::string homedir = localpath::current_user_home();
|
std::string homedir = localpath::current_user_home();
|
||||||
std::string dropshell_base = homedir + "/.local/dropshell_files";
|
std::string dropshell_base = homedir + "/.dropshell";
|
||||||
|
|
||||||
mConfig["server_definition_paths"] = {
|
mConfig["server_definition_paths"] = {
|
||||||
dropshell_base + "/servers"
|
dropshell_base + "/servers"
|
||||||
@ -81,6 +81,10 @@ bool config::save_config(bool create_aux_directories)
|
|||||||
"https://templates.dropshell.app"
|
"https://templates.dropshell.app"
|
||||||
};
|
};
|
||||||
mConfig["template_upload_token"] = "SECRETTOKEN";
|
mConfig["template_upload_token"] = "SECRETTOKEN";
|
||||||
|
|
||||||
|
mConfig["backups_path"] = {
|
||||||
|
dropshell_base + "/backups"
|
||||||
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
config_file << mConfig.dump(4);
|
config_file << mConfig.dump(4);
|
||||||
@ -116,14 +120,14 @@ bool config::is_agent_installed()
|
|||||||
return std::filesystem::exists(localfile::bb64());
|
return std::filesystem::exists(localfile::bb64());
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::string> config::get_template_registry_urls() {
|
std::vector<tRegistryEntry> config::get_template_registry_urls() {
|
||||||
nlohmann::json template_registry_urls = mConfig["template_registry_URLs"];
|
nlohmann::json template_registries = mConfig["template_registries"];
|
||||||
std::vector<std::string> urls;
|
std::vector<tRegistryEntry> registries;
|
||||||
for (auto &url : template_registry_urls) {
|
for (auto ®istry : template_registries) {
|
||||||
if (url.is_string() && !url.empty())
|
if (registry.is_object() && !registry.empty())
|
||||||
urls.push_back(url);
|
registries.push_back(tRegistryEntry(registry));
|
||||||
}
|
}
|
||||||
return urls;
|
return registries;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::string> config::get_local_template_paths()
|
std::vector<std::string> config::get_local_template_paths()
|
||||||
@ -163,16 +167,43 @@ std::string config::get_template_create_path()
|
|||||||
return paths[0];
|
return paths[0];
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string config::get_template_upload_url()
|
std::string config::get_backups_path()
|
||||||
{
|
{
|
||||||
std::vector<std::string> urls = get_template_registry_urls();
|
nlohmann::json backups_path = mConfig["backups_path"];
|
||||||
if (urls.empty())
|
if (backups_path.empty())
|
||||||
return "";
|
return "";
|
||||||
return urls[0];
|
if (backups_path.is_string())
|
||||||
|
return backups_path;
|
||||||
|
warning << "backups_path is not a string: " << backups_path << std::endl;
|
||||||
|
return "";
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string config::get_template_upload_token() {
|
dropshell::tRegistryEntry::tRegistryEntry(nlohmann::json json)
|
||||||
return mConfig["template_upload_token"];
|
{
|
||||||
|
valid = false;
|
||||||
|
if (json.is_object() && !json.empty()) {
|
||||||
|
for (auto &[key, value] : json.items()) {
|
||||||
|
if (value.is_string() && !value.empty())
|
||||||
|
switch (switchhash(key.c_str())) {
|
||||||
|
case switchhash("name"):
|
||||||
|
name = value;
|
||||||
|
break;
|
||||||
|
case switchhash("url"):
|
||||||
|
url = value;
|
||||||
|
break;
|
||||||
|
case switchhash("token"):
|
||||||
|
token = value;
|
||||||
|
break;
|
||||||
|
default:
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
valid = (!url.empty()&&!name.empty()); // token can be empty.
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
tRegistryEntry::~tRegistryEntry()
|
||||||
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
} // namespace dropshell
|
} // namespace dropshell
|
@ -4,10 +4,23 @@
|
|||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
#define JSON_INLINE_ALL
|
#define JSON_INLINE_ALL
|
||||||
#include "json.hpp"
|
#include <nlohmann/json.hpp>
|
||||||
|
|
||||||
namespace dropshell {
|
namespace dropshell {
|
||||||
|
|
||||||
|
class tRegistryEntry {
|
||||||
|
|
||||||
|
public:
|
||||||
|
tRegistryEntry(nlohmann::json json);
|
||||||
|
~tRegistryEntry();
|
||||||
|
|
||||||
|
public:
|
||||||
|
std::string name;
|
||||||
|
std::string url;
|
||||||
|
std::string token;
|
||||||
|
bool valid;
|
||||||
|
};
|
||||||
|
|
||||||
class config {
|
class config {
|
||||||
public:
|
public:
|
||||||
config();
|
config();
|
||||||
@ -19,14 +32,13 @@ class config {
|
|||||||
bool is_config_set() const;
|
bool is_config_set() const;
|
||||||
static bool is_agent_installed();
|
static bool is_agent_installed();
|
||||||
|
|
||||||
std::vector<std::string> get_template_registry_urls();
|
std::vector<tRegistryEntry> get_template_registry_urls();
|
||||||
std::vector<std::string> get_local_template_paths();
|
std::vector<std::string> get_local_template_paths();
|
||||||
std::vector<std::string> get_local_server_definition_paths();
|
std::vector<std::string> get_local_server_definition_paths();
|
||||||
|
|
||||||
std::string get_server_create_path();
|
std::string get_server_create_path();
|
||||||
std::string get_template_create_path();
|
std::string get_template_create_path();
|
||||||
std::string get_template_upload_url();
|
std::string get_backups_path();
|
||||||
std::string get_template_upload_token();
|
|
||||||
|
|
||||||
private:
|
private:
|
||||||
nlohmann::json mConfig;
|
nlohmann::json mConfig;
|
||||||
|
10509
source/src/contrib/httplib.hpp
Normal file
10509
source/src/contrib/httplib.hpp
Normal file
File diff suppressed because it is too large
Load Diff
@ -6,8 +6,8 @@ _dropshell_completions() {
|
|||||||
cur="${COMP_WORDS[COMP_CWORD]}"
|
cur="${COMP_WORDS[COMP_CWORD]}"
|
||||||
|
|
||||||
# call dropshell to get the list of possiblities for the current argument. Supply all previous arguments.
|
# call dropshell to get the list of possiblities for the current argument. Supply all previous arguments.
|
||||||
local completions=($(dropshell autocomplete "${COMP_WORDS[@]:1:${COMP_CWORD}-1}"))
|
mapfile -t completions < <(dropshell autocomplete "${COMP_WORDS[@]:1:${COMP_CWORD}-1}")
|
||||||
COMPREPLY=( $(compgen -W "${completions[*]}" -- ${cur}) )
|
mapfile -t COMPREPLY < <(compgen -W "${completions[*]}" -- "$cur")
|
||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -80,7 +80,7 @@ int main(int argc, char* argv[]) {
|
|||||||
|
|
||||||
}
|
}
|
||||||
catch (const std::exception& e) {
|
catch (const std::exception& e) {
|
||||||
std::cerr << "Error: " << e.what() << std::endl;
|
error << "Uncaught Exception: " << e.what() << std::endl;
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -104,7 +104,7 @@ namespace dropshell
|
|||||||
catch (const std::exception &e)
|
catch (const std::exception &e)
|
||||||
{
|
{
|
||||||
error << "Failed to parse " << server_json_path << std::endl;
|
error << "Failed to parse " << server_json_path << std::endl;
|
||||||
error << "Error: " << e.what() << std::endl;
|
error << "Exception: " << e.what() << std::endl;
|
||||||
|
|
||||||
mValid = false;
|
mValid = false;
|
||||||
}
|
}
|
||||||
@ -149,12 +149,20 @@ namespace dropshell
|
|||||||
|
|
||||||
std::string get_user_for_service(const std::string &server, const std::string &service)
|
std::string get_user_for_service(const std::string &server, const std::string &service)
|
||||||
{
|
{
|
||||||
|
if (!legal_service_name(service))
|
||||||
|
{
|
||||||
|
error << "Service name contains illegal characters: " + service << std::endl;
|
||||||
|
return "";
|
||||||
|
}
|
||||||
|
|
||||||
auto services_info = get_server_services_info(server);
|
auto services_info = get_server_services_info(server);
|
||||||
auto it = std::find_if(services_info.begin(), services_info.end(),
|
auto it = std::find_if(services_info.begin(), services_info.end(),
|
||||||
[&service](const LocalServiceInfo &si)
|
[&service](const LocalServiceInfo &si)
|
||||||
{ return si.service_name == service; });
|
{ return si.service_name == service; });
|
||||||
if (it != services_info.end() && SIvalid(*it))
|
if (it != services_info.end() && SIvalid(*it))
|
||||||
return it->user;
|
return it->user;
|
||||||
|
|
||||||
|
debug << "Couldn't find user for service \"" << service << "\" on server \"" << server << "\"" << std::endl;
|
||||||
return "";
|
return "";
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -241,7 +249,7 @@ namespace dropshell
|
|||||||
bool okay = execute_ssh_command(sshinfo, scommand, cMode::Silent);
|
bool okay = execute_ssh_command(sshinfo, scommand, cMode::Silent);
|
||||||
if (!okay)
|
if (!okay)
|
||||||
{
|
{
|
||||||
std::cerr << "Error: Required items not found on remote server: " << file_names_str << std::endl;
|
error << "Required items not found on remote server: " << file_names_str << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
return true;
|
return true;
|
||||||
@ -341,10 +349,12 @@ namespace dropshell
|
|||||||
std::map<std::string, std::string> env_vars;
|
std::map<std::string, std::string> env_vars;
|
||||||
if (!get_all_service_env_vars(mServerName, service_name, env_vars))
|
if (!get_all_service_env_vars(mServerName, service_name, env_vars))
|
||||||
{
|
{
|
||||||
std::cerr << "Error: Failed to get all service env vars for " << service_name << std::endl;
|
error << "Failed to get all service env vars for " << service_name << std::endl;
|
||||||
return std::nullopt;
|
return std::nullopt;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
env_vars["HOST_NAME"] = get_SSH_HOST();
|
||||||
|
|
||||||
std::string argstr = "";
|
std::string argstr = "";
|
||||||
for (const auto &arg : args)
|
for (const auto &arg : args)
|
||||||
{
|
{
|
||||||
@ -358,7 +368,7 @@ namespace dropshell
|
|||||||
|
|
||||||
if (sc.empty())
|
if (sc.empty())
|
||||||
{
|
{
|
||||||
std::cerr << "Error: Failed to construct command for " << service_name << " " << command << std::endl;
|
error << "Failed to construct command for " << service_name << " " << command << std::endl;
|
||||||
return std::nullopt;
|
return std::nullopt;
|
||||||
}
|
}
|
||||||
return sc;
|
return sc;
|
||||||
@ -388,7 +398,7 @@ namespace dropshell
|
|||||||
ServerConfig env(server_name);
|
ServerConfig env(server_name);
|
||||||
if (!env.is_valid())
|
if (!env.is_valid())
|
||||||
{
|
{
|
||||||
std::cerr << "Error: Invalid server environment file: " << entry.path().string() << std::endl;
|
error << "Invalid server environment file: " << entry.path().string() << std::endl;
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
servers.push_back(env);
|
servers.push_back(env);
|
||||||
@ -406,7 +416,7 @@ namespace dropshell
|
|||||||
std::string server_existing_dir = localpath::server(server_name);
|
std::string server_existing_dir = localpath::server(server_name);
|
||||||
if (!server_existing_dir.empty())
|
if (!server_existing_dir.empty())
|
||||||
{
|
{
|
||||||
error << "Error: Server name already exists: " << server_name << std::endl;
|
error << "Server name already exists: " << server_name << std::endl;
|
||||||
info << "Current server path: " << server_existing_dir << std::endl;
|
info << "Current server path: " << server_existing_dir << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
@ -415,7 +425,7 @@ namespace dropshell
|
|||||||
auto lsdp = gConfig().get_local_server_definition_paths();
|
auto lsdp = gConfig().get_local_server_definition_paths();
|
||||||
if (lsdp.empty() || lsdp[0].empty())
|
if (lsdp.empty() || lsdp[0].empty())
|
||||||
{
|
{
|
||||||
error << "Error: Local server definition path not found" << std::endl;
|
error << "Local server definition path not found" << std::endl;
|
||||||
info << "Run 'dropshell edit' to configure DropShell" << std::endl;
|
info << "Run 'dropshell edit' to configure DropShell" << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
@ -434,7 +444,9 @@ namespace dropshell
|
|||||||
server_env_file << " \"USER\": \"" << user << "\"," << std::endl;
|
server_env_file << " \"USER\": \"" << user << "\"," << std::endl;
|
||||||
server_env_file << " \"DIR\": \"" << "/home/" + user << "/.dropshell\"" << std::endl;
|
server_env_file << " \"DIR\": \"" << "/home/" + user << "/.dropshell\"" << std::endl;
|
||||||
server_env_file << " }" << std::endl;
|
server_env_file << " }" << std::endl;
|
||||||
server_env_file << " ]" << std::endl;
|
server_env_file << " ]," << std::endl;
|
||||||
|
server_env_file << " \"HAS_DOCKER\": \"true\"," << std::endl;
|
||||||
|
server_env_file << " \"DOCKER_ROOTLESS\": \"false\"" << std::endl;
|
||||||
server_env_file << "}" << std::endl;
|
server_env_file << "}" << std::endl;
|
||||||
server_env_file.close();
|
server_env_file.close();
|
||||||
|
|
||||||
|
@ -34,8 +34,8 @@ namespace dropshell
|
|||||||
std::vector<std::string> local_server_definition_paths = gConfig().get_local_server_definition_paths();
|
std::vector<std::string> local_server_definition_paths = gConfig().get_local_server_definition_paths();
|
||||||
if (local_server_definition_paths.empty())
|
if (local_server_definition_paths.empty())
|
||||||
{
|
{
|
||||||
std::cerr << "Error: No local server definition paths found" << std::endl;
|
error << "No local server definition paths found" << std::endl;
|
||||||
std::cerr << "Run 'dropshell edit' to configure DropShell" << std::endl;
|
info << "Run 'dropshell edit' to configure DropShell" << std::endl;
|
||||||
return services;
|
return services;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -80,6 +80,9 @@ namespace dropshell
|
|||||||
if (server_name.empty() || service_name.empty())
|
if (server_name.empty() || service_name.empty())
|
||||||
return LocalServiceInfo();
|
return LocalServiceInfo();
|
||||||
|
|
||||||
|
if (!legal_service_name(service_name))
|
||||||
|
return LocalServiceInfo();
|
||||||
|
|
||||||
service.service_name = service_name;
|
service.service_name = service_name;
|
||||||
|
|
||||||
service.local_service_path = localpath::service(server_name, service_name);
|
service.local_service_path = localpath::service(server_name, service_name);
|
||||||
@ -158,7 +161,7 @@ namespace dropshell
|
|||||||
auto service_info = get_service_info(server_name, service_name);
|
auto service_info = get_service_info(server_name, service_name);
|
||||||
if (service_info.local_template_path.empty())
|
if (service_info.local_template_path.empty())
|
||||||
{
|
{
|
||||||
std::cerr << "Error: Service not found: " << service_name << std::endl;
|
error << "Service not found: " << service_name << std::endl;
|
||||||
return commands;
|
return commands;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -183,7 +186,7 @@ namespace dropshell
|
|||||||
auto service_info = get_service_info(server_name, service_name);
|
auto service_info = get_service_info(server_name, service_name);
|
||||||
if (service_info.local_template_path.empty())
|
if (service_info.local_template_path.empty())
|
||||||
{
|
{
|
||||||
std::cerr << "Error: Service not found: " << service_name << std::endl;
|
error << "Service not found: " << service_name << std::endl;
|
||||||
return backups;
|
return backups;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -211,7 +214,7 @@ namespace dropshell
|
|||||||
|
|
||||||
if (localpath::service(server_name, service_name).empty() || !fs::exists(localpath::service(server_name, service_name)))
|
if (localpath::service(server_name, service_name).empty() || !fs::exists(localpath::service(server_name, service_name)))
|
||||||
{
|
{
|
||||||
std::cerr << "Error: Service not found: " << service_name << " on server " << server_name << std::endl;
|
error << "Service not found: " << service_name << " on server " << server_name << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -230,6 +233,13 @@ namespace dropshell
|
|||||||
warning << "Expected environment file not found: " << file << std::endl;
|
warning << "Expected environment file not found: " << file << std::endl;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
// add in some simple variables first, as others below may depend on/use these in bash.
|
||||||
|
// if we change these, we also need to update agent/_allservicesstatus.sh
|
||||||
|
all_env_vars["SERVER"] = server_name;
|
||||||
|
all_env_vars["SERVICE"] = service_name;
|
||||||
|
all_env_vars["DOCKER_CLI_HINTS"] = "false"; // turn off docker junk.
|
||||||
|
|
||||||
// Load environment files
|
// Load environment files
|
||||||
load_env_file(localfile::service_env(server_name, service_name));
|
load_env_file(localfile::service_env(server_name, service_name));
|
||||||
load_env_file(localfile::template_info_env(server_name, service_name));
|
load_env_file(localfile::template_info_env(server_name, service_name));
|
||||||
@ -243,13 +253,10 @@ namespace dropshell
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
// add in some handy variables.
|
// more additional, these depend on others above.
|
||||||
// if we change these, we also need to update agent/_allservicesstatus.sh
|
|
||||||
all_env_vars["CONFIG_PATH"] = remotepath(server_name, user).service_config(service_name);
|
all_env_vars["CONFIG_PATH"] = remotepath(server_name, user).service_config(service_name);
|
||||||
all_env_vars["SERVER"] = server_name;
|
|
||||||
all_env_vars["SERVICE"] = service_name;
|
|
||||||
all_env_vars["AGENT_PATH"] = remotepath(server_name, user).agent();
|
all_env_vars["AGENT_PATH"] = remotepath(server_name, user).agent();
|
||||||
all_env_vars["DOCKER_CLI_HINTS"] = "false"; // turn off docker junk.
|
|
||||||
|
|
||||||
// determine template name.
|
// determine template name.
|
||||||
auto it = all_env_vars.find("TEMPLATE");
|
auto it = all_env_vars.find("TEMPLATE");
|
||||||
@ -265,7 +272,7 @@ namespace dropshell
|
|||||||
template_info tinfo = gTemplateManager().get_template_info(it->second);
|
template_info tinfo = gTemplateManager().get_template_info(it->second);
|
||||||
if (!tinfo.is_set())
|
if (!tinfo.is_set())
|
||||||
{
|
{
|
||||||
std::cerr << "Error: Template '" << it->second << "' not found" << std::endl;
|
error << "Template '" << it->second << "' not found" << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -163,7 +163,7 @@
|
|||||||
ASSERT(mLoaded && mSources.size() > 0, "Template manager not loaded, or no template sources found.");
|
ASSERT(mLoaded && mSources.size() > 0, "Template manager not loaded, or no template sources found.");
|
||||||
template_source_interface* source = get_source(template_name);
|
template_source_interface* source = get_source(template_name);
|
||||||
if (!source) {
|
if (!source) {
|
||||||
std::cerr << "Error: Template '" << template_name << "' not found" << std::endl;
|
error << "Template '" << template_name << "' not found" << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
return source->template_command_exists(template_name, command);
|
return source->template_command_exists(template_name, command);
|
||||||
@ -171,25 +171,30 @@
|
|||||||
|
|
||||||
bool template_manager::create_template(const std::string &template_name) const
|
bool template_manager::create_template(const std::string &template_name) const
|
||||||
{
|
{
|
||||||
|
if (!legal_service_name(template_name)) {
|
||||||
|
error << "Template name contains illegal characters: " << template_name << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
// 1. Create a new directory in the user templates directory
|
// 1. Create a new directory in the user templates directory
|
||||||
std::vector<std::string> local_server_definition_paths = gConfig().get_local_server_definition_paths();
|
std::vector<std::string> local_server_definition_paths = gConfig().get_local_server_definition_paths();
|
||||||
|
|
||||||
if (local_server_definition_paths.empty()) {
|
if (local_server_definition_paths.empty()) {
|
||||||
std::cerr << "Error: No local server definition paths found" << std::endl;
|
error << "No local server definition paths found" << std::endl;
|
||||||
std::cerr << "Run 'dropshell edit' to configure DropShell" << std::endl;
|
info << "Run 'dropshell edit' to configure DropShell" << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
auto info = get_template_info(template_name);
|
auto tinfo = get_template_info(template_name);
|
||||||
if (info.is_set()) {
|
if (tinfo.is_set()) {
|
||||||
std::cerr << "Error: Template '" << template_name << "' already exists at " << info.locationID() << std::endl;
|
error << "Template '" << template_name << "' already exists at " << tinfo.locationID() << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
auto local_template_paths = gConfig().get_local_template_paths();
|
auto local_template_paths = gConfig().get_local_template_paths();
|
||||||
if (local_template_paths.empty()) {
|
if (local_template_paths.empty()) {
|
||||||
std::cerr << "Error: No local template paths found" << std::endl;
|
error << "No local template paths found" << std::endl;
|
||||||
std::cerr << "Run 'dropshell edit' to add one to the DropShell config" << std::endl;
|
info << "Run 'dropshell edit' to add one to the DropShell config" << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
std::string new_template_path = local_template_paths[0] + "/" + template_name;
|
std::string new_template_path = local_template_paths[0] + "/" + template_name;
|
||||||
@ -200,7 +205,7 @@
|
|||||||
// 2. Copy the example template from the system templates directory
|
// 2. Copy the example template from the system templates directory
|
||||||
auto example_info = gTemplateManager().get_template_info("example-nginx");
|
auto example_info = gTemplateManager().get_template_info("example-nginx");
|
||||||
if (!example_info.is_set()) {
|
if (!example_info.is_set()) {
|
||||||
std::cerr << "Error: Example template not found" << std::endl;
|
error << "Example template not found" << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
std::string example_template_path = example_info.local_template_path();
|
std::string example_template_path = example_info.local_template_path();
|
||||||
@ -222,7 +227,7 @@
|
|||||||
std::string replacement_line = "TEMPLATE=" + template_name;
|
std::string replacement_line = "TEMPLATE=" + template_name;
|
||||||
std::string service_env_path = new_template_path + "/config/" + filenames::template_info_env;
|
std::string service_env_path = new_template_path + "/config/" + filenames::template_info_env;
|
||||||
if (!replace_line_in_file(service_env_path, search_string, replacement_line)) {
|
if (!replace_line_in_file(service_env_path, search_string, replacement_line)) {
|
||||||
std::cerr << "Error: Failed to replace TEMPLATE= line in the " << filenames::template_info_env <<" file" << std::endl;
|
error << "Failed to replace TEMPLATE= line in the " << filenames::template_info_env <<" file" << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -278,7 +283,7 @@
|
|||||||
bool template_manager::required_file(std::string path, std::string template_name)
|
bool template_manager::required_file(std::string path, std::string template_name)
|
||||||
{
|
{
|
||||||
if (!std::filesystem::exists(path)) {
|
if (!std::filesystem::exists(path)) {
|
||||||
std::cerr << "Error: " << path << " file not found in template - REQUIRED." << template_name << std::endl;
|
error << path << " file not found in template - REQUIRED." << template_name << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
return true;
|
return true;
|
||||||
@ -323,7 +328,7 @@
|
|||||||
std::filesystem::path path = template_path + "/" + file;
|
std::filesystem::path path = template_path + "/" + file;
|
||||||
auto perms = std::filesystem::status(path).permissions();
|
auto perms = std::filesystem::status(path).permissions();
|
||||||
if ((perms & std::filesystem::perms::owner_exec) == std::filesystem::perms::none)
|
if ((perms & std::filesystem::perms::owner_exec) == std::filesystem::perms::none)
|
||||||
std::cerr << "Error: " << file << " is not executable" << std::endl;
|
error << file << " is not executable" << std::endl;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -347,18 +352,18 @@
|
|||||||
// determine template name.
|
// determine template name.
|
||||||
auto it = all_env_vars.find("TEMPLATE");
|
auto it = all_env_vars.find("TEMPLATE");
|
||||||
if (it == all_env_vars.end()) {
|
if (it == all_env_vars.end()) {
|
||||||
std::cerr << "Error: TEMPLATE variable not found in " << template_path << std::endl;
|
error << "TEMPLATE variable not found in " << template_path << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string env_template_name = it->second;
|
std::string env_template_name = it->second;
|
||||||
if (env_template_name.empty()) {
|
if (env_template_name.empty()) {
|
||||||
std::cerr << "Error: TEMPLATE variable is empty in " << template_path << std::endl;
|
error << "TEMPLATE variable is empty in " << template_path << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (env_template_name != template_name) {
|
if (env_template_name != template_name) {
|
||||||
std::cerr << "Error: TEMPLATE variable is wrong in " << template_path << std::endl;
|
error << "TEMPLATE variable is wrong in " << template_path << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -4,8 +4,9 @@
|
|||||||
#include <memory>
|
#include <memory>
|
||||||
#include <set>
|
#include <set>
|
||||||
|
|
||||||
|
#include "config.hpp"
|
||||||
#define JSON_INLINE_ALL
|
#define JSON_INLINE_ALL
|
||||||
#include "json.hpp"
|
#include <nlohmann/json.hpp>
|
||||||
|
|
||||||
|
|
||||||
namespace dropshell {
|
namespace dropshell {
|
||||||
@ -50,7 +51,7 @@ class template_source_interface {
|
|||||||
|
|
||||||
class template_source_registry : public template_source_interface {
|
class template_source_registry : public template_source_interface {
|
||||||
public:
|
public:
|
||||||
template_source_registry(std::string URL) : mURL(URL) {}
|
template_source_registry(tRegistryEntry registry) : mRegistry(registry) {}
|
||||||
|
|
||||||
~template_source_registry() {}
|
~template_source_registry() {}
|
||||||
|
|
||||||
@ -59,11 +60,11 @@ class template_source_registry : public template_source_interface {
|
|||||||
template_info get_template_info(const std::string& template_name);
|
template_info get_template_info(const std::string& template_name);
|
||||||
bool template_command_exists(const std::string& template_name,const std::string& command);
|
bool template_command_exists(const std::string& template_name,const std::string& command);
|
||||||
|
|
||||||
std::string get_description() { return "Registry: " + mURL; }
|
std::string get_description() { return "Registry: " + mRegistry.name + " (" + mRegistry.url + ")"; }
|
||||||
private:
|
private:
|
||||||
std::filesystem::path get_cache_dir();
|
std::filesystem::path get_cache_dir();
|
||||||
private:
|
private:
|
||||||
std::string mURL;
|
tRegistryEntry mRegistry;
|
||||||
std::vector<nlohmann::json> mTemplates; // cached list.
|
std::vector<nlohmann::json> mTemplates; // cached list.
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -7,86 +7,85 @@
|
|||||||
#include <string>
|
#include <string>
|
||||||
#include <filesystem>
|
#include <filesystem>
|
||||||
|
|
||||||
|
|
||||||
namespace fs = std::filesystem;
|
namespace fs = std::filesystem;
|
||||||
|
|
||||||
namespace dropshell {
|
namespace dropshell
|
||||||
|
{
|
||||||
|
|
||||||
|
namespace localfile
|
||||||
|
{
|
||||||
|
|
||||||
namespace localfile {
|
std::string dropshell_json()
|
||||||
|
{
|
||||||
std::string dropshell_json() {
|
return localpath::dropshell_dir() + "/" + filenames::dropshell_json;
|
||||||
// Try ~/.config/dropshell/dropshell.json
|
|
||||||
std::string homedir = localpath::current_user_home();
|
|
||||||
if (!homedir.empty()) {
|
|
||||||
fs::path user_path = fs::path(homedir) / ".config" / "dropshell" / filenames::dropshell_json;
|
|
||||||
return user_path.string();
|
|
||||||
}
|
}
|
||||||
return std::string();
|
|
||||||
}
|
|
||||||
|
|
||||||
std::string server_json(const std::string &server_name) {
|
std::string server_json(const std::string &server_name)
|
||||||
std::string serverpath = localpath::server(server_name);
|
{
|
||||||
return (serverpath.empty() ? "" : (fs::path(serverpath) / filenames::server_json).string());
|
std::string serverpath = localpath::server(server_name);
|
||||||
}
|
return (serverpath.empty() ? "" : (fs::path(serverpath) / filenames::server_json).string());
|
||||||
|
}
|
||||||
|
|
||||||
std::string service_env(const std::string &server_name, const std::string &service_name) {
|
std::string service_env(const std::string &server_name, const std::string &service_name)
|
||||||
std::string servicepath = localpath::service(server_name, service_name);
|
{
|
||||||
return (servicepath.empty() ? "" : (fs::path(servicepath) / filenames::service_env).string());
|
std::string servicepath = localpath::service(server_name, service_name);
|
||||||
}
|
return (servicepath.empty() ? "" : (fs::path(servicepath) / filenames::service_env).string());
|
||||||
|
}
|
||||||
|
|
||||||
std::string template_info_env(const std::string &server_name, const std::string &service_name)
|
std::string template_info_env(const std::string &server_name, const std::string &service_name)
|
||||||
|
{
|
||||||
|
std::string servicepath = localpath::service(server_name, service_name);
|
||||||
|
return (servicepath.empty() ? "" : (fs::path(servicepath) / filenames::template_info_env).string());
|
||||||
|
}
|
||||||
|
|
||||||
|
std::string template_example()
|
||||||
|
{
|
||||||
|
return localpath::agent_local() + "/template_example";
|
||||||
|
}
|
||||||
|
|
||||||
|
std::string bb64()
|
||||||
|
{
|
||||||
|
return localpath::agent_local() + "/bb64";
|
||||||
|
}
|
||||||
|
|
||||||
|
} // namespace localfile
|
||||||
|
|
||||||
|
// ------------------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
namespace localpath
|
||||||
{
|
{
|
||||||
std::string servicepath = localpath::service(server_name, service_name);
|
|
||||||
return (servicepath.empty() ? "" : (fs::path(servicepath) / filenames::template_info_env).string());
|
|
||||||
}
|
|
||||||
|
|
||||||
std::string template_example()
|
std::string dropshell_dir()
|
||||||
{
|
{
|
||||||
return localpath::agent_local() + "/template_example";
|
return current_user_home() + "/.dropshell";
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string bb64()
|
std::string server(const std::string &server_name)
|
||||||
{
|
{
|
||||||
return localpath::agent_local() + "/bb64";
|
for (std::filesystem::path dir : gConfig().get_local_server_definition_paths())
|
||||||
}
|
|
||||||
|
|
||||||
} // namespace localfile
|
|
||||||
|
|
||||||
|
|
||||||
// ------------------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
namespace localpath {
|
|
||||||
std::string server(const std::string &server_name) {
|
|
||||||
for (std::filesystem::path dir : gConfig().get_local_server_definition_paths())
|
|
||||||
if (fs::exists(dir / server_name))
|
if (fs::exists(dir / server_name))
|
||||||
return dir / server_name;
|
return dir / server_name;
|
||||||
|
|
||||||
return "";
|
return "";
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string service(const std::string &server_name, const std::string &service_name) {
|
std::string service(const std::string &server_name, const std::string &service_name)
|
||||||
|
{
|
||||||
std::string serverpath = localpath::server(server_name);
|
std::string serverpath = localpath::server(server_name);
|
||||||
return ((serverpath.empty() || service_name.empty()) ? "" : (serverpath+"/"+service_name));
|
return ((serverpath.empty() || service_name.empty()) ? "" : (serverpath + "/" + service_name));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string remote_versions(const std::string &server_name, const std::string &service_name)
|
|
||||||
{
|
|
||||||
std::string template_cache_path = localpath::template_cache();
|
|
||||||
return ((template_cache_path.empty() || service_name.empty()) ? "" :
|
|
||||||
(template_cache_path+"/remote_versions/"+service_name+".json"));
|
|
||||||
}
|
|
||||||
std::string agent_local()
|
std::string agent_local()
|
||||||
{
|
{
|
||||||
return current_user_home()+"/.local/dropshell_agent/agent-local";
|
return dropshell_dir() + "/agent-local";
|
||||||
}
|
}
|
||||||
std::string agent_remote()
|
std::string agent_remote()
|
||||||
{
|
{
|
||||||
return current_user_home() + "/.local/dropshell_agent/agent-remote";
|
return dropshell_dir() + "/agent-remote";
|
||||||
}
|
}
|
||||||
std::string current_user_home()
|
std::string current_user_home()
|
||||||
{
|
{
|
||||||
char * homedir = std::getenv("HOME");
|
char *homedir = std::getenv("HOME");
|
||||||
if (homedir)
|
if (homedir)
|
||||||
{
|
{
|
||||||
std::filesystem::path homedir_path(homedir);
|
std::filesystem::path homedir_path(homedir);
|
||||||
@ -96,37 +95,32 @@ namespace localpath {
|
|||||||
return std::string();
|
return std::string();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string dropshell_files()
|
|
||||||
{
|
|
||||||
return current_user_home() + "/.local/dropshell_files";
|
|
||||||
return std::string();
|
|
||||||
}
|
|
||||||
|
|
||||||
std::string backups()
|
std::string backups()
|
||||||
{
|
{
|
||||||
return dropshell_files() + "/backups";
|
if (!gConfig().is_config_set())
|
||||||
|
return "";
|
||||||
|
return gConfig().get_backups_path();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string temp_files()
|
std::string temp_files()
|
||||||
{
|
{
|
||||||
return dropshell_files() + "/temp_files";
|
return dropshell_dir() + "/temp_files";
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string template_cache()
|
std::string template_cache()
|
||||||
{
|
{
|
||||||
return dropshell_files() + "template_cache";
|
return dropshell_dir() + "/template_cache";
|
||||||
}
|
}
|
||||||
|
|
||||||
bool create_directories()
|
bool create_directories()
|
||||||
{
|
{
|
||||||
std::vector<std::filesystem::path> paths = {
|
std::vector<std::filesystem::path> paths = {
|
||||||
dropshell_files(),
|
dropshell_dir(),
|
||||||
agent_local(),
|
agent_local(),
|
||||||
agent_remote(),
|
agent_remote(),
|
||||||
template_cache(),
|
template_cache(),
|
||||||
backups(),
|
backups(),
|
||||||
temp_files()
|
temp_files()};
|
||||||
};
|
|
||||||
for (auto &p : gConfig().get_local_server_definition_paths())
|
for (auto &p : gConfig().get_local_server_definition_paths())
|
||||||
paths.push_back(p);
|
paths.push_back(p);
|
||||||
|
|
||||||
@ -139,7 +133,7 @@ namespace localpath {
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
} // namespace localpath
|
} // namespace localpath
|
||||||
|
|
||||||
//------------------------------------------------------------------------------------------------
|
//------------------------------------------------------------------------------------------------
|
||||||
// remote paths
|
// remote paths
|
||||||
@ -159,21 +153,26 @@ namespace localpath {
|
|||||||
// |-- service.env (default service config)
|
// |-- service.env (default service config)
|
||||||
// |-- (other config files for specific server&service)
|
// |-- (other config files for specific server&service)
|
||||||
|
|
||||||
|
remotefile::remotefile(const std::string &server_name, const std::string &user) : mServer_name(server_name), mUser(user) {}
|
||||||
remotefile::remotefile(const std::string &server_name, const std::string &user) :
|
|
||||||
mServer_name(server_name), mUser(user) {}
|
|
||||||
|
|
||||||
std::string remotefile::service_env(const std::string &service_name) const
|
std::string remotefile::service_env(const std::string &service_name) const
|
||||||
{
|
{
|
||||||
return remotepath(mServer_name,mUser).service_config(service_name) + "/" + filenames::service_env;
|
return remotepath(mServer_name, mUser).service_config(service_name) + "/" + filenames::service_env;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
remotepath::remotepath(const std::string &server_name, const std::string &user) : mServer_name(server_name), mUser(user) {}
|
remotepath::remotepath(const std::string &server_name, const std::string &user) : mServer_name(server_name), mUser(user) {}
|
||||||
|
|
||||||
std::string remotepath::DROPSHELL_DIR() const
|
std::string remotepath::DROPSHELL_DIR() const
|
||||||
{
|
{
|
||||||
return ServerConfig(mServer_name).get_user_dir(mUser);
|
try
|
||||||
|
{
|
||||||
|
return ServerConfig(mServer_name).get_user_dir(mUser);
|
||||||
|
} catch (const std::exception &e)
|
||||||
|
{
|
||||||
|
error << "Failed to get remote dropshell directory for " << mServer_name << "@" << mUser << std::endl;
|
||||||
|
error << "Exception: " << e.what() << std::endl;
|
||||||
|
return "";
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string remotepath::services() const
|
std::string remotepath::services() const
|
||||||
@ -218,22 +217,21 @@ namespace localpath {
|
|||||||
return (dsp.empty() ? "" : (dsp + "/agent"));
|
return (dsp.empty() ? "" : (dsp + "/agent"));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ------------------------------------------------------------------------------------------
|
||||||
|
// Utility functions
|
||||||
|
|
||||||
// ------------------------------------------------------------------------------------------
|
std::string get_parent(const std::filesystem::path path)
|
||||||
// Utility functions
|
{
|
||||||
|
if (path.empty())
|
||||||
|
return std::string();
|
||||||
|
return path.parent_path().string();
|
||||||
|
}
|
||||||
|
|
||||||
std::string get_parent(const std::filesystem::path path)
|
std::string get_child(const std::filesystem::path path)
|
||||||
{
|
{
|
||||||
if (path.empty())
|
if (path.empty())
|
||||||
return std::string();
|
return std::string();
|
||||||
return path.parent_path().string();
|
return path.filename().string();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string get_child(const std::filesystem::path path)
|
|
||||||
{
|
|
||||||
if (path.empty())
|
|
||||||
return std::string();
|
|
||||||
return path.filename().string();
|
|
||||||
}
|
|
||||||
|
|
||||||
} // namespace dropshell
|
} // namespace dropshell
|
||||||
|
@ -10,20 +10,15 @@ namespace dropshell {
|
|||||||
|
|
||||||
//------------------------------------------------------------------------------------------------
|
//------------------------------------------------------------------------------------------------
|
||||||
// local user config directories
|
// local user config directories
|
||||||
|
|
||||||
// ~/.config/dropshell/dropshell.json
|
// ~/.dropshell
|
||||||
|
// |-- dropshell.json
|
||||||
// ~/.local/dropshell_agent
|
|
||||||
// |-- agent-local
|
// |-- agent-local
|
||||||
// |-- agent-install.sh
|
// |-- agent-install.sh
|
||||||
// |-- bb64 (only used locally, as it's for the local machine's architecture!)
|
// |-- bb64 (only used locally, as it's for the local machine's architecture!)
|
||||||
// |-- template_example
|
// |-- template_example
|
||||||
// |-- agent-remote
|
// |-- agent-remote
|
||||||
// |-- (remote agent files, including _allservicesstatus.sh)
|
// |-- (remote agent files, including _allservicesstatus.sh)
|
||||||
|
|
||||||
// ~/.local/dropshell_files
|
|
||||||
// |-- backups
|
|
||||||
// |-- katie-_-squashkiwi-_-squashkiwi-test-_-2025-04-28_21-23-59.tgz
|
|
||||||
// |-- temp_files
|
// |-- temp_files
|
||||||
// |-- template_cache
|
// |-- template_cache
|
||||||
// |-- templates
|
// |-- templates
|
||||||
@ -35,6 +30,10 @@ namespace dropshell {
|
|||||||
// | |-- .template_info.env
|
// | |-- .template_info.env
|
||||||
// | |-- (...other service config files...)
|
// | |-- (...other service config files...)
|
||||||
|
|
||||||
|
// backups_path
|
||||||
|
// |-- katie-_-squashkiwi-_-squashkiwi-test-_-2025-04-28_21-23-59.tgz
|
||||||
|
|
||||||
|
|
||||||
// server_definition_path
|
// server_definition_path
|
||||||
// |-- <server_name>
|
// |-- <server_name>
|
||||||
// |-- server.json
|
// |-- server.json
|
||||||
@ -53,7 +52,6 @@ namespace dropshell {
|
|||||||
} // namespace filenames.
|
} // namespace filenames.
|
||||||
|
|
||||||
namespace localfile {
|
namespace localfile {
|
||||||
// ~/.config/dropshell/dropshell.json
|
|
||||||
std::string dropshell_json();
|
std::string dropshell_json();
|
||||||
std::string server_json(const std::string &server_name);
|
std::string server_json(const std::string &server_name);
|
||||||
std::string service_env(const std::string &server_name, const std::string &service_name);
|
std::string service_env(const std::string &server_name, const std::string &service_name);
|
||||||
@ -63,16 +61,15 @@ namespace dropshell {
|
|||||||
} // namespace localfile
|
} // namespace localfile
|
||||||
|
|
||||||
namespace localpath {
|
namespace localpath {
|
||||||
|
std::string dropshell_dir();
|
||||||
|
|
||||||
std::string server(const std::string &server_name);
|
std::string server(const std::string &server_name);
|
||||||
std::string service(const std::string &server_name, const std::string &service_name);
|
std::string service(const std::string &server_name, const std::string &service_name);
|
||||||
|
|
||||||
std::string remote_versions(const std::string &server_name, const std::string &service_name);
|
|
||||||
|
|
||||||
std::string agent_local();
|
std::string agent_local();
|
||||||
std::string agent_remote();
|
std::string agent_remote();
|
||||||
std::string current_user_home();
|
std::string current_user_home();
|
||||||
|
|
||||||
std::string dropshell_files();
|
|
||||||
std::string backups();
|
std::string backups();
|
||||||
std::string temp_files();
|
std::string temp_files();
|
||||||
std::string template_cache();
|
std::string template_cache();
|
||||||
|
@ -183,7 +183,7 @@ namespace dropshell
|
|||||||
|
|
||||||
if (!rval && !hasFlag(mode, cMode::Silent))
|
if (!rval && !hasFlag(mode, cMode::Silent))
|
||||||
{
|
{
|
||||||
error << "Error: Failed to execute ssh command:" << std::endl;
|
error << "Failed to execute ssh command:" << std::endl;
|
||||||
debug << ssh_cmd.str() + " " + remote_command.construct_cmd(remote_bb64_path) << std::endl;
|
debug << ssh_cmd.str() + " " + remote_command.construct_cmd(remote_bb64_path) << std::endl;
|
||||||
}
|
}
|
||||||
return rval;
|
return rval;
|
||||||
|
@ -3,6 +3,8 @@
|
|||||||
#define XXH_INLINE_ALL
|
#define XXH_INLINE_ALL
|
||||||
#include "contrib/xxhash.hpp"
|
#include "contrib/xxhash.hpp"
|
||||||
|
|
||||||
|
#include "output.hpp"
|
||||||
|
|
||||||
#include <fstream>
|
#include <fstream>
|
||||||
#include <filesystem>
|
#include <filesystem>
|
||||||
#include <iostream>
|
#include <iostream>
|
||||||
@ -13,7 +15,7 @@ uint64_t hash_file(const std::string &path) {
|
|||||||
// Create hash state
|
// Create hash state
|
||||||
XXH64_state_t* const state = XXH64_createState();
|
XXH64_state_t* const state = XXH64_createState();
|
||||||
if (state == nullptr) {
|
if (state == nullptr) {
|
||||||
std::cerr << "Failed to create hash state" << std::endl;
|
error << "Failed to create hash state" << std::endl;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -24,7 +26,7 @@ uint64_t hash_file(const std::string &path) {
|
|||||||
// Open file
|
// Open file
|
||||||
std::ifstream file(path, std::ios::binary);
|
std::ifstream file(path, std::ios::binary);
|
||||||
if (!file.is_open()) {
|
if (!file.is_open()) {
|
||||||
std::cerr << "Failed to open file: " << path << std::endl;
|
error << "Failed to open file: " << path << std::endl;
|
||||||
XXH64_freeState(state);
|
XXH64_freeState(state);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
@ -34,7 +36,7 @@ uint64_t hash_file(const std::string &path) {
|
|||||||
char buffer[buffer_size];
|
char buffer[buffer_size];
|
||||||
while (file.read(buffer, buffer_size)) {
|
while (file.read(buffer, buffer_size)) {
|
||||||
if (XXH64_update(state, buffer, file.gcount()) == XXH_ERROR) {
|
if (XXH64_update(state, buffer, file.gcount()) == XXH_ERROR) {
|
||||||
std::cerr << "Failed to update hash" << std::endl;
|
error << "Failed to update hash" << std::endl;
|
||||||
XXH64_freeState(state);
|
XXH64_freeState(state);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
@ -43,7 +45,7 @@ uint64_t hash_file(const std::string &path) {
|
|||||||
// Handle any remaining bytes
|
// Handle any remaining bytes
|
||||||
if (file.gcount() > 0) {
|
if (file.gcount() > 0) {
|
||||||
if (XXH64_update(state, buffer, file.gcount()) == XXH_ERROR) {
|
if (XXH64_update(state, buffer, file.gcount()) == XXH_ERROR) {
|
||||||
std::cerr << "Failed to update hash" << std::endl;
|
error << "Failed to update hash" << std::endl;
|
||||||
XXH64_freeState(state);
|
XXH64_freeState(state);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
@ -59,14 +61,14 @@ uint64_t hash_directory_recursive(const std::string &path) {
|
|||||||
// Create hash state
|
// Create hash state
|
||||||
XXH64_state_t* const state = XXH64_createState();
|
XXH64_state_t* const state = XXH64_createState();
|
||||||
if (state == nullptr) {
|
if (state == nullptr) {
|
||||||
std::cerr << "Failed to create hash state" << std::endl;
|
error << "Failed to create hash state" << std::endl;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Initialize state with seed 0
|
// Initialize state with seed 0
|
||||||
XXH64_hash_t const seed = 0; /* or any other value */
|
XXH64_hash_t const seed = 0; /* or any other value */
|
||||||
if (XXH64_reset(state, seed) == XXH_ERROR) {
|
if (XXH64_reset(state, seed) == XXH_ERROR) {
|
||||||
std::cerr << "Failed to reset hash state" << std::endl;
|
error << "Failed to reset hash state" << std::endl;
|
||||||
XXH64_freeState(state);
|
XXH64_freeState(state);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
@ -81,7 +83,7 @@ uint64_t hash_directory_recursive(const std::string &path) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
} catch (const std::filesystem::filesystem_error& e) {
|
} catch (const std::filesystem::filesystem_error& e) {
|
||||||
std::cerr << "Filesystem error: " << e.what() << std::endl;
|
error << "Filesystem error: " << e.what() << std::endl;
|
||||||
XXH64_freeState(state);
|
XXH64_freeState(state);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
@ -94,7 +96,7 @@ uint64_t hash_directory_recursive(const std::string &path) {
|
|||||||
|
|
||||||
uint64_t hash_path(const std::string &path) {
|
uint64_t hash_path(const std::string &path) {
|
||||||
if (!std::filesystem::exists(path)) {
|
if (!std::filesystem::exists(path)) {
|
||||||
std::cerr << "Path does not exist: " << path << std::endl;
|
error << "Path does not exist: " << path << std::endl;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -103,28 +105,28 @@ uint64_t hash_path(const std::string &path) {
|
|||||||
} else if (std::filesystem::is_regular_file(path)) {
|
} else if (std::filesystem::is_regular_file(path)) {
|
||||||
return hash_file(path);
|
return hash_file(path);
|
||||||
} else {
|
} else {
|
||||||
std::cerr << "Path is neither a file nor a directory: " << path << std::endl;
|
error << "Path is neither a file nor a directory: " << path << std::endl;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void hash_demo(const std::string & path)
|
void hash_demo(const std::string & path)
|
||||||
{
|
{
|
||||||
std::cout << "Hashing path: " << path << std::endl;
|
info << "Hashing path: " << path << std::endl;
|
||||||
auto start = std::chrono::high_resolution_clock::now();
|
auto start = std::chrono::high_resolution_clock::now();
|
||||||
XXH64_hash_t hash = hash_path(path);
|
XXH64_hash_t hash = hash_path(path);
|
||||||
auto end = std::chrono::high_resolution_clock::now();
|
auto end = std::chrono::high_resolution_clock::now();
|
||||||
auto duration = std::chrono::duration_cast<std::chrono::milliseconds>(end - start);
|
auto duration = std::chrono::duration_cast<std::chrono::milliseconds>(end - start);
|
||||||
std::cout << "Hash: " << hash << " (took " << duration.count() << "ms)" << std::endl;
|
info << "Hash: " << hash << " (took " << duration.count() << "ms)" << std::endl;
|
||||||
}
|
}
|
||||||
|
|
||||||
int hash_demo_raw(const std::string & path)
|
int hash_demo_raw(const std::string & path)
|
||||||
{
|
{
|
||||||
if (!std::filesystem::exists(path)) {
|
if (!std::filesystem::exists(path)) {
|
||||||
std::cout << 0 <<std::endl; return 1;
|
info << 0 <<std::endl; return 1;
|
||||||
}
|
}
|
||||||
XXH64_hash_t hash = hash_path(path);
|
XXH64_hash_t hash = hash_path(path);
|
||||||
std::cout << hash << std::endl;
|
info << hash << std::endl;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
File diff suppressed because it is too large
Load Diff
@ -1,187 +0,0 @@
|
|||||||
// __ _____ _____ _____
|
|
||||||
// __| | __| | | | JSON for Modern C++
|
|
||||||
// | | |__ | | | | | | version 3.12.0
|
|
||||||
// |_____|_____|_____|_|___| https://github.com/nlohmann/json
|
|
||||||
//
|
|
||||||
// SPDX-FileCopyrightText: 2013 - 2025 Niels Lohmann <https://nlohmann.me>
|
|
||||||
// SPDX-License-Identifier: MIT
|
|
||||||
|
|
||||||
#ifndef INCLUDE_NLOHMANN_JSON_FWD_HPP_
|
|
||||||
#define INCLUDE_NLOHMANN_JSON_FWD_HPP_
|
|
||||||
|
|
||||||
#include <cstdint> // int64_t, uint64_t
|
|
||||||
#include <map> // map
|
|
||||||
#include <memory> // allocator
|
|
||||||
#include <string> // string
|
|
||||||
#include <vector> // vector
|
|
||||||
|
|
||||||
// #include <nlohmann/detail/abi_macros.hpp>
|
|
||||||
// __ _____ _____ _____
|
|
||||||
// __| | __| | | | JSON for Modern C++
|
|
||||||
// | | |__ | | | | | | version 3.12.0
|
|
||||||
// |_____|_____|_____|_|___| https://github.com/nlohmann/json
|
|
||||||
//
|
|
||||||
// SPDX-FileCopyrightText: 2013 - 2025 Niels Lohmann <https://nlohmann.me>
|
|
||||||
// SPDX-License-Identifier: MIT
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
// This file contains all macro definitions affecting or depending on the ABI
|
|
||||||
|
|
||||||
#ifndef JSON_SKIP_LIBRARY_VERSION_CHECK
|
|
||||||
#if defined(NLOHMANN_JSON_VERSION_MAJOR) && defined(NLOHMANN_JSON_VERSION_MINOR) && defined(NLOHMANN_JSON_VERSION_PATCH)
|
|
||||||
#if NLOHMANN_JSON_VERSION_MAJOR != 3 || NLOHMANN_JSON_VERSION_MINOR != 12 || NLOHMANN_JSON_VERSION_PATCH != 0
|
|
||||||
#warning "Already included a different version of the library!"
|
|
||||||
#endif
|
|
||||||
#endif
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#define NLOHMANN_JSON_VERSION_MAJOR 3 // NOLINT(modernize-macro-to-enum)
|
|
||||||
#define NLOHMANN_JSON_VERSION_MINOR 12 // NOLINT(modernize-macro-to-enum)
|
|
||||||
#define NLOHMANN_JSON_VERSION_PATCH 0 // NOLINT(modernize-macro-to-enum)
|
|
||||||
|
|
||||||
#ifndef JSON_DIAGNOSTICS
|
|
||||||
#define JSON_DIAGNOSTICS 0
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#ifndef JSON_DIAGNOSTIC_POSITIONS
|
|
||||||
#define JSON_DIAGNOSTIC_POSITIONS 0
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#ifndef JSON_USE_LEGACY_DISCARDED_VALUE_COMPARISON
|
|
||||||
#define JSON_USE_LEGACY_DISCARDED_VALUE_COMPARISON 0
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#if JSON_DIAGNOSTICS
|
|
||||||
#define NLOHMANN_JSON_ABI_TAG_DIAGNOSTICS _diag
|
|
||||||
#else
|
|
||||||
#define NLOHMANN_JSON_ABI_TAG_DIAGNOSTICS
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#if JSON_DIAGNOSTIC_POSITIONS
|
|
||||||
#define NLOHMANN_JSON_ABI_TAG_DIAGNOSTIC_POSITIONS _dp
|
|
||||||
#else
|
|
||||||
#define NLOHMANN_JSON_ABI_TAG_DIAGNOSTIC_POSITIONS
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#if JSON_USE_LEGACY_DISCARDED_VALUE_COMPARISON
|
|
||||||
#define NLOHMANN_JSON_ABI_TAG_LEGACY_DISCARDED_VALUE_COMPARISON _ldvcmp
|
|
||||||
#else
|
|
||||||
#define NLOHMANN_JSON_ABI_TAG_LEGACY_DISCARDED_VALUE_COMPARISON
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#ifndef NLOHMANN_JSON_NAMESPACE_NO_VERSION
|
|
||||||
#define NLOHMANN_JSON_NAMESPACE_NO_VERSION 0
|
|
||||||
#endif
|
|
||||||
|
|
||||||
// Construct the namespace ABI tags component
|
|
||||||
#define NLOHMANN_JSON_ABI_TAGS_CONCAT_EX(a, b, c) json_abi ## a ## b ## c
|
|
||||||
#define NLOHMANN_JSON_ABI_TAGS_CONCAT(a, b, c) \
|
|
||||||
NLOHMANN_JSON_ABI_TAGS_CONCAT_EX(a, b, c)
|
|
||||||
|
|
||||||
#define NLOHMANN_JSON_ABI_TAGS \
|
|
||||||
NLOHMANN_JSON_ABI_TAGS_CONCAT( \
|
|
||||||
NLOHMANN_JSON_ABI_TAG_DIAGNOSTICS, \
|
|
||||||
NLOHMANN_JSON_ABI_TAG_LEGACY_DISCARDED_VALUE_COMPARISON, \
|
|
||||||
NLOHMANN_JSON_ABI_TAG_DIAGNOSTIC_POSITIONS)
|
|
||||||
|
|
||||||
// Construct the namespace version component
|
|
||||||
#define NLOHMANN_JSON_NAMESPACE_VERSION_CONCAT_EX(major, minor, patch) \
|
|
||||||
_v ## major ## _ ## minor ## _ ## patch
|
|
||||||
#define NLOHMANN_JSON_NAMESPACE_VERSION_CONCAT(major, minor, patch) \
|
|
||||||
NLOHMANN_JSON_NAMESPACE_VERSION_CONCAT_EX(major, minor, patch)
|
|
||||||
|
|
||||||
#if NLOHMANN_JSON_NAMESPACE_NO_VERSION
|
|
||||||
#define NLOHMANN_JSON_NAMESPACE_VERSION
|
|
||||||
#else
|
|
||||||
#define NLOHMANN_JSON_NAMESPACE_VERSION \
|
|
||||||
NLOHMANN_JSON_NAMESPACE_VERSION_CONCAT(NLOHMANN_JSON_VERSION_MAJOR, \
|
|
||||||
NLOHMANN_JSON_VERSION_MINOR, \
|
|
||||||
NLOHMANN_JSON_VERSION_PATCH)
|
|
||||||
#endif
|
|
||||||
|
|
||||||
// Combine namespace components
|
|
||||||
#define NLOHMANN_JSON_NAMESPACE_CONCAT_EX(a, b) a ## b
|
|
||||||
#define NLOHMANN_JSON_NAMESPACE_CONCAT(a, b) \
|
|
||||||
NLOHMANN_JSON_NAMESPACE_CONCAT_EX(a, b)
|
|
||||||
|
|
||||||
#ifndef NLOHMANN_JSON_NAMESPACE
|
|
||||||
#define NLOHMANN_JSON_NAMESPACE \
|
|
||||||
nlohmann::NLOHMANN_JSON_NAMESPACE_CONCAT( \
|
|
||||||
NLOHMANN_JSON_ABI_TAGS, \
|
|
||||||
NLOHMANN_JSON_NAMESPACE_VERSION)
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#ifndef NLOHMANN_JSON_NAMESPACE_BEGIN
|
|
||||||
#define NLOHMANN_JSON_NAMESPACE_BEGIN \
|
|
||||||
namespace nlohmann \
|
|
||||||
{ \
|
|
||||||
inline namespace NLOHMANN_JSON_NAMESPACE_CONCAT( \
|
|
||||||
NLOHMANN_JSON_ABI_TAGS, \
|
|
||||||
NLOHMANN_JSON_NAMESPACE_VERSION) \
|
|
||||||
{
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#ifndef NLOHMANN_JSON_NAMESPACE_END
|
|
||||||
#define NLOHMANN_JSON_NAMESPACE_END \
|
|
||||||
} /* namespace (inline namespace) NOLINT(readability/namespace) */ \
|
|
||||||
} // namespace nlohmann
|
|
||||||
#endif
|
|
||||||
|
|
||||||
|
|
||||||
/*!
|
|
||||||
@brief namespace for Niels Lohmann
|
|
||||||
@see https://github.com/nlohmann
|
|
||||||
@since version 1.0.0
|
|
||||||
*/
|
|
||||||
NLOHMANN_JSON_NAMESPACE_BEGIN
|
|
||||||
|
|
||||||
/*!
|
|
||||||
@brief default JSONSerializer template argument
|
|
||||||
|
|
||||||
This serializer ignores the template arguments and uses ADL
|
|
||||||
([argument-dependent lookup](https://en.cppreference.com/w/cpp/language/adl))
|
|
||||||
for serialization.
|
|
||||||
*/
|
|
||||||
template<typename T = void, typename SFINAE = void>
|
|
||||||
struct adl_serializer;
|
|
||||||
|
|
||||||
/// a class to store JSON values
|
|
||||||
/// @sa https://json.nlohmann.me/api/basic_json/
|
|
||||||
template<template<typename U, typename V, typename... Args> class ObjectType =
|
|
||||||
std::map,
|
|
||||||
template<typename U, typename... Args> class ArrayType = std::vector,
|
|
||||||
class StringType = std::string, class BooleanType = bool,
|
|
||||||
class NumberIntegerType = std::int64_t,
|
|
||||||
class NumberUnsignedType = std::uint64_t,
|
|
||||||
class NumberFloatType = double,
|
|
||||||
template<typename U> class AllocatorType = std::allocator,
|
|
||||||
template<typename T, typename SFINAE = void> class JSONSerializer =
|
|
||||||
adl_serializer,
|
|
||||||
class BinaryType = std::vector<std::uint8_t>, // cppcheck-suppress syntaxError
|
|
||||||
class CustomBaseClass = void>
|
|
||||||
class basic_json;
|
|
||||||
|
|
||||||
/// @brief JSON Pointer defines a string syntax for identifying a specific value within a JSON document
|
|
||||||
/// @sa https://json.nlohmann.me/api/json_pointer/
|
|
||||||
template<typename RefStringType>
|
|
||||||
class json_pointer;
|
|
||||||
|
|
||||||
/*!
|
|
||||||
@brief default specialization
|
|
||||||
@sa https://json.nlohmann.me/api/json/
|
|
||||||
*/
|
|
||||||
using json = basic_json<>;
|
|
||||||
|
|
||||||
/// @brief a minimal map-like container that preserves insertion order
|
|
||||||
/// @sa https://json.nlohmann.me/api/ordered_map/
|
|
||||||
template<class Key, class T, class IgnoredLess, class Allocator>
|
|
||||||
struct ordered_map;
|
|
||||||
|
|
||||||
/// @brief specialization that maintains the insertion order of object keys
|
|
||||||
/// @sa https://json.nlohmann.me/api/ordered_json/
|
|
||||||
using ordered_json = basic_json<nlohmann::ordered_map>;
|
|
||||||
|
|
||||||
NLOHMANN_JSON_NAMESPACE_END
|
|
||||||
|
|
||||||
#endif // INCLUDE_NLOHMANN_JSON_FWD_HPP_
|
|
@ -1,4 +1,8 @@
|
|||||||
#include "utils.hpp"
|
#include "utils.hpp"
|
||||||
|
#include "httplib.hpp"
|
||||||
|
#include <nlohmann/json.hpp>
|
||||||
|
|
||||||
|
|
||||||
#include <iostream>
|
#include <iostream>
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <fstream>
|
#include <fstream>
|
||||||
@ -10,6 +14,7 @@
|
|||||||
#include <sys/ioctl.h>
|
#include <sys/ioctl.h>
|
||||||
#include <unistd.h>
|
#include <unistd.h>
|
||||||
#include <cctype>
|
#include <cctype>
|
||||||
|
#include <sstream>
|
||||||
|
|
||||||
namespace dropshell {
|
namespace dropshell {
|
||||||
|
|
||||||
@ -35,7 +40,7 @@ bool replace_line_in_file(const std::string& file_path, const std::string& searc
|
|||||||
std::string line;
|
std::string line;
|
||||||
|
|
||||||
if (!input_file.is_open()) {
|
if (!input_file.is_open()) {
|
||||||
std::cerr << "Error: Unable to open file: " << file_path << std::endl;
|
error << "Unable to open file: " << file_path << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -50,7 +55,7 @@ bool replace_line_in_file(const std::string& file_path, const std::string& searc
|
|||||||
std::ofstream output_file(file_path);
|
std::ofstream output_file(file_path);
|
||||||
if (!output_file.is_open())
|
if (!output_file.is_open())
|
||||||
{
|
{
|
||||||
std::cerr << "Error: Unable to open file: " << file_path << std::endl;
|
error << "Unable to open file: " << file_path << std::endl;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
for (const auto& modified_line : file_lines)
|
for (const auto& modified_line : file_lines)
|
||||||
@ -151,7 +156,7 @@ int str2int(const std::string &str)
|
|||||||
try {
|
try {
|
||||||
return std::stoi(str);
|
return std::stoi(str);
|
||||||
} catch (const std::exception& e) {
|
} catch (const std::exception& e) {
|
||||||
std::cerr << "Error: Invalid integer string: [" << str << "]" << std::endl;
|
error << "Invalid integer string: [" << str << "]" << std::endl;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -307,8 +312,9 @@ std::string requote(std::string str) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
int die(const std::string & msg) {
|
int return_die(const std::string & msg) {
|
||||||
std::cerr << msg << std::endl;
|
error << "Fatal error:" << std::endl;
|
||||||
|
error << msg << std::endl;
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -445,4 +451,223 @@ std::string tolower(const std::string& str) {
|
|||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Common utility function to make HTTP requests
|
||||||
|
struct HttpResult {
|
||||||
|
bool success;
|
||||||
|
int status;
|
||||||
|
std::string body;
|
||||||
|
std::string error;
|
||||||
|
};
|
||||||
|
|
||||||
|
HttpResult make_http_request(const std::string& url) {
|
||||||
|
try {
|
||||||
|
// Parse the URL to get host and path
|
||||||
|
std::string host;
|
||||||
|
std::string path;
|
||||||
|
size_t protocol_end = url.find("://");
|
||||||
|
if (protocol_end != std::string::npos) {
|
||||||
|
size_t host_start = protocol_end + 3;
|
||||||
|
size_t path_start = url.find('/', host_start);
|
||||||
|
if (path_start != std::string::npos) {
|
||||||
|
host = url.substr(host_start, path_start - host_start);
|
||||||
|
path = url.substr(path_start);
|
||||||
|
} else {
|
||||||
|
host = url.substr(host_start);
|
||||||
|
path = "/";
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
return {false, 0, "", "Invalid URL format"};
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create HTTP client
|
||||||
|
httplib::Client cli(host);
|
||||||
|
cli.set_connection_timeout(10); // 10 second timeout
|
||||||
|
|
||||||
|
// Make GET request
|
||||||
|
auto res = cli.Get(path);
|
||||||
|
if (!res) {
|
||||||
|
return {false, 0, "", "Failed to connect to server"};
|
||||||
|
}
|
||||||
|
if (res->status != 200) {
|
||||||
|
return {false, res->status, "", "HTTP request failed with status " + std::to_string(res->status)};
|
||||||
|
}
|
||||||
|
|
||||||
|
return {true, res->status, res->body, ""};
|
||||||
|
} catch (const std::exception& e) {
|
||||||
|
return {false, 0, "", std::string("Exception: ") + e.what()};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
bool download_file(const std::string &url, const std::string &destination) {
|
||||||
|
auto result = make_http_request(url);
|
||||||
|
if (!result.success) {
|
||||||
|
warning << "Failed to download file from URL: " << url << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
std::ofstream out_file(destination, std::ios::binary);
|
||||||
|
if (!out_file) {
|
||||||
|
warning << "Failed to open file for writing: " << destination << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
out_file.write(result.body.c_str(), result.body.size());
|
||||||
|
out_file.close();
|
||||||
|
return true;
|
||||||
|
} catch (const std::exception& e) {
|
||||||
|
warning << "Failed to download file from URL: " << url << std::endl;
|
||||||
|
warning << "Exception: " << e.what() << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
nlohmann::json get_json_from_url(const std::string &url) {
|
||||||
|
auto result = make_http_request(url);
|
||||||
|
if (!result.success) {
|
||||||
|
warning << "Failed to get JSON from URL: " << url << std::endl;
|
||||||
|
return nlohmann::json();
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
return nlohmann::json::parse(result.body);
|
||||||
|
} catch (const nlohmann::json::parse_error& e) {
|
||||||
|
warning << "Failed to parse JSON from URL: " << url << std::endl;
|
||||||
|
warning << "JSON: " << result.body << std::endl;
|
||||||
|
return nlohmann::json();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
std::string get_string_from_url(const std::string &url) {
|
||||||
|
auto result = make_http_request(url);
|
||||||
|
if (!result.success) {
|
||||||
|
warning << "Failed to get string from URL: " << url << std::endl;
|
||||||
|
return std::string();
|
||||||
|
}
|
||||||
|
return result.body;
|
||||||
|
}
|
||||||
|
|
||||||
|
bool match_line(const std::string &line, const std::string &pattern) {
|
||||||
|
return trim(line) == trim(pattern);
|
||||||
|
}
|
||||||
|
|
||||||
|
// replace or append a block of text to a file, matching first and last lines if replacing.
|
||||||
|
// edits file in-place.
|
||||||
|
bool file_replace_or_add_segment(std::string filepath, std::string segment)
|
||||||
|
{
|
||||||
|
// Create a backup of the original file
|
||||||
|
std::string backup_path = filepath + ".bak";
|
||||||
|
try {
|
||||||
|
std::filesystem::copy_file(filepath, backup_path, std::filesystem::copy_options::overwrite_existing);
|
||||||
|
} catch (const std::exception& e) {
|
||||||
|
std::cerr << "Error creating backup file: " << e.what() << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Handle empty segment
|
||||||
|
if (segment.empty()) {
|
||||||
|
error << "Empty segment provided" << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
// split the segment into lines
|
||||||
|
std::vector<std::string> segment_lines = split(segment, "\n");
|
||||||
|
// remove empty lines
|
||||||
|
segment_lines.erase(std::remove_if(segment_lines.begin(), segment_lines.end(), [](const std::string& line) {
|
||||||
|
return trim(line).empty();
|
||||||
|
}), segment_lines.end());
|
||||||
|
// remove any lines that are just whitespace
|
||||||
|
segment_lines.erase(std::remove_if(segment_lines.begin(), segment_lines.end(), [](const std::string& line) { return trim(line).empty(); }), segment_lines.end());
|
||||||
|
|
||||||
|
// check that the segment has at least two lines
|
||||||
|
if (segment_lines.size() < 2) {
|
||||||
|
error << "Segment must contain at least two non-empty lines" << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Read the entire file into memory
|
||||||
|
std::ifstream input_file(filepath);
|
||||||
|
if (!input_file.is_open()) {
|
||||||
|
error << "Unable to open file: " << filepath << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
std::vector<std::string> file_lines;
|
||||||
|
std::string line;
|
||||||
|
while (std::getline(input_file, line)) {
|
||||||
|
file_lines.push_back(line);
|
||||||
|
}
|
||||||
|
input_file.close();
|
||||||
|
|
||||||
|
// Store original file size for verification
|
||||||
|
size_t original_size = file_lines.size();
|
||||||
|
if (original_size == 0) {
|
||||||
|
warning << "File is empty" << std::endl;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Try to find the matching block
|
||||||
|
bool found_match = false;
|
||||||
|
for (size_t i = 0; i < file_lines.size(); i++) {
|
||||||
|
if (match_line(file_lines[i], segment_lines[0])) {
|
||||||
|
// Found potential start, look for end
|
||||||
|
for (size_t j = i + 1; j < file_lines.size(); j++) {
|
||||||
|
if (match_line(file_lines[j], segment_lines[segment_lines.size() - 1])) {
|
||||||
|
// Found matching block, replace it
|
||||||
|
file_lines.erase(file_lines.begin() + i, file_lines.begin() + j + 1);
|
||||||
|
file_lines.insert(file_lines.begin() + i, segment_lines.begin(), segment_lines.end());
|
||||||
|
|
||||||
|
found_match = true;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (found_match) break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// If no match found, append the segment
|
||||||
|
if (!found_match) {
|
||||||
|
file_lines.insert(file_lines.end(), segment_lines.begin(), segment_lines.end());
|
||||||
|
}
|
||||||
|
|
||||||
|
// Write back to file
|
||||||
|
std::ofstream output_file(filepath);
|
||||||
|
if (!output_file.is_open()) {
|
||||||
|
error << "Unable to open file for writing: " << filepath << std::endl;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
for (const auto& line : file_lines) {
|
||||||
|
output_file << line << "\n";
|
||||||
|
}
|
||||||
|
output_file.close();
|
||||||
|
|
||||||
|
// If everything succeeded, remove the backup
|
||||||
|
try {
|
||||||
|
std::filesystem::remove(backup_path);
|
||||||
|
} catch (const std::exception& e) {
|
||||||
|
warning << "Could not remove backup file: " << e.what() << std::endl;
|
||||||
|
}
|
||||||
|
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
bool legal_service_name(const std::string &service_name) {
|
||||||
|
static bool initialized = false;
|
||||||
|
static bool legal_chars[256] = {false}; // Initialize all to false
|
||||||
|
|
||||||
|
// One-time initialization
|
||||||
|
if (!initialized) {
|
||||||
|
// Set true for valid characters
|
||||||
|
for (unsigned char c : "0123456789"
|
||||||
|
"ABCDEFGHIJKLMNOPQRSTUVWXYZ"
|
||||||
|
"abcdefghijklmnopqrstuvwxyz"
|
||||||
|
"._-") {
|
||||||
|
legal_chars[c] = true;
|
||||||
|
}
|
||||||
|
initialized = true;
|
||||||
|
}
|
||||||
|
|
||||||
|
return std::all_of(service_name.begin(), service_name.end(),
|
||||||
|
[](unsigned char c) { return legal_chars[c]; });
|
||||||
|
}
|
||||||
|
|
||||||
} // namespace dropshell
|
} // namespace dropshell
|
@ -5,6 +5,7 @@
|
|||||||
#include <map>
|
#include <map>
|
||||||
|
|
||||||
#include "output.hpp"
|
#include "output.hpp"
|
||||||
|
#include <nlohmann/json.hpp>
|
||||||
|
|
||||||
namespace dropshell {
|
namespace dropshell {
|
||||||
|
|
||||||
@ -44,7 +45,7 @@ int count_substring(const std::string &substring, const std::string &text);
|
|||||||
|
|
||||||
std::string random_alphanumeric_string(int length);
|
std::string random_alphanumeric_string(int length);
|
||||||
|
|
||||||
int die(const std::string & msg);
|
int return_die(const std::string & msg);
|
||||||
std::string safearg(int argc, char *argv[], int index);
|
std::string safearg(int argc, char *argv[], int index);
|
||||||
std::string safearg(const std::vector<std::string> & args, int index);
|
std::string safearg(const std::vector<std::string> & args, int index);
|
||||||
|
|
||||||
@ -61,4 +62,20 @@ std::string get_line_wrap(std::string & src, int maxchars);
|
|||||||
|
|
||||||
std::string tolower(const std::string& str);
|
std::string tolower(const std::string& str);
|
||||||
|
|
||||||
|
bool download_file(const std::string& url, const std::string& destination);
|
||||||
|
nlohmann::json get_json_from_url(const std::string& url);
|
||||||
|
std::string get_string_from_url(const std::string& url);
|
||||||
|
|
||||||
|
// replace or append a block of text to a file, matching first and last lines if replacing.
|
||||||
|
bool file_replace_or_add_segment(std::string filepath, std::string segment);
|
||||||
|
|
||||||
|
|
||||||
|
constexpr unsigned int switchhash(const char *s, int off = 0)
|
||||||
|
{
|
||||||
|
return !s[off] ? 5381 : (switchhash(s, off + 1) * 33) ^ s[off];
|
||||||
|
}
|
||||||
|
|
||||||
|
bool legal_service_name(const std::string & service_name);
|
||||||
|
|
||||||
|
|
||||||
} // namespace dropshell
|
} // namespace dropshell
|
12
source/test.sh
Executable file
12
source/test.sh
Executable file
@ -0,0 +1,12 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -euo pipefail
|
||||||
|
SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" &> /dev/null && pwd )"
|
||||||
|
|
||||||
|
ARCH=$(uname -m)
|
||||||
|
|
||||||
|
PREV_DIR=$(pwd)
|
||||||
|
trap 'cd "$PREV_DIR"' EXIT
|
||||||
|
|
||||||
|
"$SCRIPT_DIR/output/dropshell.${ARCH}" hash "${SCRIPT_DIR}/test.sh"
|
||||||
|
"$SCRIPT_DIR/output/dropshell.${ARCH}" help
|
||||||
|
|
Loading…
x
Reference in New Issue
Block a user