Compare commits
114 Commits
| Author | SHA1 | Date |
|---|---|---|
|
|
fae4a113d6 | |
|
|
af77b89a68 | |
|
|
ee64e39c50 | |
|
|
02f46839bd | |
|
|
94ed7af436 | |
|
|
4eb603afd7 | |
|
|
d4e9a2f875 | |
|
|
4133d2f348 | |
|
|
0634ecd0d3 | |
|
|
ef8d7e9b4d | |
|
|
5fabed1e56 | |
|
|
196280f783 | |
|
|
da276b034f | |
|
|
d7f39d8a99 | |
|
|
262b739c0c | |
|
|
2d8f270c54 | |
|
|
99b1d14dba | |
|
|
c59ac23997 | |
|
|
1c1252ecfe | |
|
|
888fb41885 | |
|
|
8ba911385f | |
|
|
a9c0aa14fa | |
|
|
56f8474c01 | |
|
|
c76b615341 | |
|
|
e9f9bbc0b4 | |
|
|
dafc6dcbeb | |
|
|
5a5b7dc315 | |
|
|
9133ce5de7 | |
|
|
0d9020d3bf | |
|
|
5f63459e4f | |
|
|
e0720a5f54 | |
|
|
1c651c93ce | |
|
|
b456c7aa18 | |
|
|
3d0b7766ac | |
|
|
19b3477560 | |
|
|
08e99385b9 | |
|
|
81df3f8435 | |
|
|
75940cb187 | |
|
|
ac0c3cb8c1 | |
|
|
3147c73006 | |
|
|
bd5e21fce7 | |
|
|
54d2b1aaf7 | |
|
|
3f98fd7f24 | |
|
|
a4afd430af | |
|
|
5fe95bfcd9 | |
|
|
df3934452b | |
|
|
5c94e366eb | |
|
|
92aa6f5b66 | |
|
|
afeece915c | |
|
|
848b321228 | |
|
|
895dc34557 | |
|
|
d6d45fe4aa | |
|
|
6ff17d9620 | |
|
|
4b9db02528 | |
|
|
551223dd91 | |
|
|
62af29b9e0 | |
|
|
b4c42dfc00 | |
|
|
1f4281e877 | |
|
|
38c9268db3 | |
|
|
6e4a889a10 | |
|
|
fac74663fa | |
|
|
71e9e8c032 | |
|
|
b77241fc4c | |
|
|
2559570702 | |
|
|
0c44491955 | |
|
|
529247a839 | |
|
|
577c62ee9b | |
|
|
cc8d949dc4 | |
|
|
e8317600df | |
|
|
1c3ac320b8 | |
|
|
aae0391a0c | |
|
|
3dd238cf97 | |
|
|
403285a937 | |
|
|
a16eb78b79 | |
|
|
281841b68a | |
|
|
34979a035d | |
|
|
09c1baed8e | |
|
|
3c22a67052 | |
|
|
052448a7b9 | |
|
|
6d56d1e39c | |
|
|
541b0f52dd | |
|
|
2495fb84cf | |
|
|
c3fd0dd09f | |
|
|
502ea114a6 | |
|
|
28092d945a | |
|
|
0d68efd461 | |
|
|
4fc90300fc | |
|
|
e3f07f42a6 | |
|
|
cd7669d942 | |
|
|
721fa6c758 | |
|
|
c50c444f21 | |
|
|
71acb4a32e | |
|
|
f504632c4d | |
|
|
2b82fb7aac | |
|
|
584404c050 | |
|
|
886ae6308b | |
|
|
8f1214bd9a | |
|
|
011c479550 | |
|
|
27e79ce731 | |
|
|
026a502044 | |
|
|
163887d42c | |
|
|
35a21da431 | |
|
|
064611823a | |
|
|
e4c3e5f46f | |
|
|
6adab1b903 | |
|
|
98da769dd3 | |
|
|
a8a7fd8a72 | |
|
|
3d88967281 | |
|
|
8c1998c93f | |
|
|
f560dfebc5 | |
|
|
0f160f4dcd | |
|
|
7a5704dd93 | |
|
|
df03bd5346 | |
|
|
e9b6abefdf |
|
|
@ -0,0 +1,2 @@
|
||||||
|
target/
|
||||||
|
Cargo.lock
|
||||||
|
|
@ -0,0 +1,40 @@
|
||||||
|
name: Rust riscv64 build
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches: [ master , rc ]
|
||||||
|
pull_request:
|
||||||
|
branches: [ master , rc ]
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
test:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Checkout code
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Install dependencies
|
||||||
|
run: sudo apt-get update && sudo apt-get install -y libssl-dev pkg-config build-essential gcc-riscv64-unknown-elf gcc-riscv64-linux-gnu binutils-riscv64-linux-gnu
|
||||||
|
|
||||||
|
- name: Set up Rust
|
||||||
|
uses: dtolnay/rust-toolchain@stable
|
||||||
|
with:
|
||||||
|
toolchain: stable
|
||||||
|
|
||||||
|
- name: Cache dependencies
|
||||||
|
uses: actions/cache@v4
|
||||||
|
with:
|
||||||
|
path: |
|
||||||
|
~/.cargo/registry
|
||||||
|
~/.cargo/git
|
||||||
|
target
|
||||||
|
key: ${{ runner.os }}-cargo-${{ hashFiles('**/Cargo.lock') }}
|
||||||
|
restore-keys: |
|
||||||
|
${{ runner.os }}-cargo-
|
||||||
|
|
||||||
|
- name: Set up riscv64gc toolchain
|
||||||
|
run: rustup target add riscv64gc-unknown-linux-gnu
|
||||||
|
|
||||||
|
- name: Build x86_64-unknown-linux-gnu
|
||||||
|
run: cargo riscv64 --verbose
|
||||||
|
|
@ -0,0 +1,37 @@
|
||||||
|
name: Rust test
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches: [ master , rc ]
|
||||||
|
pull_request:
|
||||||
|
branches: [ master , rc ]
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
test:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Checkout code
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Install dependencies
|
||||||
|
run: sudo apt-get update && sudo apt-get install -y libssl-dev pkg-config build-essential
|
||||||
|
|
||||||
|
- name: Set up Rust
|
||||||
|
uses: dtolnay/rust-toolchain@stable
|
||||||
|
with:
|
||||||
|
toolchain: stable
|
||||||
|
|
||||||
|
- name: Cache dependencies
|
||||||
|
uses: actions/cache@v4
|
||||||
|
with:
|
||||||
|
path: |
|
||||||
|
~/.cargo/registry
|
||||||
|
~/.cargo/git
|
||||||
|
target
|
||||||
|
key: ${{ runner.os }}-cargo-${{ hashFiles('**/Cargo.lock') }}
|
||||||
|
restore-keys: |
|
||||||
|
${{ runner.os }}-cargo-
|
||||||
|
|
||||||
|
- name: Run tests
|
||||||
|
run: cargo test --verbose
|
||||||
|
|
@ -0,0 +1,37 @@
|
||||||
|
name: Rust x86_64 build
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches: [ master , rc ]
|
||||||
|
pull_request:
|
||||||
|
branches: [ master , rc ]
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
test:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Checkout code
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Install dependencies
|
||||||
|
run: sudo apt-get update && sudo apt-get install -y libssl-dev pkg-config build-essential
|
||||||
|
|
||||||
|
- name: Set up Rust
|
||||||
|
uses: dtolnay/rust-toolchain@stable
|
||||||
|
with:
|
||||||
|
toolchain: stable
|
||||||
|
|
||||||
|
- name: Cache dependencies
|
||||||
|
uses: actions/cache@v4
|
||||||
|
with:
|
||||||
|
path: |
|
||||||
|
~/.cargo/registry
|
||||||
|
~/.cargo/git
|
||||||
|
target
|
||||||
|
key: ${{ runner.os }}-cargo-${{ hashFiles('**/Cargo.lock') }}
|
||||||
|
restore-keys: |
|
||||||
|
${{ runner.os }}-cargo-
|
||||||
|
|
||||||
|
- name: Build x86_64-unknown-linux-gnu
|
||||||
|
run: cargo x86_64 --verbose
|
||||||
|
|
@ -1,6 +1,9 @@
|
||||||
/target
|
/target
|
||||||
.idea
|
.idea
|
||||||
/.env
|
.env
|
||||||
Cargo.lock
|
Cargo.lock
|
||||||
hagent_test.sock
|
hagent_test.sock
|
||||||
release
|
release
|
||||||
|
*.sock
|
||||||
|
*.bak
|
||||||
|
docker-compose.yml
|
||||||
|
|
@ -2,7 +2,7 @@
|
||||||
resolver = "2"
|
resolver = "2"
|
||||||
members = [
|
members = [
|
||||||
"noxis-rs",
|
"noxis-rs",
|
||||||
"noxis-cli",
|
"noxis-cli", "noxis-proxy",
|
||||||
]
|
]
|
||||||
|
|
||||||
[profile.dev]
|
[profile.dev]
|
||||||
|
|
|
||||||
|
|
@ -1,31 +1,13 @@
|
||||||
FROM ubuntu:22.04
|
FROM registry.entcor/trust-module/rust:1.88.0-extended
|
||||||
|
|
||||||
USER root
|
USER root
|
||||||
|
|
||||||
RUN apt update && apt install -y \
|
|
||||||
curl \
|
|
||||||
build-essential \
|
|
||||||
libssl-dev \
|
|
||||||
pkg-config \
|
|
||||||
libudev-dev \
|
|
||||||
procps \
|
|
||||||
gcc-riscv64-unknown-elf \
|
|
||||||
gcc-riscv64-linux-gnu \
|
|
||||||
binutils-riscv64-linux-gnu \
|
|
||||||
&& rm -rf /var/lib/apt/lists/*
|
|
||||||
|
|
||||||
RUN curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh -s -- -y
|
|
||||||
|
|
||||||
ENV PATH="/root/.cargo/bin:${PATH}"
|
|
||||||
|
|
||||||
WORKDIR /usr/src/kii/
|
WORKDIR /usr/src/kii/
|
||||||
|
|
||||||
COPY . ./
|
COPY . ./
|
||||||
|
|
||||||
RUN chmod +x noxis-rs/temp-process
|
RUN chmod +x noxis-rs/temp-process
|
||||||
|
RUN rustup target add riscv64gc-unknown-linux-gnu
|
||||||
RUN rustup target add riscv64gc-unknown-linux-gnu && rustup target add x86_64-unknown-linux-gnu
|
|
||||||
|
|
||||||
RUN cargo unibuild
|
RUN cargo unibuild
|
||||||
|
|
||||||
ENTRYPOINT ["cargo", "test"]
|
ENTRYPOINT ["cargo", "test"]
|
||||||
|
|
|
||||||
|
|
@ -1,10 +1,20 @@
|
||||||
|
|
||||||
# noxis-rs
|
# noxis
|
||||||
|
|
||||||

|

|
||||||
### In-container integrating util to handle processes runtime
|
|
||||||
( with amd64 and riscv64 support )
|
|
||||||
|
|
||||||
## Depends on
|
`noxis` - monitoring util with special attention on
|
||||||
|
1) **Speed**
|
||||||
|
2) **Multiplatform** execution *(with `amd64` and `riscv64` **support**)*
|
||||||
|
3) **Smallness** and **Optimization**
|
||||||
|
|
||||||
|
It's **main tasks** are
|
||||||
|
- to manage the processes that occur inside the container or in the target system.
|
||||||
|
- collect data (metrics);
|
||||||
|
- monitor the availability of system files necessary for the operation of processes;
|
||||||
|
- check whether there is a connection between processes and services, where the information comes from or where it is sent.
|
||||||
|
|
||||||
|
## Build requirements
|
||||||
- `rustup (>=1.27.1)`
|
- `rustup (>=1.27.1)`
|
||||||
- `gcc-riscv64-unknown-elf`
|
- `gcc-riscv64-unknown-elf`
|
||||||
- `build-essential`
|
- `build-essential`
|
||||||
|
|
@ -12,60 +22,70 @@
|
||||||
- `binutils-riscv64-linux-gnu`
|
- `binutils-riscv64-linux-gnu`
|
||||||
|
|
||||||
|
|
||||||
## Setting up
|
## Key items in repo
|
||||||
Download and execute rustup.sh
|
|
||||||
~~~bash
|
1) Main daemon `noxis-rs`
|
||||||
|
2) CLI `noxis-cli`
|
||||||
|
3) Unix-Socket to Web-Socket **Proxy** for integrations `noxis-proxy`
|
||||||
|
|
||||||
|
## Setting up device
|
||||||
|
|
||||||
|
Download and execute rustup.sh *(for building)*
|
||||||
|
~~~ bash
|
||||||
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh
|
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh
|
||||||
~~~
|
~~~
|
||||||
|
|
||||||
## Building
|
## Building
|
||||||
|
|
||||||
1. Clone this repo `runner-rs`
|
1. Clone `noxis`
|
||||||
|
|
||||||
~~~bash
|
~~~ bash
|
||||||
git clone https://github.com/prplV/runner-rs
|
git clone https://github.com/prplV/noxis
|
||||||
~~~
|
~~~
|
||||||
|
|
||||||
2. Enter project's dir and set up toolchain list to compile code for RISC-V and AMD64
|
2. Enter project's dir and set up toolchain list to compile code for RISC-V or x86_64
|
||||||
|
|
||||||
~~~bash
|
~~~ bash
|
||||||
cd runner-rs/ && rustup target add riscv64gc-unknown-linux-gnu && rustup target add x86_64-unknown-linux-gnu
|
cd noxis/ && rustup target add riscv64gc-unknown-linux-gnu && rustup target add x86_64-unknown-linux-gnu
|
||||||
~~~
|
~~~
|
||||||
> [!NOTE]
|
> [!NOTE]
|
||||||
> Cargo is configured to build an app for amd64/linux defaultly. RISCV-based compilation is optional.
|
> Cargo is configured to build an apputil for x86_64/linux defaultly. RISCV-based compilation is optional.
|
||||||
|
|
||||||
3.1. Release build of app for amd64/linux
|
3.1. Release build of util for x86_64/linux
|
||||||
|
|
||||||
~~~bash
|
~~~bash
|
||||||
cargo x86_64
|
cargo x86_64
|
||||||
~~~
|
~~~
|
||||||
|
|
||||||
3.2. Release build of app for riscv64/linux
|
3.2. Release build of util for riscv64/linux
|
||||||
|
|
||||||
~~~bash
|
~~~bash
|
||||||
cargo riscv64
|
cargo riscv64
|
||||||
~~~
|
~~~
|
||||||
|
|
||||||
3.3. Release build of app for both (riscv64 and amd64)
|
3.3. Release build of util for both (riscv64 and x86_64)
|
||||||
|
|
||||||
~~~bash
|
~~~bash
|
||||||
cargo unibuild
|
cargo unibuild
|
||||||
~~~
|
~~~
|
||||||
|
|
||||||
|
|
||||||
## Execution for amd64/linux
|
## Execution **DAEMON** for x86_64/linux
|
||||||
|
|
||||||
|
1) If you work on x86_64/linux machine execute:
|
||||||
|
|
||||||
~~~bash
|
~~~bash
|
||||||
./target/x86_64-unknown-linux-gnu/release/runner-rs
|
./target/x86_64-unknown-linux-gnu/release/noxis-rs
|
||||||
~~~
|
~~~
|
||||||
or
|
or
|
||||||
~~~bash
|
~~~bash
|
||||||
cargo run_x86
|
cargo run_x86
|
||||||
~~~
|
~~~
|
||||||
|
|
||||||
|
2) If you work on riscv64/linux machine execute:
|
||||||
|
|
||||||
## Execution for riscv64/linux
|
|
||||||
~~~bash
|
~~~bash
|
||||||
./target/riscv64gc-unknown-linux-gnu/release/runner-rs
|
./target/riscv64gc-unknown-linux-gnu/release/noxis-rs
|
||||||
~~~
|
~~~
|
||||||
or
|
or
|
||||||
> [!CAUTION]
|
> [!CAUTION]
|
||||||
|
|
|
||||||
50
build.sh
50
build.sh
|
|
@ -5,7 +5,7 @@
|
||||||
# Поддерживаемые архитектуры: amd64, riscv64
|
# Поддерживаемые архитектуры: amd64, riscv64
|
||||||
|
|
||||||
if [ -z "$1" ]; then
|
if [ -z "$1" ]; then
|
||||||
echo "Ошибка: Необходимо указать архитектуру (например, amd64 или riscv64)."
|
echo "Ошибка: Необходимо указать архитектуру (например, amd64)."
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
|
@ -13,7 +13,7 @@ ARCH="$1"
|
||||||
TARGET_DIR="release/${ARCH}"
|
TARGET_DIR="release/${ARCH}"
|
||||||
CONTAINER_NAME="e-monitor"
|
CONTAINER_NAME="e-monitor"
|
||||||
|
|
||||||
SUPPORTED_ARCHS=("amd64" "riscv64")
|
SUPPORTED_ARCHS=("amd64" "riscv64" "supported")
|
||||||
if [[ ! " ${SUPPORTED_ARCHS[@]} " =~ " ${ARCH} " ]]; then
|
if [[ ! " ${SUPPORTED_ARCHS[@]} " =~ " ${ARCH} " ]]; then
|
||||||
echo "Ошибка: Неизвестная архитектура $ARCH. Допустимые значения: ${SUPPORTED_ARCHS[*]}."
|
echo "Ошибка: Неизвестная архитектура $ARCH. Допустимые значения: ${SUPPORTED_ARCHS[*]}."
|
||||||
exit 1
|
exit 1
|
||||||
|
|
@ -52,6 +52,11 @@ case "$ARCH" in
|
||||||
docker stop "$CONTAINER_NAME" && docker rm "$CONTAINER_NAME"
|
docker stop "$CONTAINER_NAME" && docker rm "$CONTAINER_NAME"
|
||||||
exit 1
|
exit 1
|
||||||
}
|
}
|
||||||
|
docker cp "$CONTAINER_NAME:/usr/src/kii/target/x86_64-unknown-linux-gnu/release/noxis-proxy" "$TARGET_DIR/" || {
|
||||||
|
echo "Ошибка: Не удалось скопировать noxis-proxy для amd64."
|
||||||
|
docker stop "$CONTAINER_NAME" && docker rm "$CONTAINER_NAME"
|
||||||
|
exit 1
|
||||||
|
}
|
||||||
;;
|
;;
|
||||||
riscv64)
|
riscv64)
|
||||||
echo "Copying binaries for architecture: riscv64"
|
echo "Copying binaries for architecture: riscv64"
|
||||||
|
|
@ -65,6 +70,47 @@ case "$ARCH" in
|
||||||
docker stop "$CONTAINER_NAME" && docker rm "$CONTAINER_NAME"
|
docker stop "$CONTAINER_NAME" && docker rm "$CONTAINER_NAME"
|
||||||
exit 1
|
exit 1
|
||||||
}
|
}
|
||||||
|
docker cp "$CONTAINER_NAME:/usr/src/kii/target/riscv64gc-unknown-linux-gnu/release/noxis-proxy" "$TARGET_DIR/" || {
|
||||||
|
echo "Ошибка: Не удалось скопировать noxis-proxy для riscv64."
|
||||||
|
docker stop "$CONTAINER_NAME" && docker rm "$CONTAINER_NAME"
|
||||||
|
exit 1
|
||||||
|
}
|
||||||
|
;;
|
||||||
|
supported)
|
||||||
|
TARGET_DIR="release/supported/amd64"
|
||||||
|
echo "Copying binaries for architecture: amd64"
|
||||||
|
docker cp "$CONTAINER_NAME:/usr/src/kii/target/x86_64-unknown-linux-gnu/release/noxis-cli" "$TARGET_DIR/amd64/" || {
|
||||||
|
echo "Ошибка: Не удалось скопировать noxis-cli для amd64."
|
||||||
|
docker stop "$CONTAINER_NAME" && docker rm "$CONTAINER_NAME"
|
||||||
|
exit 1
|
||||||
|
}
|
||||||
|
docker cp "$CONTAINER_NAME:/usr/src/kii/target/x86_64-unknown-linux-gnu/release/noxis-rs" "$TARGET_DIR/amd64/" || {
|
||||||
|
echo "Ошибка: Не удалось скопировать noxis-rs для amd64."
|
||||||
|
docker stop "$CONTAINER_NAME" && docker rm "$CONTAINER_NAME"
|
||||||
|
exit 1
|
||||||
|
}
|
||||||
|
docker cp "$CONTAINER_NAME:/usr/src/kii/target/x86_64-unknown-linux-gnu/release/noxis-proxy" "$TARGET_DIR/amd64/" || {
|
||||||
|
echo "Ошибка: Не удалось скопировать noxis-proxy для amd64."
|
||||||
|
docker stop "$CONTAINER_NAME" && docker rm "$CONTAINER_NAME"
|
||||||
|
exit 1
|
||||||
|
}
|
||||||
|
TARGET_DIR="release/supported/riscv64"
|
||||||
|
echo "Copying binaries for architecture: riscv64"
|
||||||
|
docker cp "$CONTAINER_NAME:/usr/src/kii/target/riscv64gc-unknown-linux-gnu/release/noxis-cli" "$TARGET_DIR/riscv64/" || {
|
||||||
|
echo "Ошибка: Не удалось скопировать noxis-cli для riscv64."
|
||||||
|
docker stop "$CONTAINER_NAME" && docker rm "$CONTAINER_NAME"
|
||||||
|
exit 1
|
||||||
|
}
|
||||||
|
docker cp "$CONTAINER_NAME:/usr/src/kii/target/riscv64gc-unknown-linux-gnu/release/noxis-rs" "$TARGET_DIR/riscv64/" || {
|
||||||
|
echo "Ошибка: Не удалось скопировать noxis-rs для riscv64."
|
||||||
|
docker stop "$CONTAINER_NAME" && docker rm "$CONTAINER_NAME"
|
||||||
|
exit 1
|
||||||
|
}
|
||||||
|
docker cp "$CONTAINER_NAME:/usr/src/kii/target/riscv64gc-unknown-linux-gnu/release/noxis-proxy" "$TARGET_DIR/riscv64/" || {
|
||||||
|
echo "Ошибка: Не удалось скопировать noxis-proxy для riscv64."
|
||||||
|
docker stop "$CONTAINER_NAME" && docker rm "$CONTAINER_NAME"
|
||||||
|
exit 1
|
||||||
|
}
|
||||||
;;
|
;;
|
||||||
esac
|
esac
|
||||||
|
|
||||||
|
|
|
||||||
Binary file not shown.
|
Before Width: | Height: | Size: 152 KiB After Width: | Height: | Size: 221 KiB |
|
|
@ -0,0 +1 @@
|
||||||
|
NOXIS_SOCKET_PATH = "/home/vladislavd/diplom_code/noxis-rs/noxis.sock"
|
||||||
|
|
@ -1,11 +1,12 @@
|
||||||
[package]
|
[package]
|
||||||
name = "noxis-cli"
|
name = "noxis-cli"
|
||||||
version = "0.2.4"
|
version = "0.2.7"
|
||||||
edition = "2021"
|
edition = "2021"
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
anyhow = "1.0.94"
|
anyhow = "1.0.94"
|
||||||
clap = { version = "4.5.22", features = ["derive"] }
|
clap = { version = "4.5.22", features = ["derive"] }
|
||||||
|
dotenv = "0.15.0"
|
||||||
serde = { version = "1.0.215", features = ["derive"] }
|
serde = { version = "1.0.215", features = ["derive"] }
|
||||||
serde_json = "1.0.133"
|
serde_json = "1.0.133"
|
||||||
thiserror = "2.0.11"
|
thiserror = "2.0.11"
|
||||||
|
|
|
||||||
|
|
@ -1,46 +1,43 @@
|
||||||
use clap::{Parser, Subcommand};
|
use clap::{Parser, Subcommand};
|
||||||
|
use metrics_models::MetricsMode;
|
||||||
|
|
||||||
#[derive(Debug, Parser, serde::Serialize, serde::Deserialize)]
|
#[derive(Debug, Parser, serde::Serialize, serde::Deserialize)]
|
||||||
pub struct Cli {
|
pub struct Cli {
|
||||||
#[command(
|
#[arg(
|
||||||
subcommand,
|
short,
|
||||||
help = "to manage Noxis work",
|
default_value = "noxis-rs.sock",
|
||||||
|
help = "explicit specify of NOXIS Socket file"
|
||||||
)]
|
)]
|
||||||
command : Commands,
|
pub socket: String,
|
||||||
|
#[command(subcommand, help = "to manage Noxis work")]
|
||||||
|
pub command: Commands,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Subcommand, serde::Serialize, serde::Deserialize)]
|
#[derive(Debug, Subcommand, serde::Serialize, serde::Deserialize)]
|
||||||
pub enum Commands {
|
pub enum Commands {
|
||||||
#[command(
|
#[command(about = "To get info about current Noxis status")]
|
||||||
about = "To get info about current Noxis status",
|
|
||||||
)]
|
|
||||||
Status,
|
Status,
|
||||||
#[command(
|
#[command(about = "To start Noxis process")]
|
||||||
about = "To start Noxis process",
|
|
||||||
)]
|
|
||||||
Start(StartAction),
|
Start(StartAction),
|
||||||
#[command(
|
#[command(about = "To stop Noxis process")]
|
||||||
about = "To stop Noxis process",
|
|
||||||
)]
|
|
||||||
Stop,
|
Stop,
|
||||||
#[command(
|
#[command(about = "To restart Noxis process")]
|
||||||
about = "To restart Noxis process",
|
|
||||||
)]
|
|
||||||
Restart(StartAction),
|
Restart(StartAction),
|
||||||
#[command(
|
#[command(about = "To get list of processes that are being monitoring")]
|
||||||
about = "To get list of processes that are being monitoring",
|
|
||||||
)]
|
|
||||||
Processes,
|
Processes,
|
||||||
// process command
|
// process command
|
||||||
#[command(
|
#[command(about = "To manage current process that is being monitoring")]
|
||||||
about = "To manage current process that is being monitoring",
|
|
||||||
)]
|
|
||||||
Process(ProcessCommand),
|
Process(ProcessCommand),
|
||||||
// config command =
|
#[command(about = "To manage config settings")]
|
||||||
#[command(
|
|
||||||
about = "To manage config settings",
|
|
||||||
)]
|
|
||||||
Config(ConfigCommand),
|
Config(ConfigCommand),
|
||||||
|
#[command(about = "To inspect system metrics in restricted mode")]
|
||||||
|
Inspect(MetricsCommand),
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Parser, serde::Serialize, serde::Deserialize)]
|
||||||
|
pub struct MetricsCommand {
|
||||||
|
#[command(subcommand)]
|
||||||
|
pub mode: MetricsMode,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Parser, serde::Serialize, serde::Deserialize)]
|
#[derive(Debug, Parser, serde::Serialize, serde::Deserialize)]
|
||||||
|
|
@ -50,96 +47,100 @@ pub struct StartAction {
|
||||||
num_args = 1..,
|
num_args = 1..,
|
||||||
value_delimiter = ' '
|
value_delimiter = ' '
|
||||||
)]
|
)]
|
||||||
flags : Vec<String>,
|
pub flags: Vec<String>,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Parser, serde::Serialize, serde::Deserialize)]
|
#[derive(Debug, Parser, serde::Serialize, serde::Deserialize)]
|
||||||
pub struct ConfigCommand {
|
pub struct ConfigCommand {
|
||||||
#[command(subcommand)]
|
#[command(subcommand)]
|
||||||
action : ConfigAction,
|
pub action: ConfigAction,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Subcommand, serde::Serialize, serde::Deserialize)]
|
#[derive(Debug, Subcommand, serde::Serialize, serde::Deserialize)]
|
||||||
pub enum ConfigAction {
|
pub enum ConfigAction {
|
||||||
#[command(
|
#[command(about = "To change current Noxis configuration")]
|
||||||
about = "To change current Noxis configuration",
|
|
||||||
)]
|
|
||||||
Local(LocalConfig),
|
Local(LocalConfig),
|
||||||
#[command(
|
#[command(about = "To change credentials of the remote config server")]
|
||||||
about = "To change credentials of the remote config server",
|
|
||||||
)]
|
|
||||||
Remote,
|
Remote,
|
||||||
#[command(
|
#[command(about = "To reset all config settings")]
|
||||||
about = "To reset all config settings",
|
|
||||||
)]
|
|
||||||
Reset,
|
Reset,
|
||||||
|
#[command(about = "To get current Noxis configuration", name = "ls")]
|
||||||
|
Show(EnvConfig),
|
||||||
|
}
|
||||||
|
#[derive(Debug, Parser, serde::Serialize, serde::Deserialize)]
|
||||||
|
pub struct EnvConfig {
|
||||||
|
// flag
|
||||||
|
#[arg(long = "env", action, help = "to read environment vars configuration")]
|
||||||
|
pub is_env: bool,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Parser, serde::Serialize, serde::Deserialize)]
|
#[derive(Debug, Parser, serde::Serialize, serde::Deserialize)]
|
||||||
pub struct LocalConfig {
|
pub struct LocalConfig {
|
||||||
// flag
|
// flag
|
||||||
#[arg(
|
#[arg(long = "json", action, help = "to read following input as JSON")]
|
||||||
long = "json",
|
pub is_json: bool,
|
||||||
action,
|
|
||||||
help = "to read following input as JSON",
|
|
||||||
)]
|
|
||||||
is_json : bool,
|
|
||||||
// value
|
// value
|
||||||
#[arg(
|
#[arg(help = "path to config file or config String (with --json flag)")]
|
||||||
help = "path to config file or config String (with --json flag)",
|
pub config: String,
|
||||||
)]
|
|
||||||
config : String,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Parser, serde::Serialize, serde::Deserialize)]
|
#[derive(Debug, Parser, serde::Serialize, serde::Deserialize)]
|
||||||
pub struct ProcessCommand {
|
pub struct ProcessCommand {
|
||||||
#[arg(
|
#[arg(help = "name of needed process")]
|
||||||
help = "name of needed process",
|
pub process: String,
|
||||||
)]
|
#[command(subcommand, help = "To get current process's status")]
|
||||||
process : String,
|
pub action: ProcessAction,
|
||||||
#[command(
|
|
||||||
subcommand,
|
|
||||||
help = "To get current process's status",
|
|
||||||
)]
|
|
||||||
action : ProcessAction,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Subcommand, serde::Serialize, serde::Deserialize)]
|
#[derive(Debug, Subcommand, serde::Serialize, serde::Deserialize)]
|
||||||
enum ProcessAction {
|
pub enum ProcessAction {
|
||||||
#[command(
|
#[command(about = "To get info about current process status")]
|
||||||
about = "To get info about current process status",
|
|
||||||
)]
|
|
||||||
Status,
|
Status,
|
||||||
#[command(
|
#[command(about = "To start current process")]
|
||||||
about = "To start current process",
|
|
||||||
)]
|
|
||||||
Start,
|
Start,
|
||||||
#[command(
|
#[command(about = "To stop current process")]
|
||||||
about = "To stop current process",
|
|
||||||
)]
|
|
||||||
Stop,
|
Stop,
|
||||||
#[command(
|
#[command(about = "To freeze (hybernaze) current process")]
|
||||||
about = "To freeze (hybernaze) current process",
|
|
||||||
)]
|
|
||||||
Freeze,
|
Freeze,
|
||||||
#[command(
|
#[command(about = "To unfreeze (unhybernaze) current process")]
|
||||||
about = "To unfreeze (unhybernaze) current process",
|
|
||||||
)]
|
|
||||||
Unfreeze,
|
Unfreeze,
|
||||||
#[command(
|
#[command(about = "To restart current process")]
|
||||||
about = "To restart current process",
|
|
||||||
)]
|
|
||||||
Restart,
|
Restart,
|
||||||
#[command(
|
#[command(about = "To get info about current process's dependencies")]
|
||||||
about = "To get info about current process's dependencies",
|
|
||||||
)]
|
|
||||||
Deps,
|
Deps,
|
||||||
#[command(
|
#[command(about = "To get info about current process's files-dependencies")]
|
||||||
about = "To get info about current process's files-dependencies",
|
|
||||||
)]
|
|
||||||
Files,
|
Files,
|
||||||
#[command(
|
#[command(about = "To get info about current process's services-dependencies")]
|
||||||
about = "To get info about current process's services-dependencies",
|
|
||||||
)]
|
|
||||||
Services,
|
Services,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub mod metrics_models {
|
||||||
|
#[derive(Debug, clap::Parser, serde::Serialize, serde::Deserialize)]
|
||||||
|
pub enum MetricsMode {
|
||||||
|
#[command(about = "To capture all metrics about undercontrolled system")]
|
||||||
|
Full,
|
||||||
|
// system
|
||||||
|
#[command(about = "To capture general host info")]
|
||||||
|
Host,
|
||||||
|
#[command(about = "To capture detailed CPU metrics")]
|
||||||
|
Cpu,
|
||||||
|
#[command(about = "To capture RAM metrics")]
|
||||||
|
Ram,
|
||||||
|
#[command(about = "To capture disk environment metrics")]
|
||||||
|
Rom,
|
||||||
|
#[command(about = "To capture system net interfaces metrics")]
|
||||||
|
Network,
|
||||||
|
// processes
|
||||||
|
#[command(about = "To capture monitoring processes metrics")]
|
||||||
|
Processes, // Config
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Cli {
|
||||||
|
pub fn validate_socket(mut self) -> Self {
|
||||||
|
if let Ok(path) = std::env::var("NOXIS_SOCKET_PATH") {
|
||||||
|
self.socket = path;
|
||||||
|
}
|
||||||
|
self
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
|
||||||
|
|
@ -1,14 +1,16 @@
|
||||||
use thiserror::Error;
|
use thiserror::Error;
|
||||||
use super::cli_net::NOXIS_RS_CREDS;
|
|
||||||
|
|
||||||
#[derive(Debug, Error)]
|
#[derive(Debug, Error)]
|
||||||
|
#[allow(dead_code)]
|
||||||
pub enum NoxisCliError {
|
pub enum NoxisCliError {
|
||||||
#[error("Can't send any data to {:?}. Noxis-rs daemon is disabled or can't be accessed", NOXIS_RS_CREDS)]
|
#[error("Can't find socket `{0}`. {1}")]
|
||||||
NoxisDaemonMissing,
|
NoxisDaemonMissing(String, String),
|
||||||
#[error("Noxis CLI can't write any data to the Noxis-rs port. Check daemon and it's web-functionality")]
|
#[error("Noxis CLI can't write any data to the Noxis-rs port. Check daemon and it's runtime!")]
|
||||||
PortIsNotWritable,
|
PortIsNotWritable,
|
||||||
#[error("Can't send Cli-prompt to the Noxis-rs. Check it's state")]
|
#[error("Can't send Cli-prompt to the Noxis-rs. Check it's state")]
|
||||||
CliPromptCanNotBeSent,
|
CliPromptCanNotBeSent,
|
||||||
#[error("Can't parse CLI struct and send as byte stream")]
|
#[error("Can't parse CLI struct and send as byte stream")]
|
||||||
ToStringCliParsingParsing,
|
ToStringCliParsingParsing,
|
||||||
|
#[error("Can't read Noxis response due to {0}")]
|
||||||
|
CliResponseReadError(String),
|
||||||
}
|
}
|
||||||
|
|
@ -1,32 +1,37 @@
|
||||||
use tokio::net::TcpStream;
|
|
||||||
use tokio::io::AsyncWriteExt;
|
|
||||||
use tokio::time::{Duration, sleep};
|
|
||||||
use anyhow::Result;
|
|
||||||
use super::Cli;
|
|
||||||
use super::cli_error::NoxisCliError;
|
use super::cli_error::NoxisCliError;
|
||||||
|
use super::Cli;
|
||||||
|
use anyhow::Result;
|
||||||
|
use tokio::io::{AsyncReadExt, AsyncWriteExt};
|
||||||
|
use tokio::net::UnixStream;
|
||||||
|
|
||||||
pub const NOXIS_RS_CREDS: &str = "127.0.0.1:7753";
|
#[allow(dead_code)]
|
||||||
|
async fn create_us_stream(cli: &Cli) -> Result<UnixStream> {
|
||||||
|
Ok(UnixStream::connect(&cli.socket).await.map_err(|er| {
|
||||||
pub async fn create_tcp_stream() -> Result<TcpStream> {
|
NoxisCliError::NoxisDaemonMissing((&cli.socket).to_string(), er.to_string())
|
||||||
Ok(TcpStream::connect(NOXIS_RS_CREDS).await.map_err(|_| NoxisCliError::NoxisDaemonMissing)?)
|
})?)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn try_send(stream: Result<TcpStream>, params: Cli) -> Result<()> {
|
#[allow(dead_code)]
|
||||||
use serde_json::to_string;
|
pub async fn try_send(cli: Cli) -> Result<()> {
|
||||||
let mut stream = stream.map_err(|_| NoxisCliError::NoxisDaemonMissing)?;
|
let mut stream = create_us_stream(&cli).await?;
|
||||||
loop {
|
|
||||||
if stream.writable().await.is_err() {
|
|
||||||
sleep(Duration::from_millis(100)).await;
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
// let msg: Cli = from_str(&format!("{:?}", params))?;
|
|
||||||
let msg= to_string(¶ms).map_err(|_| NoxisCliError::ToStringCliParsingParsing)?;
|
|
||||||
// let msg = r"HTTP/1.1 POST\r\nContent-Length: 14\r\nContent-Type: text/plain\r\n\r\nHello, World!@";
|
|
||||||
|
|
||||||
stream.write_all(msg.as_bytes()).await.map_err(|_| NoxisCliError::CliPromptCanNotBeSent)?;
|
let msg = serde_json::to_vec(&cli).map_err(|_| NoxisCliError::ToStringCliParsingParsing)?;
|
||||||
// ...
|
|
||||||
break;
|
stream
|
||||||
|
.write_all(&msg)
|
||||||
|
.await
|
||||||
|
.map_err(|_| NoxisCliError::CliPromptCanNotBeSent)?;
|
||||||
|
|
||||||
|
let mut response = Vec::new();
|
||||||
|
stream
|
||||||
|
.read_to_end(&mut response)
|
||||||
|
.await
|
||||||
|
.map_err(|er| NoxisCliError::CliResponseReadError(er.to_string()))?;
|
||||||
|
|
||||||
|
let response = String::from_utf8_lossy(&response);
|
||||||
|
for line in response.lines() {
|
||||||
|
println!("{}", line);
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
@ -1,5 +1,5 @@
|
||||||
mod cli;
|
mod cli;
|
||||||
mod cli_net;
|
|
||||||
mod cli_error;
|
mod cli_error;
|
||||||
|
mod cli_net;
|
||||||
|
|
||||||
pub use cli::*;
|
pub use cli::*;
|
||||||
|
|
@ -1,15 +1,16 @@
|
||||||
mod cli;
|
mod cli;
|
||||||
mod cli_net;
|
|
||||||
mod cli_error;
|
mod cli_error;
|
||||||
|
mod cli_net;
|
||||||
|
|
||||||
|
use anyhow::Result;
|
||||||
use clap::Parser;
|
use clap::Parser;
|
||||||
use cli::Cli;
|
use cli::Cli;
|
||||||
use cli_net::{create_tcp_stream, try_send};
|
use cli_net::try_send;
|
||||||
use anyhow::Result;
|
|
||||||
|
|
||||||
#[tokio::main]
|
#[tokio::main]
|
||||||
async fn main() -> Result<()>{
|
async fn main() -> Result<()> {
|
||||||
let cli = Cli::parse();
|
dotenv::dotenv().ok();
|
||||||
try_send(create_tcp_stream().await, cli).await?;
|
let cli = Cli::parse().validate_socket();
|
||||||
|
try_send(cli).await?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,4 @@
|
||||||
|
.env
|
||||||
|
.env.example
|
||||||
|
README.md
|
||||||
|
target
|
||||||
|
|
@ -0,0 +1,2 @@
|
||||||
|
NOXIS_SOCKET_PATH = "/path/to/noxis.sock"
|
||||||
|
NOXIS_PROXY_PORT = "numport"
|
||||||
|
|
@ -0,0 +1,12 @@
|
||||||
|
[package]
|
||||||
|
name = "noxis-proxy"
|
||||||
|
version = "0.1.0"
|
||||||
|
edition = "2024"
|
||||||
|
|
||||||
|
[dependencies]
|
||||||
|
anyhow = "1.0.98"
|
||||||
|
axum = { version = "0.8.4", features = ["ws"] }
|
||||||
|
dotenv = "0.15.0"
|
||||||
|
tokio = { version = "1.45.1", features = ["full"] }
|
||||||
|
tracing = "0.1.41"
|
||||||
|
tracing-subscriber = "0.3.19"
|
||||||
|
|
@ -0,0 +1,19 @@
|
||||||
|
FROM rust:latest AS builder
|
||||||
|
WORKDIR /app
|
||||||
|
|
||||||
|
RUN apt update && apt install -y musl-tools
|
||||||
|
RUN rustup target add x86_64-unknown-linux-musl
|
||||||
|
|
||||||
|
COPY . .
|
||||||
|
|
||||||
|
RUN cargo build --release --target=x86_64-unknown-linux-musl
|
||||||
|
|
||||||
|
FROM alpine:latest
|
||||||
|
WORKDIR /app
|
||||||
|
|
||||||
|
COPY --from=builder /app/target/x86_64-unknown-linux-musl/release/noxis-proxy /app/noxis-proxy
|
||||||
|
RUN apk add --no-cache ca-certificates
|
||||||
|
|
||||||
|
EXPOSE 7654
|
||||||
|
|
||||||
|
ENTRYPOINT ["/app/noxis-proxy"]
|
||||||
|
|
@ -0,0 +1,20 @@
|
||||||
|
services:
|
||||||
|
noxis-proxy:
|
||||||
|
container_name: noxis-proxy
|
||||||
|
image: noxis-proxy:0.1.0
|
||||||
|
networks:
|
||||||
|
- noxis-net
|
||||||
|
environment:
|
||||||
|
- NOXIS_SOCKET_PATH=./noxis.sock
|
||||||
|
- NOXIS_PROXY_PORT=7654
|
||||||
|
- NOXIS_LOG_LEVEL=TRACE
|
||||||
|
volumes:
|
||||||
|
- /home/user/diplom_code/noxis-rs/noxis.sock:/app/noxis.sock
|
||||||
|
ports:
|
||||||
|
- 7654:7654
|
||||||
|
restart: always
|
||||||
|
|
||||||
|
networks:
|
||||||
|
noxis-net:
|
||||||
|
driver: bridge
|
||||||
|
|
||||||
|
|
@ -0,0 +1,96 @@
|
||||||
|
use axum::{
|
||||||
|
extract::{
|
||||||
|
ws::{Message, WebSocket, WebSocketUpgrade},
|
||||||
|
State,
|
||||||
|
},
|
||||||
|
response::IntoResponse,
|
||||||
|
routing::get,
|
||||||
|
Router,
|
||||||
|
};
|
||||||
|
use std::{
|
||||||
|
path::PathBuf, str::FromStr,
|
||||||
|
};
|
||||||
|
use tokio::net::UnixStream;
|
||||||
|
use tokio::io::{AsyncReadExt, AsyncWriteExt};
|
||||||
|
|
||||||
|
#[derive(Clone)]
|
||||||
|
struct AppState {
|
||||||
|
socket_path: PathBuf,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::main]
|
||||||
|
async fn main() -> anyhow::Result<()> {
|
||||||
|
dotenv::dotenv().ok();
|
||||||
|
|
||||||
|
tracing_subscriber::fmt()
|
||||||
|
.with_max_level(tracing::Level::from_str(&std::env::var("NOXIS_LOG_LEVEL").unwrap_or_else(|_| String::from("INFO"))).unwrap_or_else(|_| tracing::Level::INFO))
|
||||||
|
.with_writer(std::io::stdout)
|
||||||
|
.compact()
|
||||||
|
.init();
|
||||||
|
|
||||||
|
let app_state = AppState {
|
||||||
|
socket_path : PathBuf::new().join(std::env::var("NOXIS_SOCKET_PATH").unwrap_or_else(|_| String::from("./noxis.sock")))
|
||||||
|
};
|
||||||
|
let app = Router::new()
|
||||||
|
.route("/ws", get(ws_handler))
|
||||||
|
.route("/hello", get(hello))
|
||||||
|
.with_state(app_state);
|
||||||
|
|
||||||
|
let bind = format!("0.0.0.0:{}", std::env::var("NOXIS_PROXY_PORT").unwrap_or_else(|_| String::from("7654")));
|
||||||
|
|
||||||
|
tracing::info!("Serving on {}", &bind);
|
||||||
|
|
||||||
|
let listener = tokio::net::TcpListener::bind(bind)
|
||||||
|
.await?;
|
||||||
|
axum::serve(listener, app).await?;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn ws_handler(
|
||||||
|
ws: WebSocketUpgrade,
|
||||||
|
State(state): State<AppState>,
|
||||||
|
) -> impl IntoResponse {
|
||||||
|
tracing::info!("New WebSocket connection");
|
||||||
|
ws.on_upgrade(|socket| handle_socket(socket, state))
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn hello(
|
||||||
|
State(_state): State<AppState>,
|
||||||
|
) -> impl IntoResponse {
|
||||||
|
String::from("HELLO")
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn handle_socket(mut ws: WebSocket, state: AppState) {
|
||||||
|
tracing::info!("handle websocket");
|
||||||
|
|
||||||
|
let ws_receiver = tokio::spawn(async move {
|
||||||
|
while let Some(Ok(msg)) = ws.recv().await {
|
||||||
|
let mut unix_socket = match UnixStream::connect(&state.socket_path).await {
|
||||||
|
Ok(socket) => socket,
|
||||||
|
Err(e) => {
|
||||||
|
tracing::error!("Failed to connect to Unix socket: {}", e);
|
||||||
|
let _ = ws.send(Message::Text("ERROR: Unix socket connection failed".into())).await;
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
if let Message::Text(text) = msg {
|
||||||
|
if let Err(e) = unix_socket.write_all(text.as_bytes()).await {
|
||||||
|
tracing::error!("Failed to write to Unix socket: {}", e);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
let mut buf = Vec::new();
|
||||||
|
match unix_socket.read_to_end(&mut buf).await {
|
||||||
|
Ok(n) if n > 0 => {
|
||||||
|
let response = String::from_utf8_lossy(&buf[..n]);
|
||||||
|
if ws.send(Message::Text(response.into_owned().into())).await.is_err() {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Ok(_) | Err(_) => break,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
let _ = ws_receiver.await;
|
||||||
|
}
|
||||||
|
|
@ -0,0 +1,16 @@
|
||||||
|
# ACTIONS (trigger on exist)
|
||||||
|
NOXIS_NO_HAGENT = "true"
|
||||||
|
# NOXIS_NO_LOGS = "true"
|
||||||
|
NOXIS_REFRESH_LOGS = "true"
|
||||||
|
NOXIS_NO_SUB = "true"
|
||||||
|
|
||||||
|
# VALUES
|
||||||
|
NOXIS_HAGENT_SOCKET_PATH = "/var/run/example/hostagent.sock"
|
||||||
|
NOXIS_LOG_TO = "/var/log/noxis/noxis.log"
|
||||||
|
NOXIS_REMOTE_SERVER_URL = "ip.ip.ip.ip:port"
|
||||||
|
NOXIS_CONFIG_PATH = "./settings.json"
|
||||||
|
NOXIS_METRICS_MODE = "full"
|
||||||
|
NOXIS_SOCKET_PATH = "/path/to/noxis.sock"
|
||||||
|
NOXIS_BACKUP_FOLDER = "/path/to/backups/folder"
|
||||||
|
|
||||||
|
NOXIS_MAX_LOG_LEVEL = "TRACE"
|
||||||
|
|
@ -1,6 +1,6 @@
|
||||||
[package]
|
[package]
|
||||||
name = "noxis-rs"
|
name = "noxis-rs"
|
||||||
version = "0.11.10"
|
version = "0.12.0"
|
||||||
edition = "2021"
|
edition = "2021"
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
|
|
@ -8,13 +8,18 @@ anyhow = "1.0.93"
|
||||||
chrono = "0.4.38"
|
chrono = "0.4.38"
|
||||||
clap = { version = "4.5.21", features = ["derive"] }
|
clap = { version = "4.5.21", features = ["derive"] }
|
||||||
env_logger = "0.11.3"
|
env_logger = "0.11.3"
|
||||||
inotify = "0.10.2"
|
inotify = "0.11.0"
|
||||||
log = "0.4.22"
|
log = "0.4.22"
|
||||||
pcap = "2.2.0"
|
pcap = "2.2.0"
|
||||||
redis = "0.25.4"
|
redis = "0.29.2"
|
||||||
serde = { version = "1.0.203", features = ["derive"] }
|
serde = { version = "1.0.203", features = ["derive"] }
|
||||||
serde_json = "1.0.118"
|
serde_json = "1.0.118"
|
||||||
sysinfo = "0.32.0"
|
sysinfo = "0.32.0"
|
||||||
tokio = { version = "1.38.0", features = ["full", "time"] }
|
tokio = { version = "1.38.0", features = ["full", "time"] }
|
||||||
noxis-cli = { path = "../noxis-cli" }
|
noxis-cli = { path = "../noxis-cli" }
|
||||||
dotenv = "0.15.0"
|
dotenv = "0.15.0"
|
||||||
|
futures = "0.3.31"
|
||||||
|
async-trait = "0.1.88"
|
||||||
|
crossbeam = { version = "0.8.4", features = ["crossbeam-channel"] }
|
||||||
|
lazy_static = "1.5.0"
|
||||||
|
ulid = "1.2.1"
|
||||||
|
|
|
||||||
|
|
@ -1,6 +1,5 @@
|
||||||
{
|
{
|
||||||
"dateOfCreation": "1721381809104",
|
"dateOfCreation": "1721381809122",
|
||||||
"configServer": "localhost",
|
|
||||||
"processes": [
|
"processes": [
|
||||||
{
|
{
|
||||||
"name": "temp-process",
|
"name": "temp-process",
|
||||||
|
|
@ -12,7 +11,35 @@
|
||||||
"src": "./tests/examples/",
|
"src": "./tests/examples/",
|
||||||
"triggers": {
|
"triggers": {
|
||||||
"onDelete": "stop",
|
"onDelete": "stop",
|
||||||
"onChange": "stay"
|
"onChange": "restart",
|
||||||
|
"doRestore": true
|
||||||
|
}
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"filename": "none.json",
|
||||||
|
"src": "./tests/examples/",
|
||||||
|
"triggers": {
|
||||||
|
"onDelete": "stop",
|
||||||
|
"onChange": "restart",
|
||||||
|
"doRestore": false
|
||||||
|
}
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"filename": "invalid_config.json",
|
||||||
|
"src": "./tests/examples/",
|
||||||
|
"triggers": {
|
||||||
|
"onDelete": "stop",
|
||||||
|
"onChange": "restart",
|
||||||
|
"doRestore": false
|
||||||
|
}
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"filename": "save-conf.json",
|
||||||
|
"src": "./tests/examples/",
|
||||||
|
"triggers": {
|
||||||
|
"onDelete": "stop",
|
||||||
|
"onChange": "restart",
|
||||||
|
"doRestore": true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
],
|
],
|
||||||
|
|
@ -21,9 +48,16 @@
|
||||||
"hostname": "ya.ru",
|
"hostname": "ya.ru",
|
||||||
"port": 443,
|
"port": 443,
|
||||||
"triggers": {
|
"triggers": {
|
||||||
"wait": 10,
|
"wait": 2,
|
||||||
"delay": 2,
|
"onLost": "stop"
|
||||||
"onLost": "hold"
|
}
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"hostname": "8.8.8.8",
|
||||||
|
"port": 443,
|
||||||
|
"triggers": {
|
||||||
|
"wait": 2,
|
||||||
|
"onLost": "stop"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
|
|
|
||||||
|
|
@ -1,101 +1,133 @@
|
||||||
mod options;
|
mod options;
|
||||||
mod utils;
|
mod utils;
|
||||||
|
|
||||||
use anyhow::Error;
|
|
||||||
use clap::Parser;
|
|
||||||
use log::{error, info};
|
use log::{error, info};
|
||||||
use options::config::*;
|
|
||||||
use options::logger::setup_logger;
|
|
||||||
use options::signals::set_valid_destructor;
|
|
||||||
use options::structs::Processes;
|
|
||||||
use options::cli_pipeline::init_cli_pipeline;
|
use options::cli_pipeline::init_cli_pipeline;
|
||||||
|
use options::config::v2::init_config_mechanism;
|
||||||
|
use options::logger::setup_logger;
|
||||||
|
use options::preboot::PrebootParams;
|
||||||
|
use options::signals::set_valid_destructor;
|
||||||
|
use options::structs::ProcessUnit;
|
||||||
|
use options::structs::{bus::BusMessage, Processes};
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use std::time::Duration;
|
use std::time::Duration;
|
||||||
use tokio::sync::mpsc;
|
use tokio::sync::{broadcast, mpsc, oneshot};
|
||||||
use utils::*;
|
use utils::bus::Bus;
|
||||||
use options::preboot::PrebootParams;
|
use utils::metrics::init_metrics_grubber;
|
||||||
|
use utils::v2::init_monitoring;
|
||||||
#[tokio::main(flavor = "multi_thread")]
|
|
||||||
async fn main() -> anyhow::Result<()>{
|
|
||||||
let preboot = Arc::new(PrebootParams::parse().validate()?);
|
|
||||||
|
|
||||||
|
#[tokio::main(flavor = "multi_thread", worker_threads = 4)]
|
||||||
|
async fn main() -> anyhow::Result<()> {
|
||||||
|
let preboot = Arc::new(PrebootParams::validate());
|
||||||
let _ = setup_logger();
|
let _ = setup_logger();
|
||||||
|
|
||||||
info!("Runner is configurating...");
|
info!("Noxis is configurating...");
|
||||||
|
//
|
||||||
// setting up redis connection \
|
let (tx_brd, mut rx_brd) = broadcast::channel::<Processes>(1);
|
||||||
// then conf checks to choose the most actual \
|
// for cli to get config
|
||||||
let processes: Processes = get_actual_config(preboot.clone()).await.unwrap_or_else(|| {
|
let mut rx_cli_brd = tx_brd.subscribe();
|
||||||
error!("No actual configuration for runner. Stopping...");
|
// cli <-> config
|
||||||
std::process::exit(1);
|
let (tx_oneshot, rx_oneshot) = oneshot::channel::<Processes>();
|
||||||
});
|
|
||||||
|
|
||||||
info!(
|
|
||||||
"Current runner configuration: {}",
|
|
||||||
&processes.date_of_creation
|
|
||||||
);
|
|
||||||
info!("Runner is ready. Initializing...");
|
|
||||||
|
|
||||||
if processes.processes.is_empty() {
|
|
||||||
error!("Processes list is null, runner-rs initialization is stopped");
|
|
||||||
return Err(Error::msg("Empty processes segment in config"));
|
|
||||||
}
|
|
||||||
let mut handler: Vec<tokio::task::JoinHandle<()>> = vec![];
|
let mut handler: Vec<tokio::task::JoinHandle<()>> = vec![];
|
||||||
// is in need to send to the signals handler thread
|
|
||||||
let mut senders: Vec<Arc<mpsc::Sender<u8>>> = vec![];
|
|
||||||
|
|
||||||
for proc in processes.processes.iter() {
|
// to BUS channel
|
||||||
info!(
|
let (tx_to_bus, rx_to_bus) = mpsc::channel::<BusMessage>(5);
|
||||||
"Process '{}' on stage: {}. Depends on {} file(s), {} service(s)",
|
// from BUS channels
|
||||||
proc.name,
|
let (tx_to_cli, rx_to_cli) = mpsc::channel::<BusMessage>(5);
|
||||||
proc.path,
|
let (tx_to_supervisor, rx_to_supervisor) = mpsc::channel::<BusMessage>(5);
|
||||||
proc.dependencies.files.len(),
|
let (tx_to_metrics, rx_to_metrics) = mpsc::channel::<BusMessage>(5);
|
||||||
proc.dependencies.services.len()
|
|
||||||
|
let tx_to_bus = Arc::new(tx_to_bus);
|
||||||
|
let tx_to_cli = Arc::new(tx_to_cli);
|
||||||
|
let tx_to_supervisor = Arc::new(tx_to_supervisor);
|
||||||
|
let tx_to_metrics = Arc::new(tx_to_metrics);
|
||||||
|
|
||||||
|
let bus_module = tokio::spawn(async move {
|
||||||
|
let mut bus = Bus::new(
|
||||||
|
rx_to_bus,
|
||||||
|
tx_to_cli.clone(),
|
||||||
|
tx_to_supervisor.clone(),
|
||||||
|
tx_to_metrics.clone(),
|
||||||
);
|
);
|
||||||
|
bus.process().await;
|
||||||
// creating msg channel
|
error!("Info Bus crushed !");
|
||||||
// can or should be executed in new thread
|
|
||||||
let (tx, mut rx) = mpsc::channel::<u8>(1);
|
|
||||||
let proc = Arc::new(proc.clone());
|
|
||||||
let tx = Arc::new(tx.clone());
|
|
||||||
|
|
||||||
senders.push(Arc::clone(&tx.clone()));
|
|
||||||
|
|
||||||
let event = tokio::spawn(async move {
|
|
||||||
run_daemons(proc.clone(), tx.clone(), &mut rx).await;
|
|
||||||
});
|
});
|
||||||
handler.push(event);
|
handler.push(bus_module);
|
||||||
}
|
|
||||||
|
|
||||||
// destructor addition
|
// initilaizing task for config manipulations
|
||||||
handler.push(tokio::spawn(async move {
|
let preboot_config = preboot.clone();
|
||||||
if set_valid_destructor(Arc::new(senders)).await.is_err() {
|
let config_module = tokio::spawn(async move {
|
||||||
error!("Linux signals handler creation failed. Terminating main thread...");
|
let _ = init_config_mechanism(rx_oneshot, tx_brd, preboot_config).await;
|
||||||
return;
|
});
|
||||||
}
|
handler.push(config_module);
|
||||||
|
|
||||||
tokio::time::sleep(Duration::from_millis(200)).await;
|
// initilaizing task for cli manipulation
|
||||||
info!("End of job. Terminating main thread...");
|
let tx_bus = tx_to_bus.clone();
|
||||||
|
let preboot_cli = preboot.clone();
|
||||||
|
let cli_module = tokio::spawn(async move {
|
||||||
|
let config = {
|
||||||
|
let mut tick = tokio::time::interval(Duration::from_millis(500));
|
||||||
|
loop {
|
||||||
|
tick.tick().await;
|
||||||
|
break match rx_cli_brd.try_recv() {
|
||||||
|
Ok(conf) => conf,
|
||||||
|
Err(_) => continue,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
};
|
||||||
|
if let Err(er) = init_cli_pipeline(
|
||||||
|
preboot_cli,
|
||||||
|
Arc::new(config),
|
||||||
|
tx_oneshot,
|
||||||
|
rx_to_cli,
|
||||||
|
tx_bus.clone(),
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
{
|
||||||
|
error!("CLI pipeline failed due to {}", er)
|
||||||
|
}
|
||||||
|
});
|
||||||
|
handler.push(cli_module);
|
||||||
|
|
||||||
|
// metrics
|
||||||
|
let tx_bus = tx_to_bus.clone();
|
||||||
|
let metrics_module = tokio::spawn(async move {
|
||||||
|
if let Err(er) = init_metrics_grubber(tx_bus.clone(), rx_to_metrics).await {
|
||||||
|
error!("Metrics module crushed : {}", er);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
handler.push(metrics_module);
|
||||||
|
|
||||||
|
// initilaizing task for deinitializing `Noxis`
|
||||||
|
let ctrlc = tokio::spawn(async move {
|
||||||
|
if let Err(er) = set_valid_destructor(vec![].into()).await {
|
||||||
|
error!("Destructor mod failed due to {}", er);
|
||||||
|
}
|
||||||
std::process::exit(0);
|
std::process::exit(0);
|
||||||
}));
|
});
|
||||||
|
handler.push(ctrlc);
|
||||||
|
|
||||||
// remote config update subscription
|
let tx_bus = tx_to_bus.clone();
|
||||||
handler.push(tokio::spawn(async move {
|
let preboot_cli = preboot.clone();
|
||||||
let _ = subscribe_config_stream(Arc::new(processes), preboot.clone()).await;
|
let monitoring = tokio::spawn(async move {
|
||||||
}));
|
let config = {
|
||||||
|
let mut tick = tokio::time::interval(Duration::from_millis(500));
|
||||||
// cli pipeline
|
loop {
|
||||||
handler.push(tokio::spawn(async move {
|
tick.tick().await;
|
||||||
let _ = init_cli_pipeline().await;
|
break match rx_brd.try_recv() {
|
||||||
}));
|
Ok(conf) => conf,
|
||||||
|
Err(_) => continue,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
};
|
||||||
|
if let Err(er) = init_monitoring(config, preboot_cli, rx_to_supervisor, tx_bus).await {
|
||||||
|
error!("Monitoring mod failed due to {}", er);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
handler.push(monitoring);
|
||||||
|
|
||||||
for i in handler {
|
for i in handler {
|
||||||
let _ = i.await;
|
let _ = i.await;
|
||||||
}
|
}
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
// todo: integration tests
|
|
||||||
// todo: config pulling mechanism rework (socket)
|
|
||||||
// todo: tasks management after killing all processes
|
|
||||||
// todo:
|
|
||||||
|
|
|
||||||
|
|
@ -1,8 +1,8 @@
|
||||||
// ! gathering optional items module
|
// ! gathering optional items module
|
||||||
|
|
||||||
|
pub mod cli_pipeline;
|
||||||
pub mod config;
|
pub mod config;
|
||||||
pub mod logger;
|
pub mod logger;
|
||||||
|
pub mod preboot;
|
||||||
pub mod signals;
|
pub mod signals;
|
||||||
pub mod structs;
|
pub mod structs;
|
||||||
pub mod preboot;
|
|
||||||
pub mod cli_pipeline;
|
|
||||||
|
|
@ -1,12 +1,23 @@
|
||||||
use log::{error, info, warn};
|
use super::structs::bus::BusMessage;
|
||||||
use tokio::net::{TcpListener, TcpStream};
|
use super::structs::Processes;
|
||||||
use anyhow::{Result as DynResult, Error};
|
use crate::options::structs::bus::InternalCli;
|
||||||
|
use log::{error, info};
|
||||||
|
use noxis_cli::{Cli, ProcessAction};
|
||||||
|
use std::any::Any;
|
||||||
|
use std::fs;
|
||||||
|
use std::sync::Arc;
|
||||||
|
use tokio::io::{AsyncReadExt, AsyncWriteExt};
|
||||||
|
use tokio::net::{UnixListener, UnixStream};
|
||||||
|
use tokio::sync::{Mutex, OnceCell};
|
||||||
use tokio::time::{sleep, Duration};
|
use tokio::time::{sleep, Duration};
|
||||||
use std::{borrow::BorrowMut, net::{IpAddr, Ipv4Addr}};
|
|
||||||
// use std::io::BufReader;
|
use super::preboot::PrebootParams;
|
||||||
use tokio::io::{BufReader, AsyncWriteExt, AsyncBufReadExt};
|
|
||||||
use noxis_cli::Cli;
|
type ConfigGateway = tokio::sync::oneshot::Sender<Processes>;
|
||||||
use serde_json::from_str;
|
type ProcessedConfigGateway = Arc<Mutex<OnceCell<ConfigGateway>>>;
|
||||||
|
type BusReciever = tokio::sync::mpsc::Receiver<BusMessage>;
|
||||||
|
type BusSender = Arc<tokio::sync::mpsc::Sender<BusMessage>>;
|
||||||
|
type ReadyBusReciever = Arc<Mutex<tokio::sync::mpsc::Receiver<BusMessage>>>;
|
||||||
|
|
||||||
/// # Fn `init_cli_pipeline`
|
/// # Fn `init_cli_pipeline`
|
||||||
/// ## for catching all input requests from CLI
|
/// ## for catching all input requests from CLI
|
||||||
|
|
@ -21,48 +32,56 @@ use serde_json::from_str;
|
||||||
///
|
///
|
||||||
/// *depends on* : -
|
/// *depends on* : -
|
||||||
///
|
///
|
||||||
pub async fn init_cli_pipeline() -> DynResult<()> {
|
pub async fn init_cli_pipeline(
|
||||||
match init_listener().await {
|
params: Arc<PrebootParams>,
|
||||||
Some(list) => {
|
config: Arc<Processes>,
|
||||||
|
config_gateway: ConfigGateway,
|
||||||
|
bus_reciever: BusReciever,
|
||||||
|
bus_sender: BusSender,
|
||||||
|
) -> anyhow::Result<()> {
|
||||||
|
let socket_path = ¶ms.self_socket;
|
||||||
|
let _ = fs::remove_file(socket_path);
|
||||||
|
|
||||||
|
let config_gateway = Arc::new(Mutex::new(OnceCell::new_with(Some(config_gateway))));
|
||||||
|
let bus_reciever = Arc::new(Mutex::new(bus_reciever));
|
||||||
|
|
||||||
|
match UnixListener::bind(socket_path) {
|
||||||
|
Ok(list) => {
|
||||||
|
// TODO: remove `unwrap`s
|
||||||
|
info!("Listening on {}", socket_path.display());
|
||||||
|
std::env::set_var("NOXIS_SOCKET_PATH", socket_path);
|
||||||
loop {
|
loop {
|
||||||
if let Ok((socket, addr)) = list.accept().await {
|
match list.accept().await {
|
||||||
// isolation
|
Ok((socket, _)) => {
|
||||||
if IpAddr::V4(Ipv4Addr::new(127, 0, 0, 1)) != addr.ip() {
|
// ??? maybe errors on async work with data transfering between modules
|
||||||
warn!("Declined attempt to connect TCP-socket from {}", addr);
|
let params = params.clone();
|
||||||
continue;
|
let config = config.clone();
|
||||||
|
let config_gateway = config_gateway.clone();
|
||||||
|
let bus_reciever = bus_reciever.clone();
|
||||||
|
let bus_sender = bus_sender.clone();
|
||||||
|
tokio::spawn(async move {
|
||||||
|
process_connection(
|
||||||
|
socket,
|
||||||
|
params.clone(),
|
||||||
|
config.clone(),
|
||||||
|
config_gateway.clone(),
|
||||||
|
bus_reciever,
|
||||||
|
bus_sender,
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
});
|
||||||
|
}
|
||||||
|
Err(er) => {
|
||||||
|
error!("Cannot poll connection to CLI due to {}", er);
|
||||||
|
sleep(Duration::from_millis(300)).await;
|
||||||
}
|
}
|
||||||
process_connection(socket).await;
|
|
||||||
}
|
}
|
||||||
sleep(Duration::from_millis(500)).await;
|
|
||||||
}
|
}
|
||||||
// Ok(())
|
// Ok(())
|
||||||
},
|
|
||||||
None => Err(Error::msg("Addr 127.0.0.1:7753 is already in use"))
|
|
||||||
}
|
}
|
||||||
}
|
Err(er) => {
|
||||||
|
error!("Failed to open UnixListener for CLI");
|
||||||
/// # Fn `init_listener`
|
Err(er.into())
|
||||||
/// ## for creating TCP-listener for communicating with CLI
|
|
||||||
///
|
|
||||||
/// *input* : -
|
|
||||||
///
|
|
||||||
/// *output* : `Some<TcpListener>` if port 7753 was opened | None if not
|
|
||||||
///
|
|
||||||
/// *initiator* : fn `init_cli_pipeline`
|
|
||||||
///
|
|
||||||
/// *managing* : `TcpListener` object to handle requests
|
|
||||||
///
|
|
||||||
/// *depends on* : `tokio::net::TcpListener`
|
|
||||||
///
|
|
||||||
async fn init_listener() -> Option<TcpListener> {
|
|
||||||
match TcpListener::bind("127.0.0.1:7753").await {
|
|
||||||
Ok(listener) => {
|
|
||||||
info!("Runner is listening localhost:7753");
|
|
||||||
Some(listener)
|
|
||||||
},
|
|
||||||
Err(_) => {
|
|
||||||
error!("Cannot create TCP listener for CLI");
|
|
||||||
None
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -80,27 +99,252 @@ async fn init_listener() -> Option<TcpListener> {
|
||||||
///
|
///
|
||||||
/// *depends on* : `tokio::net::TcpStream`
|
/// *depends on* : `tokio::net::TcpStream`
|
||||||
///
|
///
|
||||||
async fn process_connection(mut stream: TcpStream) {
|
async fn process_connection(
|
||||||
let buf_reader = BufReader::new(stream.borrow_mut());
|
mut stream: UnixStream,
|
||||||
let mut rqst = buf_reader.lines();
|
params: Arc<PrebootParams>,
|
||||||
|
config: Arc<Processes>,
|
||||||
|
cfg_gateway: ProcessedConfigGateway,
|
||||||
while let Ok(Some(line)) = rqst.next_line().await {
|
bus_reciever: ReadyBusReciever,
|
||||||
if line.is_empty() {
|
bus_sender: BusSender,
|
||||||
break
|
) {
|
||||||
|
let mut buf = vec![0; 1024];
|
||||||
|
match stream.read(&mut buf).await {
|
||||||
|
Ok(0) => {
|
||||||
|
info!("Client disconnected ");
|
||||||
}
|
}
|
||||||
match from_str::<Cli>(&line) {
|
Ok(n) => {
|
||||||
Ok(req) => {
|
buf.truncate(n);
|
||||||
// TODO: func wrapper
|
info!("CLI have sent {} bytes", n);
|
||||||
dbg!(req);
|
match serde_json::from_slice::<Cli>(&buf) {
|
||||||
},
|
Ok(cli) => {
|
||||||
Err(_) => {
|
info!("Received CLI request: {:?}", cli);
|
||||||
break
|
let response = match process_cli_cmd(
|
||||||
},
|
cli,
|
||||||
|
params.clone(),
|
||||||
|
config,
|
||||||
|
cfg_gateway.clone(),
|
||||||
|
bus_reciever.clone(),
|
||||||
|
bus_sender.clone(),
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
{
|
||||||
|
Ok(response) => response,
|
||||||
|
Err(er) => {
|
||||||
|
let error_msg = format!("Error: {}", er);
|
||||||
|
error!("{}", &error_msg);
|
||||||
|
error_msg
|
||||||
}
|
}
|
||||||
println!("{}", line);
|
};
|
||||||
|
for line in response.lines() {
|
||||||
|
if let Err(er) = stream.write_all(line.as_bytes()).await {
|
||||||
|
error!("Failed to send response: {}", er);
|
||||||
}
|
}
|
||||||
|
}
|
||||||
let response = "HTTP/1.1 200 OK\r\nContent-Length: 13\r\nContent-Type: text/plain\r\n\r\nHello, World!";
|
}
|
||||||
stream.write_all(response.as_bytes()).await.unwrap();
|
Err(e) => {
|
||||||
|
error!("Failed to parse CLI request: {}", e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(e) => error!("Failed to read from socket: {}", e),
|
||||||
|
}
|
||||||
|
let _ = stream.shutdown().await;
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn process_cli_cmd(
|
||||||
|
cli: Cli,
|
||||||
|
params: Arc<PrebootParams>,
|
||||||
|
global_config: Arc<Processes>,
|
||||||
|
cfg_gateway: ProcessedConfigGateway,
|
||||||
|
bus_reciever: ReadyBusReciever,
|
||||||
|
bus_sender: BusSender,
|
||||||
|
) -> anyhow::Result<String> {
|
||||||
|
use noxis_cli::{Commands, ConfigAction};
|
||||||
|
return match cli.command {
|
||||||
|
Commands::Config(config) => {
|
||||||
|
match config.action {
|
||||||
|
ConfigAction::Show(env) => {
|
||||||
|
if env.is_env {
|
||||||
|
Ok(serde_json::to_string_pretty(params.as_ref())?)
|
||||||
|
} else {
|
||||||
|
/* */
|
||||||
|
Ok(serde_json::to_string_pretty(global_config.as_ref())?)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
ConfigAction::Reset => Err(anyhow::Error::msg(
|
||||||
|
"It's temporarly forbidden to reset current config using CLI-util",
|
||||||
|
)),
|
||||||
|
ConfigAction::Local(cfg) => {
|
||||||
|
if cfg.is_json {
|
||||||
|
/* */
|
||||||
|
let new_config = serde_json::from_str::<Processes>(&cfg.config)?;
|
||||||
|
let new_version = new_config.get_version().to_string();
|
||||||
|
|
||||||
|
use super::{config::config_comparing, structs::ConfigActuality};
|
||||||
|
|
||||||
|
return match config_comparing(&global_config, &new_config) {
|
||||||
|
ConfigActuality::Remote => {
|
||||||
|
let cfg_gateway = cfg_gateway.clone();
|
||||||
|
tokio::spawn(async move {
|
||||||
|
let mut lock = cfg_gateway.lock().await;
|
||||||
|
match lock.take() {
|
||||||
|
Some(channel) => {
|
||||||
|
let _ = channel.send(new_config);
|
||||||
|
}
|
||||||
|
None => error!(
|
||||||
|
"Cannot update confif due to channel sender loss"
|
||||||
|
),
|
||||||
|
}
|
||||||
|
});
|
||||||
|
Ok(format!(
|
||||||
|
"Ok. Saving and reloading with version {}",
|
||||||
|
new_version
|
||||||
|
))
|
||||||
|
}
|
||||||
|
_ => Err(anyhow::Error::msg(format!(
|
||||||
|
"Local config (version: {}) is more actual",
|
||||||
|
global_config.get_version()
|
||||||
|
))),
|
||||||
|
};
|
||||||
|
} else {
|
||||||
|
Err(anyhow::Error::msg(
|
||||||
|
"It's temporarly forbidden to set config in non-json mode",
|
||||||
|
))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
ConfigAction::Remote => Ok(params.remote_server_url.clone()),
|
||||||
|
/* */
|
||||||
|
// _ => Err(anyhow::Error::msg("Unrecognized command from CLI"))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Commands::Processes => {
|
||||||
|
use crate::options::structs::bus::{BusMessageContentType, BusMessageDirection};
|
||||||
|
use crate::utils::metrics::processes::ProcessesQuery;
|
||||||
|
|
||||||
|
let _ = bus_sender
|
||||||
|
.send(BusMessage::Request(
|
||||||
|
BusMessageDirection::ToSupervisor,
|
||||||
|
BusMessageContentType::ProcessQuery,
|
||||||
|
Box::new(ProcessesQuery::QueryAll),
|
||||||
|
))
|
||||||
|
.await;
|
||||||
|
|
||||||
|
let mut bus = bus_reciever.lock().await;
|
||||||
|
let resp = tokio::time::timeout(std::time::Duration::from_secs(5), async move {
|
||||||
|
loop {
|
||||||
|
if let Ok(cont) = bus.try_recv() {
|
||||||
|
return cont;
|
||||||
|
}
|
||||||
|
tokio::time::sleep(std::time::Duration::from_millis(500)).await;
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.await?;
|
||||||
|
|
||||||
|
if let BusMessage::Response(_, _, content) = resp {
|
||||||
|
let content: Box<dyn Any> = content;
|
||||||
|
if let Ok(resp) = content.downcast::<anyhow::Result<String>>() {
|
||||||
|
return Ok((*resp)?);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(anyhow::Error::msg(format!(
|
||||||
|
"Unknown type of response from the Supervisor"
|
||||||
|
)))
|
||||||
|
}
|
||||||
|
Commands::Process(prc) => {
|
||||||
|
use crate::options::structs::bus::{
|
||||||
|
BusMessageContentType, BusMessageDirection, CLiCommand,
|
||||||
|
};
|
||||||
|
|
||||||
|
let proc_name = prc.process;
|
||||||
|
let req = BusMessage::Request(
|
||||||
|
BusMessageDirection::ToSupervisor,
|
||||||
|
BusMessageContentType::Cli,
|
||||||
|
Box::new(match prc.action {
|
||||||
|
ProcessAction::Start => InternalCli {
|
||||||
|
prc: proc_name,
|
||||||
|
cmd: CLiCommand::Start,
|
||||||
|
},
|
||||||
|
ProcessAction::Stop => InternalCli {
|
||||||
|
prc: proc_name,
|
||||||
|
cmd: CLiCommand::Stop,
|
||||||
|
},
|
||||||
|
ProcessAction::Restart => InternalCli {
|
||||||
|
prc: proc_name,
|
||||||
|
cmd: CLiCommand::Restart,
|
||||||
|
},
|
||||||
|
ProcessAction::Freeze => InternalCli {
|
||||||
|
prc: proc_name,
|
||||||
|
cmd: CLiCommand::Freeze,
|
||||||
|
},
|
||||||
|
ProcessAction::Unfreeze => InternalCli {
|
||||||
|
prc: proc_name,
|
||||||
|
cmd: CLiCommand::Unfreeze,
|
||||||
|
},
|
||||||
|
/* TODO: ALL CMDS */
|
||||||
|
_ => InternalCli {
|
||||||
|
prc: proc_name,
|
||||||
|
cmd: CLiCommand::Restart,
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
);
|
||||||
|
let mut bus = bus_reciever.lock().await;
|
||||||
|
bus_sender.send(req).await?;
|
||||||
|
tokio::time::sleep(std::time::Duration::from_millis(200)).await;
|
||||||
|
let resp = tokio::time::timeout(std::time::Duration::from_secs(5), async move {
|
||||||
|
loop {
|
||||||
|
if let Ok(cont) = bus.try_recv() {
|
||||||
|
return cont;
|
||||||
|
}
|
||||||
|
tokio::time::sleep(std::time::Duration::from_millis(500)).await;
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.await?;
|
||||||
|
|
||||||
|
if let BusMessage::Response(_, _, content) = resp {
|
||||||
|
let content: Box<dyn Any> = content;
|
||||||
|
if let Ok(resp) = content.downcast::<anyhow::Result<String>>() {
|
||||||
|
return Ok((*resp)?);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(anyhow::Error::msg(format!(
|
||||||
|
"Unknown type of response from the Supervisor"
|
||||||
|
)))
|
||||||
|
}
|
||||||
|
Commands::Status => Ok(String::from("Ok")),
|
||||||
|
Commands::Inspect(mode) => {
|
||||||
|
use crate::options::structs::bus::{BusMessageContentType, BusMessageDirection};
|
||||||
|
let mode = mode.mode;
|
||||||
|
if let Ok(()) = bus_sender
|
||||||
|
.send(BusMessage::Request(
|
||||||
|
BusMessageDirection::ToMetrics,
|
||||||
|
BusMessageContentType::MetricsModeTransfered,
|
||||||
|
Box::new(mode),
|
||||||
|
))
|
||||||
|
.await
|
||||||
|
{
|
||||||
|
let mut bus_reciever = bus_reciever.lock().await;
|
||||||
|
sleep(Duration::from_millis(300)).await;
|
||||||
|
let resp = tokio::time::timeout(std::time::Duration::from_secs(5), async move {
|
||||||
|
loop {
|
||||||
|
if let Ok(cont) = bus_reciever.try_recv() {
|
||||||
|
return cont;
|
||||||
|
}
|
||||||
|
tokio::time::sleep(std::time::Duration::from_millis(500)).await;
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.await?;
|
||||||
|
if let BusMessage::Response(_, _, content) = resp {
|
||||||
|
let content: Box<dyn Any> = content;
|
||||||
|
if let Ok(resp) = content.downcast::<anyhow::Result<String>>() {
|
||||||
|
return Ok((*resp)?);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return Err(anyhow::Error::msg(format!(
|
||||||
|
"Unknown type of response from CLI"
|
||||||
|
)));
|
||||||
|
}
|
||||||
|
Ok(String::from("Ok"))
|
||||||
|
}
|
||||||
|
_ => Ok(String::from("Ok")),
|
||||||
|
};
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -1,17 +1,387 @@
|
||||||
|
use super::preboot::PrebootParams;
|
||||||
use super::structs::*;
|
use super::structs::*;
|
||||||
|
use crate::utils::files::create_watcher;
|
||||||
|
use inotify::EventMask;
|
||||||
use log::{error, info, warn};
|
use log::{error, info, warn};
|
||||||
use redis::{Client, Connection};
|
use redis::{Client, Connection};
|
||||||
|
use std::fs::File;
|
||||||
use std::fs::OpenOptions;
|
use std::fs::OpenOptions;
|
||||||
use std::io::Write;
|
use std::io::Write;
|
||||||
use std::os::unix::process::CommandExt;
|
use std::os::unix::process::CommandExt;
|
||||||
use std::process::Command;
|
use std::process::Command;
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use std::{env, fs};
|
use std::{env, fs};
|
||||||
use super::preboot::PrebootParams;
|
use tokio::sync::{
|
||||||
use tokio::time::{Duration, sleep};
|
broadcast::Receiver as BroadcastReceiver,
|
||||||
|
broadcast::Sender as BroadcastSender,
|
||||||
|
oneshot,
|
||||||
|
oneshot::{Receiver as OneShotReciever, Sender as OneShotSender},
|
||||||
|
};
|
||||||
|
use tokio::time::{sleep, Duration};
|
||||||
|
|
||||||
// const CONFIG_PATH: &str = "settings.json";
|
// const CONFIG_PATH: &str = "settings.json";
|
||||||
|
|
||||||
|
pub mod v2 {
|
||||||
|
use crate::utils::get_container_id;
|
||||||
|
use std::path::PathBuf;
|
||||||
|
|
||||||
|
use super::*;
|
||||||
|
|
||||||
|
pub async fn init_config_mechanism(
|
||||||
|
// to handle cli config changes
|
||||||
|
cli_oneshot: OneShotReciever<Processes>,
|
||||||
|
// to share local config with PRCS, CLI_PIPELINE and CONFIG modules
|
||||||
|
brd_tx: BroadcastSender<Processes>,
|
||||||
|
// preboot params (args)
|
||||||
|
params: Arc<PrebootParams>, /*...*/
|
||||||
|
) {
|
||||||
|
// channel for pubsub to handle local config pulling
|
||||||
|
let local_config_brd_reciever = brd_tx.subscribe();
|
||||||
|
// channel between pub-sub mech and local config mech
|
||||||
|
let (tx_pb_lc, rx_pb_lc) = oneshot::channel::<bool>();
|
||||||
|
// channel between cli mech and local config mech
|
||||||
|
let (tx_cli_lc, rx_cli_lc) = oneshot::channel::<bool>();
|
||||||
|
|
||||||
|
// dbg!("before lc");
|
||||||
|
let params_clone = params.clone();
|
||||||
|
let for_lc_path = params.clone();
|
||||||
|
let lc_path = for_lc_path.config.to_str().unwrap_or("settings.json");
|
||||||
|
|
||||||
|
// future to init work with local config
|
||||||
|
let lc_future = tokio::spawn(
|
||||||
|
// let params = params.clone();
|
||||||
|
local_config_reciever(params_clone, rx_pb_lc, rx_cli_lc, Arc::new(brd_tx)),
|
||||||
|
);
|
||||||
|
// dbg!("before pb");
|
||||||
|
// future to init work with pub sub mechanism
|
||||||
|
let pubsub_future = tokio::spawn(pubsub_config_reciever(
|
||||||
|
tx_pb_lc,
|
||||||
|
params.clone(),
|
||||||
|
local_config_brd_reciever,
|
||||||
|
));
|
||||||
|
|
||||||
|
// dbg!("before cli");
|
||||||
|
// future to catch new configs from cli pipeline
|
||||||
|
let cli_future = tokio::spawn(from_cli_config_reciever(cli_oneshot, tx_cli_lc));
|
||||||
|
// let _ = lc_future.await;
|
||||||
|
// dbg!("before select");
|
||||||
|
tokio::select! {
|
||||||
|
lc_result = lc_future => {
|
||||||
|
// dbg!("end of lc");
|
||||||
|
match lc_result {
|
||||||
|
Ok(res) => {
|
||||||
|
if res.is_ok() {
|
||||||
|
info!("Local config warding mechanism stopped, waiting for others ...");
|
||||||
|
sleep(Duration::from_millis(500)).await;
|
||||||
|
let _ = restart_main_thread();
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
error!("Local config warding mechanism crushed, restarting ...");
|
||||||
|
let _ = restart_main_thread();
|
||||||
|
}
|
||||||
|
},
|
||||||
|
Err(_) => {
|
||||||
|
error!("Local config warding mechanism crushed, restarting ...");
|
||||||
|
let _ = restart_main_thread();
|
||||||
|
},
|
||||||
|
}
|
||||||
|
},
|
||||||
|
pb_result = pubsub_future => {
|
||||||
|
match pb_result {
|
||||||
|
Ok(res) => {
|
||||||
|
if res.is_ok() {
|
||||||
|
info!("New config was saved locally, restarting ...");
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
error!("Pubsub mechanism crushed, restarting ...");
|
||||||
|
}
|
||||||
|
},
|
||||||
|
Err(_) => {
|
||||||
|
error!("Pubsub mechanism crushed, restarting ...");
|
||||||
|
},
|
||||||
|
}
|
||||||
|
let _ = restart_main_thread();
|
||||||
|
},
|
||||||
|
cli_config_option = cli_future => {
|
||||||
|
match cli_config_option {
|
||||||
|
Err(_) => error!("CLI pulling new config mechanism crushed, restarting ..."),
|
||||||
|
Ok(option_config) => {
|
||||||
|
match option_config {
|
||||||
|
None => error!("CLI pulling new config mechanism crushed, restarting ..."),
|
||||||
|
Some(config) => {
|
||||||
|
info!("New config was pulled from CLI, saving and restarting ...");
|
||||||
|
let _ = save_new_config(&config, lc_path);
|
||||||
|
},
|
||||||
|
}
|
||||||
|
},
|
||||||
|
}
|
||||||
|
let _ = restart_main_thread();
|
||||||
|
},
|
||||||
|
}
|
||||||
|
// dbg!("after select");
|
||||||
|
// TODO! futures + select! [OK]
|
||||||
|
// TODO! tests config
|
||||||
|
}
|
||||||
|
pub async fn get_redis_connection(params: &str) -> Option<Connection> {
|
||||||
|
for i in 1..=3 {
|
||||||
|
let redis_url = format!("redis://{}/", params);
|
||||||
|
info!(
|
||||||
|
"Trying to connect Redis pubsub `{}`. Attempt {}",
|
||||||
|
&redis_url, i
|
||||||
|
);
|
||||||
|
if let Ok(client) = Client::open(redis_url) {
|
||||||
|
if let Ok(conn) = client.get_connection() {
|
||||||
|
info!("Successfully opened Redis connection");
|
||||||
|
return Some(conn);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
error!("Error with subscribing Redis stream on update. Retrying in 5 secs...");
|
||||||
|
sleep(Duration::from_secs(5)).await;
|
||||||
|
}
|
||||||
|
None
|
||||||
|
}
|
||||||
|
|
||||||
|
// loop checking redis pubsub
|
||||||
|
async fn pubsub_config_reciever(
|
||||||
|
// to stop checking local config
|
||||||
|
local_conf_tx: OneShotSender<bool>,
|
||||||
|
params: Arc<PrebootParams>,
|
||||||
|
tx_brd_local: BroadcastReceiver<Processes>,
|
||||||
|
) -> anyhow::Result<()> {
|
||||||
|
/*...*/
|
||||||
|
// dbg!("start of pb");
|
||||||
|
let mut tx_brd_local = tx_brd_local;
|
||||||
|
let local_config = if !tx_brd_local.is_empty() {
|
||||||
|
tx_brd_local.recv().await?
|
||||||
|
} else {
|
||||||
|
// Processes::default()
|
||||||
|
let mut tick = tokio::time::interval(Duration::from_millis(500));
|
||||||
|
loop {
|
||||||
|
tick.tick().await;
|
||||||
|
break match tx_brd_local.recv().await {
|
||||||
|
Ok(conf) => conf,
|
||||||
|
Err(_) => continue,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
};
|
||||||
|
match get_redis_connection(¶ms.remote_server_url).await {
|
||||||
|
Some(mut conn) => {
|
||||||
|
let mut pub_sub = conn.as_pubsub();
|
||||||
|
let channel_name = get_container_id().unwrap_or(String::from("default"));
|
||||||
|
match pub_sub.subscribe(&channel_name) {
|
||||||
|
Err(er) => {
|
||||||
|
error!("Cannot subscribe pubsub channel due to {}", &er);
|
||||||
|
return Err(anyhow::Error::msg(format!(
|
||||||
|
"Cannot subscribe pubsub channel due to {}",
|
||||||
|
er
|
||||||
|
)));
|
||||||
|
}
|
||||||
|
Ok(_) => {
|
||||||
|
info!(
|
||||||
|
"Successfully subscribed to {} pubsub channel",
|
||||||
|
&channel_name
|
||||||
|
);
|
||||||
|
let _ = pub_sub.set_read_timeout(Some(Duration::from_secs(1)));
|
||||||
|
loop {
|
||||||
|
if let Ok(msg) = pub_sub.get_message() {
|
||||||
|
// dbg!("ok on get message");
|
||||||
|
let payload: Result<String, _> = msg.get_payload();
|
||||||
|
match payload {
|
||||||
|
Err(_) => error!("Cannot read new config from Redis channel. Check network or Redis configuration "),
|
||||||
|
Ok(payload) => {
|
||||||
|
if let Some(remote) = parse_extern_config(&payload) {
|
||||||
|
match config_comparing(&local_config, &remote) {
|
||||||
|
ConfigActuality::Local => {
|
||||||
|
warn!("Pulled new config from Redis channel, it's outdated. Ignoring ...");
|
||||||
|
},
|
||||||
|
ConfigActuality::Remote => {
|
||||||
|
info!("Pulled new actual config from Redis channel, version - `{}`", remote.date_of_creation);
|
||||||
|
// to stop watching local config file mechanism
|
||||||
|
let _ = local_conf_tx.send(true);
|
||||||
|
let config_path = params.config.to_str().unwrap_or("settings.json");
|
||||||
|
|
||||||
|
if save_new_config(&remote, &config_path).is_err() {
|
||||||
|
error!("Error with saving new config to {}. Stopping pubsub mechanism...", config_path);
|
||||||
|
return Err(anyhow::Error::msg(
|
||||||
|
format!("Error with saving new config to {}. Stopping pubsub mechanism...", config_path)
|
||||||
|
))
|
||||||
|
}
|
||||||
|
return Ok(());
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
warn!("Invalid config was pulled from Redis channel")
|
||||||
|
}
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// delay
|
||||||
|
tokio::task::yield_now().await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
None => {
|
||||||
|
sleep(Duration::from_secs(20)).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
//
|
||||||
|
async fn local_config_reciever(
|
||||||
|
params: Arc<PrebootParams>,
|
||||||
|
pubsub_oneshot: OneShotReciever<bool>,
|
||||||
|
cli_oneshot: OneShotReciever<bool>,
|
||||||
|
brd_tx: Arc<BroadcastSender<Processes>>,
|
||||||
|
/*...*/
|
||||||
|
) -> anyhow::Result<()> {
|
||||||
|
/*...*/
|
||||||
|
// shadowing as mut
|
||||||
|
let mut pubsub_oneshot = pubsub_oneshot;
|
||||||
|
let mut cli_oneshot = cli_oneshot;
|
||||||
|
// fill with default empty config, mut to change later
|
||||||
|
let mut _current_config = Processes::default();
|
||||||
|
// PathBuf to &str to work with local config path as slice
|
||||||
|
let local_config_path = params.config.to_str().unwrap_or("settings.json");
|
||||||
|
|
||||||
|
match load_processes(local_config_path) {
|
||||||
|
// if local exists
|
||||||
|
Some(conf) => {
|
||||||
|
info!("Local config `{}` was found.", &conf.date_of_creation);
|
||||||
|
_current_config = conf;
|
||||||
|
if let Err(er) = brd_tx.send(_current_config.clone()) {
|
||||||
|
error!("Cannot share local config with broadcast due to {}", er);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// if local is not exist
|
||||||
|
None => {
|
||||||
|
warn!("Local config wasn't found. Waiting for new ...");
|
||||||
|
return Err(anyhow::Error::msg("No local config"));
|
||||||
|
// ...
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// 100% local exists here
|
||||||
|
// create watcher on local config file
|
||||||
|
match create_watcher("", local_config_path) {
|
||||||
|
Ok(mut watcher) => {
|
||||||
|
loop {
|
||||||
|
let mut need_to_export_config = false;
|
||||||
|
// let mut need_to_recreate_watcher = false;
|
||||||
|
// return situations here
|
||||||
|
// 1) oneshot signal
|
||||||
|
// 2) if config was deleted -> recreate and fill with current config that is held here
|
||||||
|
// 3) if config was changed -> fill with current config that is held here
|
||||||
|
|
||||||
|
// catching signal from pubsub
|
||||||
|
// it's because pubsub mech pulled new valid and actual config and now it's time to ...
|
||||||
|
// ... overwrite local config file and restart main thread
|
||||||
|
if let Ok(_) = pubsub_oneshot.try_recv() {
|
||||||
|
sleep(Duration::from_secs(1)).await;
|
||||||
|
return Ok(());
|
||||||
|
}
|
||||||
|
|
||||||
|
// catching signal from cli
|
||||||
|
// it's because cli mech pulled new valid and actual config and now it's time to ...
|
||||||
|
// ... overwrite local config file and restart main thread (like in previous mechanism)
|
||||||
|
if let Ok(_) = cli_oneshot.try_recv() {
|
||||||
|
sleep(Duration::from_secs(1)).await;
|
||||||
|
return Ok(());
|
||||||
|
}
|
||||||
|
|
||||||
|
// ! IF NOXIS NEEDS TO RECREATE OR CHANGE LOCAL CONFIG NEED TO DRAIN THIS ACTIVITY ...
|
||||||
|
// ! ... FROM WATCHER"S BUFFER
|
||||||
|
|
||||||
|
// existing check
|
||||||
|
if !params.config.exists() {
|
||||||
|
warn!("Local config file was deleted or moved. Recreating new one with saved data ...");
|
||||||
|
need_to_export_config = true;
|
||||||
|
// need_to_recreate_watcher = true;
|
||||||
|
} else {
|
||||||
|
// changes check
|
||||||
|
let mut buffer = [0; 128];
|
||||||
|
let events = watcher.read_events(&mut buffer);
|
||||||
|
if events.is_ok() {
|
||||||
|
let events: Vec<EventMask> = events
|
||||||
|
.unwrap()
|
||||||
|
.map(|mask| mask.mask)
|
||||||
|
.filter(|mask| {
|
||||||
|
*mask == EventMask::MODIFY || *mask == EventMask::DELETE_SELF
|
||||||
|
})
|
||||||
|
.collect();
|
||||||
|
if !events.is_empty() {
|
||||||
|
warn!("Local config file was overwritten. Discarding changes ...");
|
||||||
|
need_to_export_config = true;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// exporting data
|
||||||
|
if need_to_export_config {
|
||||||
|
if let Err(er) =
|
||||||
|
export_saved_config_data_locally(¶ms.config, &_current_config).await
|
||||||
|
{
|
||||||
|
error!("Cannot save actual imported config due to {}", er);
|
||||||
|
} else {
|
||||||
|
// recreation watcher (draining activity buffer mechanism)
|
||||||
|
// if local config file was deleted and recreated
|
||||||
|
// if local config file was modified locally
|
||||||
|
match create_watcher("", local_config_path) {
|
||||||
|
Ok(new) => watcher = new,
|
||||||
|
Err(er) => error!("Cannot create new watcher due to {}", er),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
sleep(Duration::from_millis(300)).await;
|
||||||
|
// tokio::task::yield_now().await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(_) => {
|
||||||
|
error!("Cannot create watcher on local config file `{}`. Deinitializing warding local config mechanism...", local_config_path);
|
||||||
|
return Err(anyhow::Error::msg(
|
||||||
|
"Cannot create watcher on local config file",
|
||||||
|
));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// [:IN-TEST]
|
||||||
|
async fn from_cli_config_reciever(
|
||||||
|
cli_oneshot: OneShotReciever<Processes>,
|
||||||
|
to_local_tx: OneShotSender<bool>,
|
||||||
|
) -> Option<Processes> {
|
||||||
|
/* match awaits til channel*/
|
||||||
|
// dbg!("start of cli");
|
||||||
|
loop {
|
||||||
|
if !cli_oneshot.is_empty() {
|
||||||
|
match cli_oneshot.await {
|
||||||
|
Ok(config_from_cli) => {
|
||||||
|
info!(
|
||||||
|
"New actual config `{}` from CLI was pulled. Saving and restaring ...",
|
||||||
|
&config_from_cli.date_of_creation
|
||||||
|
);
|
||||||
|
let _ = to_local_tx.send(true);
|
||||||
|
return Some(config_from_cli);
|
||||||
|
}
|
||||||
|
_ => return None,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
sleep(Duration::from_millis(300)).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn export_saved_config_data_locally(
|
||||||
|
config_file_path: &PathBuf,
|
||||||
|
current_config: &Processes,
|
||||||
|
) -> anyhow::Result<()> {
|
||||||
|
let mut file = File::create(config_file_path)?;
|
||||||
|
file.write_all(serde_json::to_string_pretty(current_config)?.as_bytes())?;
|
||||||
|
Ok(())
|
||||||
|
// Ok(())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// # Fn `load_processes`
|
/// # Fn `load_processes`
|
||||||
/// ## for reading and parsing *local* storing config
|
/// ## for reading and parsing *local* storing config
|
||||||
///
|
///
|
||||||
|
|
@ -34,259 +404,6 @@ fn load_processes(json_filename: &str) -> Option<Processes> {
|
||||||
None
|
None
|
||||||
}
|
}
|
||||||
|
|
||||||
/// # Fn `get_actual_config`
|
|
||||||
/// ## for getting actual Monitor's config from local and remote storages
|
|
||||||
///
|
|
||||||
/// *input* : -
|
|
||||||
///
|
|
||||||
/// *output* : `None` on fatal error in mechanisms | `Some(conf)` on finish reading and parsing
|
|
||||||
///
|
|
||||||
/// *initiator* : main thread
|
|
||||||
///
|
|
||||||
/// *managing* : -
|
|
||||||
///
|
|
||||||
/// *depends on* : struct `Processes`
|
|
||||||
///
|
|
||||||
pub async fn get_actual_config(params : Arc<PrebootParams>) -> Option<Processes> {
|
|
||||||
// * if no local conf -> loop and +inf getting conf from redis server
|
|
||||||
// * if local conf -> once getting conf from redis server
|
|
||||||
let config_path = params.config.to_str().unwrap_or_else(|| {
|
|
||||||
error!("Invalid character in config file. Config path was set to default");
|
|
||||||
"settings.json"
|
|
||||||
});
|
|
||||||
info!("Configurating config module with params: no-remote-config={}, no-sub={}, local config path={:?}, remote server={}", params.no_remote_config, params.no_sub, params.config, params.remote_server_url);
|
|
||||||
match load_processes(config_path) {
|
|
||||||
Some(local_conf) => {
|
|
||||||
info!(
|
|
||||||
"Found local configuration, version - {}",
|
|
||||||
&local_conf.date_of_creation
|
|
||||||
);
|
|
||||||
if !params.no_remote_config {
|
|
||||||
if let Some(remote_conf) =
|
|
||||||
// TODO : rework with pubsub mech
|
|
||||||
once_get_remote_configuration(&format!("redis://{}/", ¶ms.remote_server_url))
|
|
||||||
{
|
|
||||||
return match config_comparing(&local_conf, &remote_conf) {
|
|
||||||
ConfigActuality::Local => {
|
|
||||||
info!("Local config is actual");
|
|
||||||
Some(local_conf)
|
|
||||||
}
|
|
||||||
ConfigActuality::Remote => {
|
|
||||||
info!("Pulled config is more actual. Saving changes!");
|
|
||||||
if save_new_config(&remote_conf, config_path).is_err() {
|
|
||||||
error!("Saving changes process failed due to unexpected error...")
|
|
||||||
}
|
|
||||||
Some(remote_conf)
|
|
||||||
}
|
|
||||||
};
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Some(local_conf)
|
|
||||||
}
|
|
||||||
None => {
|
|
||||||
warn!("No local valid conf was found. Trying to pull remote one...");
|
|
||||||
if !params.no_remote_config {
|
|
||||||
let mut conn = get_connection_watcher(&open_watcher(&format!("redis://{}/", ¶ms.remote_server_url)));
|
|
||||||
if let Some(conf) = get_remote_conf_watcher(&mut conn).await {
|
|
||||||
info!("Config {} was pulled from Redis-Server. Starting...", &conf.date_of_creation);
|
|
||||||
let _ = save_new_config(&conf, config_path);
|
|
||||||
return Some(conf);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
None
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Fn `get_remote_conf_watcher`
|
|
||||||
/// ## for infinitive pulling remote config
|
|
||||||
///
|
|
||||||
/// *input* : `&mut Connection`
|
|
||||||
///
|
|
||||||
/// *output* : `None` on fatal error | `Some(conf)` on succesfull pulling
|
|
||||||
///
|
|
||||||
/// *initiator* : fn `get_actual_config`
|
|
||||||
///
|
|
||||||
/// *managing* : mut ref `Connection` object
|
|
||||||
///
|
|
||||||
/// *depends on* : struct `Processes`
|
|
||||||
///
|
|
||||||
async fn get_remote_conf_watcher(conn : &mut Connection) -> Option<Processes> {
|
|
||||||
let mut conn = conn.as_pubsub();
|
|
||||||
let cont = crate::utils::get_container_id();
|
|
||||||
loop {
|
|
||||||
match cont {
|
|
||||||
Some(ref cont) => {
|
|
||||||
let cont = cont.trim();
|
|
||||||
if conn.subscribe(cont).is_err() {
|
|
||||||
// todo : delay
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
match conn.get_message() {
|
|
||||||
Ok(msg) => {
|
|
||||||
let msg: Result<String, redis::RedisError> = msg.get_payload();
|
|
||||||
if let Ok(payload) = msg {
|
|
||||||
if let Some(remote) = parse_extern_config(&payload) {
|
|
||||||
return Some(remote)
|
|
||||||
}
|
|
||||||
else {
|
|
||||||
error!("Pulled invalid config, cannot start. Waiting for remote conf...");
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
error!("Cannot get Redis message payload. Waiting for remote conf...");
|
|
||||||
}
|
|
||||||
// todo : delay
|
|
||||||
continue;
|
|
||||||
},
|
|
||||||
Err(_) => {
|
|
||||||
// todo : delay
|
|
||||||
continue;
|
|
||||||
},
|
|
||||||
}
|
|
||||||
},
|
|
||||||
None => {
|
|
||||||
error!("Cannot get container id. Returning");
|
|
||||||
break
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
None
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Fn `get_remote_conf_watcher`
|
|
||||||
/// ## for trying to pull remote config
|
|
||||||
///
|
|
||||||
/// > only for situation when local isn't None (no need to fck redis server)
|
|
||||||
///
|
|
||||||
/// *input* : `&str`
|
|
||||||
///
|
|
||||||
/// *output* : `None` on empty pubsub or error | `Some(conf)` on succesfull pulling
|
|
||||||
///
|
|
||||||
/// *initiator* : fn `get_actual_config`
|
|
||||||
///
|
|
||||||
/// *managing* : &str of Redis Server credentials
|
|
||||||
///
|
|
||||||
/// *depends on* : struct `Processes`
|
|
||||||
///
|
|
||||||
fn once_get_remote_configuration(serv_info: &str) -> Option<Processes> {
|
|
||||||
let cont = crate::utils::get_container_id();
|
|
||||||
match Client::open(serv_info) {
|
|
||||||
Ok(client) => {
|
|
||||||
match client.get_connection() {
|
|
||||||
Ok(mut conn) => {
|
|
||||||
let mut conn = conn.as_pubsub();
|
|
||||||
match conn.subscribe(cont) {
|
|
||||||
Ok(_) => {
|
|
||||||
if conn.set_read_timeout(Some(Duration::from_millis(100))).is_err() {
|
|
||||||
error!("Cannot set reading pubsub timeout and pull remote config");
|
|
||||||
return None;
|
|
||||||
}
|
|
||||||
match conn.get_message() {
|
|
||||||
Ok(msg) => {
|
|
||||||
info!("Pulled config from Redis Server");
|
|
||||||
let get_payload: Result<String, redis::RedisError> = msg.get_payload();
|
|
||||||
match get_payload {
|
|
||||||
Ok(payload) => {
|
|
||||||
let remote = parse_extern_config(&payload);
|
|
||||||
if remote.is_none() {
|
|
||||||
error!("Pulled config is invalid. Check it in Redis Server");
|
|
||||||
}
|
|
||||||
remote
|
|
||||||
},
|
|
||||||
Err(_) => {
|
|
||||||
error!("Cannot extract payload from new message. Check Redis Server state");
|
|
||||||
None
|
|
||||||
},
|
|
||||||
}
|
|
||||||
},
|
|
||||||
Err(_) => {
|
|
||||||
None
|
|
||||||
},
|
|
||||||
}
|
|
||||||
},
|
|
||||||
Err(_) => {
|
|
||||||
error!("Redis subscription process failed. Check Redis configuration!");
|
|
||||||
None
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Err(_) => {
|
|
||||||
error!("Redis connection attempt is failed. Check Redis configuration!");
|
|
||||||
None
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Err(_) => {
|
|
||||||
error!("Redis-Client opening attempt is failed. Check network configuration!");
|
|
||||||
None
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// ! watchers
|
|
||||||
|
|
||||||
/// # Fn `open_watcher`
|
|
||||||
/// ## for infinitive opening Redis client
|
|
||||||
///
|
|
||||||
/// > only for situation when local isn't None (no need to fck redis server)
|
|
||||||
///
|
|
||||||
/// *input* : `Option<Processes>`
|
|
||||||
///
|
|
||||||
/// *output* : redis::Client on successful opening client
|
|
||||||
///
|
|
||||||
/// *initiator* : fn `get_actual_config`
|
|
||||||
///
|
|
||||||
/// *managing* : &str of Redis Server credentials
|
|
||||||
///
|
|
||||||
/// *depends on* : struct `redis::Client`
|
|
||||||
///
|
|
||||||
fn open_watcher(serv_info: &str) -> Client {
|
|
||||||
loop {
|
|
||||||
match Client::open(serv_info) {
|
|
||||||
Ok(redis) => {
|
|
||||||
info!("Successfully opened Redis-Client");
|
|
||||||
return redis;
|
|
||||||
}
|
|
||||||
Err(_) => {
|
|
||||||
error!("Redis-Client opening attempt is failed. Check network configuration! Retrying...");
|
|
||||||
std::thread::sleep(Duration::from_secs(4));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Fn `get_connection_watcher`
|
|
||||||
/// ## for infinitive establishing Redis connection on existing client
|
|
||||||
///
|
|
||||||
/// > only for situation when local isn't None (no need to fck redis server)
|
|
||||||
///
|
|
||||||
/// *input* : `&Client`
|
|
||||||
///
|
|
||||||
/// *output* : `Connection`
|
|
||||||
///
|
|
||||||
/// *initiator* : fn `get_actual_config`
|
|
||||||
///
|
|
||||||
/// *managing* : &Client for opening connection
|
|
||||||
///
|
|
||||||
/// *depends on* : struct `redis::Connection`
|
|
||||||
///
|
|
||||||
fn get_connection_watcher(client: &Client) -> Connection {
|
|
||||||
loop {
|
|
||||||
match client.get_connection() {
|
|
||||||
Ok(conn) => {
|
|
||||||
info!("Successfully got Redis connection object");
|
|
||||||
return conn;
|
|
||||||
}
|
|
||||||
Err(_) => {
|
|
||||||
error!(
|
|
||||||
"Redis connection attempt is failed. Check Redis configuration! Retrying..."
|
|
||||||
);
|
|
||||||
std::thread::sleep(Duration::from_secs(4));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Fn `restart_main_thread`
|
/// # Fn `restart_main_thread`
|
||||||
/// ## for restart monitor with new config
|
/// ## for restart monitor with new config
|
||||||
///
|
///
|
||||||
|
|
@ -302,87 +419,10 @@ fn get_connection_watcher(client: &Client) -> Connection {
|
||||||
///
|
///
|
||||||
fn restart_main_thread() -> std::io::Result<()> {
|
fn restart_main_thread() -> std::io::Result<()> {
|
||||||
let current_exe = env::current_exe()?;
|
let current_exe = env::current_exe()?;
|
||||||
Command::new(current_exe).exec();
|
let _ = Command::new(current_exe).exec();
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
/// # Fn `subscribe_config_stream`
|
|
||||||
/// ## for subscribe on changes, pulling to Redis pubsub to get more actual config
|
|
||||||
///
|
|
||||||
/// *input* : `Arc<Processes>`
|
|
||||||
///
|
|
||||||
/// *output* : `Ok(())` on end of work | `Err(er)` on error with subscribing mechanism
|
|
||||||
///
|
|
||||||
/// *initiator* : fn `subscribe_config_stream`
|
|
||||||
///
|
|
||||||
/// *managing* : `Arc<Processes>` to compare old config with new pulled
|
|
||||||
///
|
|
||||||
/// *depends on* : `Processes`
|
|
||||||
///
|
|
||||||
pub async fn subscribe_config_stream(actual_prcs: Arc<Processes>, params: Arc<PrebootParams>) -> Result<(), CustomError> {
|
|
||||||
let config_path = params.config.to_str().unwrap_or_else(|| "settings.json");
|
|
||||||
|
|
||||||
if params.no_sub || params.no_remote_config {
|
|
||||||
return Err(CustomError::Fatal);
|
|
||||||
}
|
|
||||||
if let Ok(client) = Client::open(format!("redis://{}/", ¶ms.remote_server_url)) {
|
|
||||||
if let Ok(mut conn) = client.get_connection() {
|
|
||||||
match crate::utils::get_container_id() {
|
|
||||||
Some(channel_name) => {
|
|
||||||
let channel_name = channel_name.trim();
|
|
||||||
let mut pubsub = conn.as_pubsub();
|
|
||||||
if pubsub.subscribe(&channel_name).is_ok() {
|
|
||||||
info!("Runner subscribed on config update publishing in channel {}", &channel_name);
|
|
||||||
loop {
|
|
||||||
if let Ok(msg) = pubsub.get_message() {
|
|
||||||
let get_remote_config: Result<String, redis::RedisError> = msg.get_payload();
|
|
||||||
match get_remote_config {
|
|
||||||
Ok(payload) => {
|
|
||||||
if let Some(remote_config) = parse_extern_config(&payload) {
|
|
||||||
match config_comparing(&actual_prcs, &remote_config) {
|
|
||||||
ConfigActuality::Remote => {
|
|
||||||
warn!("Pulled config is actual. Saving and restarting...");
|
|
||||||
if save_new_config(&remote_config, config_path).is_err() {
|
|
||||||
error!("Error with saving new config to {}. Stopping sub mechanism...", config_path);
|
|
||||||
return Err(CustomError::Fatal);
|
|
||||||
}
|
|
||||||
if restart_main_thread().is_err() {
|
|
||||||
error!("Error with restarting Runner. Stopping sub mechanism...");
|
|
||||||
return Err(CustomError::Fatal);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
_ => {
|
|
||||||
warn!("Pulled new config. Current config is more actual ...");
|
|
||||||
continue
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
else {
|
|
||||||
error!("Invalid conig was pulled");
|
|
||||||
}
|
|
||||||
},
|
|
||||||
Err(_) => {
|
|
||||||
error!("Cannot extract new config from message");
|
|
||||||
break;
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
sleep(Duration::from_secs(30)).await;
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
error!("Cannot subscribe channel {}. Check Redis Server status", &channel_name);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
None => {
|
|
||||||
error!("Cannot get channel name");
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
error!("Error with subscribing Redis stream on update. Working only with selected config...");
|
|
||||||
Err(CustomError::Fatal)
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Fn `config_comparing`
|
/// # Fn `config_comparing`
|
||||||
/// ## for compare old and new configs
|
/// ## for compare old and new configs
|
||||||
///
|
///
|
||||||
|
|
@ -396,7 +436,10 @@ pub async fn subscribe_config_stream(actual_prcs: Arc<Processes>, params: Arc<Pr
|
||||||
///
|
///
|
||||||
/// *depends on* : `Processes`, `ConfigActuality`
|
/// *depends on* : `Processes`, `ConfigActuality`
|
||||||
///
|
///
|
||||||
fn config_comparing(local: &Processes, remote: &Processes) -> ConfigActuality {
|
pub fn config_comparing(local: &Processes, remote: &Processes) -> ConfigActuality {
|
||||||
|
if local.is_default() {
|
||||||
|
return ConfigActuality::Remote;
|
||||||
|
}
|
||||||
let local_date: u64 = local.date_of_creation.parse().unwrap();
|
let local_date: u64 = local.date_of_creation.parse().unwrap();
|
||||||
let remote_date: u64 = remote.date_of_creation.parse().unwrap();
|
let remote_date: u64 = remote.date_of_creation.parse().unwrap();
|
||||||
|
|
||||||
|
|
@ -406,14 +449,6 @@ fn config_comparing(local: &Processes, remote: &Processes) -> ConfigActuality {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// ! TEMPORARILY DEPRECATED !
|
|
||||||
// fn native_date_from_millis(mls: &str) -> Option<chrono::DateTime<Utc>> {
|
|
||||||
// match mls.parse::<i64>(){
|
|
||||||
// Ok(val) => return chrono::DateTime::from_timestamp_millis(val),
|
|
||||||
// Err(_) => return None,
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
|
|
||||||
/// # Fn `save_new_config`
|
/// # Fn `save_new_config`
|
||||||
/// ## mechanism for saving new config in local storage
|
/// ## mechanism for saving new config in local storage
|
||||||
///
|
///
|
||||||
|
|
@ -492,28 +527,20 @@ mod config_unittests {
|
||||||
// old one (kinda local)
|
// old one (kinda local)
|
||||||
let a = Processes {
|
let a = Processes {
|
||||||
date_of_creation: String::from("1"),
|
date_of_creation: String::from("1"),
|
||||||
config_server: String::new(),
|
|
||||||
processes: vec![],
|
processes: vec![],
|
||||||
};
|
};
|
||||||
// new one (kinda remote)
|
// new one (kinda remote)
|
||||||
let b = Processes {
|
let b = Processes {
|
||||||
date_of_creation: String::from("2"),
|
date_of_creation: String::from("2"),
|
||||||
config_server: String::new(),
|
|
||||||
processes: vec![],
|
processes: vec![],
|
||||||
};
|
};
|
||||||
|
|
||||||
assert_eq!(config_comparing(&a, &b), ConfigActuality::Remote);
|
assert_eq!(config_comparing(&a, &b), ConfigActuality::Remote);
|
||||||
}
|
}
|
||||||
// TODO : strange output
|
|
||||||
// #[test]
|
|
||||||
// fn get_actual_config_mechanism() {
|
|
||||||
// assert!(get_actual_config().is_some())
|
|
||||||
// }
|
|
||||||
#[test]
|
#[test]
|
||||||
fn save_config() {
|
fn save_config() {
|
||||||
let a = Processes {
|
let a = Processes {
|
||||||
date_of_creation: String::from("1"),
|
date_of_creation: String::from("1"),
|
||||||
config_server: String::new(),
|
|
||||||
processes: vec![],
|
processes: vec![],
|
||||||
};
|
};
|
||||||
assert!(save_new_config(&a, "tests/examples/save-conf.json").is_ok());
|
assert!(save_new_config(&a, "tests/examples/save-conf.json").is_ok());
|
||||||
|
|
@ -523,7 +550,6 @@ mod config_unittests {
|
||||||
fn save_to_zero_file() {
|
fn save_to_zero_file() {
|
||||||
let a = Processes {
|
let a = Processes {
|
||||||
date_of_creation: String::from("1"),
|
date_of_creation: String::from("1"),
|
||||||
config_server: String::new(),
|
|
||||||
processes: vec![],
|
processes: vec![],
|
||||||
};
|
};
|
||||||
assert!(save_new_config(&a, "tests/examples/none.json").is_ok());
|
assert!(save_new_config(&a, "tests/examples/none.json").is_ok());
|
||||||
|
|
|
||||||
|
|
@ -49,7 +49,7 @@ pub fn setup_logger() -> Result<(), crate::options::structs::CustomError> {
|
||||||
record.args(),
|
record.args(),
|
||||||
)
|
)
|
||||||
})
|
})
|
||||||
.filter(None, LevelFilter::Info)
|
.filter(None, LevelFilter::from_env())
|
||||||
.target(env_logger::Target::Stdout)
|
.target(env_logger::Target::Stdout)
|
||||||
// temporary deprecated
|
// temporary deprecated
|
||||||
// .target(env_logger::Target::Pipe(log_target))
|
// .target(env_logger::Target::Pipe(log_target))
|
||||||
|
|
@ -58,6 +58,27 @@ pub fn setup_logger() -> Result<(), crate::options::structs::CustomError> {
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
trait FromEnv {
|
||||||
|
fn from_env() -> LevelFilter;
|
||||||
|
}
|
||||||
|
|
||||||
|
impl FromEnv for LevelFilter {
|
||||||
|
fn from_env() -> LevelFilter {
|
||||||
|
return match std::env::var("NOXIS_MAX_LOG_LEVEL") {
|
||||||
|
Ok(var) => match var.to_ascii_lowercase().trim().as_ref() {
|
||||||
|
"trace" => LevelFilter::Trace,
|
||||||
|
"debug" => LevelFilter::Debug,
|
||||||
|
"info" => LevelFilter::Info,
|
||||||
|
"error" => LevelFilter::Error,
|
||||||
|
"warn" => LevelFilter::Warn,
|
||||||
|
"off" => LevelFilter::Off,
|
||||||
|
_ => LevelFilter::Info,
|
||||||
|
},
|
||||||
|
Err(_) => LevelFilter::Info,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod logger_tests {
|
mod logger_tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|
|
||||||
|
|
@ -1,92 +1,12 @@
|
||||||
// module to handle pre-boot params of the monitor
|
//!
|
||||||
|
//! Module to handle `pre-boot params` of the monitor (calling also as `settings`)
|
||||||
|
//!
|
||||||
#[allow(unused_imports)]
|
#[allow(unused_imports)]
|
||||||
use anyhow::{Result, Ok, Error};
|
use anyhow::{Error, Result};
|
||||||
use clap::Parser;
|
|
||||||
use std::path::PathBuf;
|
|
||||||
use std::env::var;
|
|
||||||
use dotenv::dotenv;
|
use dotenv::dotenv;
|
||||||
|
use log::warn;
|
||||||
const SOCKET_PATH: &str = "/var/run/enode/hostagent.sock";
|
use std::env::var;
|
||||||
|
use std::path::PathBuf;
|
||||||
///
|
|
||||||
enum EnvVars {
|
|
||||||
NoxisNoHagent,
|
|
||||||
NoxisNoLogs,
|
|
||||||
NoxisRefreshLogs,
|
|
||||||
NoxisNoRemoteConfig,
|
|
||||||
NoxisNoConfigSub,
|
|
||||||
NoxisSocketPath,
|
|
||||||
NoxisLogTo,
|
|
||||||
NoxisRemoteServerUrl,
|
|
||||||
NoxisConfig,
|
|
||||||
NoxisMetrics,
|
|
||||||
}
|
|
||||||
|
|
||||||
///
|
|
||||||
impl std::fmt::Display for EnvVars {
|
|
||||||
fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result {
|
|
||||||
match self {
|
|
||||||
EnvVars::NoxisNoHagent => write!(f, "NOXIS_NO_HAGENT"),
|
|
||||||
EnvVars::NoxisNoLogs => write!(f, "NOXIS_NO_LOGS"),
|
|
||||||
EnvVars::NoxisRefreshLogs => write!(f, "NOXIS_REFRESH_LOGS"),
|
|
||||||
EnvVars::NoxisNoRemoteConfig => write!(f, "NOXIS_NO_REMOTE_CONFIG"),
|
|
||||||
EnvVars::NoxisNoConfigSub => write!(f, "NOXIS_NO_CONFIG_SUB"),
|
|
||||||
EnvVars::NoxisSocketPath => write!(f, "NOXIS_SOCKET_PATH"),
|
|
||||||
EnvVars::NoxisLogTo => write!(f, "NOXIS_LOG_TO"),
|
|
||||||
EnvVars::NoxisRemoteServerUrl => write!(f, "NOXIS_REMOTE_SERVER_URL"),
|
|
||||||
EnvVars::NoxisConfig => write!(f, "NOXIS_CONFIG"),
|
|
||||||
EnvVars::NoxisMetrics => write!(f, "NOXIS_METRICS"),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
///
|
|
||||||
impl<'a> EnvVars {
|
|
||||||
// Default trait func is not satisfying this issue
|
|
||||||
fn default(self) -> &'a str {
|
|
||||||
match self {
|
|
||||||
EnvVars::NoxisNoHagent => "false",
|
|
||||||
EnvVars::NoxisNoLogs => "false",
|
|
||||||
EnvVars::NoxisRefreshLogs => "false",
|
|
||||||
EnvVars::NoxisNoRemoteConfig => "false",
|
|
||||||
EnvVars::NoxisNoConfigSub => "false",
|
|
||||||
EnvVars::NoxisSocketPath => "/var/run/enode/hostagent.sock",
|
|
||||||
EnvVars::NoxisLogTo => "./",
|
|
||||||
EnvVars::NoxisRemoteServerUrl => "localhost",
|
|
||||||
EnvVars::NoxisConfig => "./settings.json",
|
|
||||||
EnvVars::NoxisMetrics => "full",
|
|
||||||
}
|
|
||||||
}
|
|
||||||
fn process_env_var(self, preboot_value: &str) {
|
|
||||||
// let default = self.default();
|
|
||||||
match var(self.to_string()) {
|
|
||||||
std::result::Result::Ok(val) => {
|
|
||||||
if val != preboot_value {
|
|
||||||
std::env::set_var(self.to_string(), self.default());
|
|
||||||
}
|
|
||||||
},
|
|
||||||
Err(_) => {
|
|
||||||
std::env::set_var(self.to_string(), preboot_value);
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
pub fn setup(preboot: &PrebootParams) {
|
|
||||||
// setup default if not exists
|
|
||||||
// check values and save preboot states in env vars if not equal
|
|
||||||
|
|
||||||
Self::NoxisNoHagent.process_env_var(&preboot.no_hostagent.to_string());
|
|
||||||
Self::NoxisNoLogs.process_env_var(&preboot.no_logs.to_string());
|
|
||||||
Self::NoxisRefreshLogs.process_env_var(&preboot.refresh_logs.to_string());
|
|
||||||
Self::NoxisNoRemoteConfig.process_env_var(&preboot.no_remote_config.to_string());
|
|
||||||
Self::NoxisNoConfigSub.process_env_var(&preboot.no_sub.to_string());
|
|
||||||
Self::NoxisSocketPath.process_env_var(preboot.socket_path.to_str().unwrap());
|
|
||||||
Self::NoxisLogTo.process_env_var(preboot.log_to.to_str().unwrap());
|
|
||||||
Self::NoxisRemoteServerUrl.process_env_var(&preboot.remote_server_url);
|
|
||||||
Self::NoxisConfig.process_env_var(preboot.config.to_str().unwrap());
|
|
||||||
Self::NoxisMetrics.process_env_var(&preboot.metrics.to_string());
|
|
||||||
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Enum `MetricsPrebootParams`
|
/// # Enum `MetricsPrebootParams`
|
||||||
/// ## for setting up metrics mode as preboot param from command prompt
|
/// ## for setting up metrics mode as preboot param from command prompt
|
||||||
|
|
@ -100,7 +20,7 @@ impl<'a> EnvVars {
|
||||||
/// noxis-rs ... --metrics none
|
/// noxis-rs ... --metrics none
|
||||||
/// ```
|
/// ```
|
||||||
///
|
///
|
||||||
#[derive(clap::ValueEnum, Debug, Clone)]
|
#[derive(Debug, Clone, serde::Serialize, serde::Deserialize)]
|
||||||
pub enum MetricsPrebootParams {
|
pub enum MetricsPrebootParams {
|
||||||
Full,
|
Full,
|
||||||
System,
|
System,
|
||||||
|
|
@ -109,6 +29,19 @@ pub enum MetricsPrebootParams {
|
||||||
None,
|
None,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl MetricsPrebootParams {
|
||||||
|
fn from_env(var: &str) -> Self {
|
||||||
|
match var.trim().to_lowercase().as_str() {
|
||||||
|
"full" => Self::Full,
|
||||||
|
"system" => Self::System,
|
||||||
|
"processes" => Self::Processes,
|
||||||
|
"net" => Self::Net,
|
||||||
|
"none" => Self::None,
|
||||||
|
_ => Self::Full,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// # `std::fmt::Display` implementation for `MetricsPrebootParams`
|
/// # `std::fmt::Display` implementation for `MetricsPrebootParams`
|
||||||
/// ## to enable parsing object to String
|
/// ## to enable parsing object to String
|
||||||
impl std::fmt::Display for MetricsPrebootParams {
|
impl std::fmt::Display for MetricsPrebootParams {
|
||||||
|
|
@ -123,260 +56,314 @@ impl std::fmt::Display for MetricsPrebootParams {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// # struct `PrebootParams`
|
/// struct to handle Noxis settings (that were set as `.env` or global env vars)
|
||||||
/// ## to parse and set up all modes as preboot params from command prompt
|
|
||||||
///
|
///
|
||||||
/// ### args :
|
/// to parse and set up all modes as preboot params from `.env` or using `export $VAR $VAL` command
|
||||||
///
|
///
|
||||||
/// `--no-hagent` - to disable hagent work module and set up work mode as autonomous
|
/// # Settings :
|
||||||
/// ### usage :
|
///
|
||||||
|
/// All settings are divided by `actions` and `values`.
|
||||||
|
/// 1. `Actions` - true or false on var exist check
|
||||||
|
/// > `Actions` is not about values, it can handle any data
|
||||||
|
///
|
||||||
|
/// 2. `Values` - classic string-driven environment variables
|
||||||
|
///
|
||||||
|
/// # `Actions` vars
|
||||||
|
///
|
||||||
|
/// 1. `NOXIS_NO_HAGENT` - to disable hagent work module and set up work mode as autonomous
|
||||||
|
///
|
||||||
|
/// **usage** :
|
||||||
|
/// ``` toml
|
||||||
|
/// ...
|
||||||
|
/// NOXIS_NO_HAGENT = "random-text"
|
||||||
|
/// ...
|
||||||
|
/// ```
|
||||||
|
/// or
|
||||||
/// ``` bash
|
/// ``` bash
|
||||||
/// noxis-rs ... --no-hagent ...
|
/// export NOXIS_NO_HAGENT "random-text"
|
||||||
/// ```
|
/// ```
|
||||||
///
|
///
|
||||||
///
|
///
|
||||||
/// `--no-logs` - to disable logging at all
|
/// 2. `NOXIS_NO_LOGS` - to disable logging at all
|
||||||
/// ### usage :
|
/// **usage** :
|
||||||
|
/// ``` toml
|
||||||
|
/// ...
|
||||||
|
/// NOXIS_NO_LOGS = "random-text"
|
||||||
|
/// ...
|
||||||
|
/// ```
|
||||||
|
/// or
|
||||||
/// ``` bash
|
/// ``` bash
|
||||||
/// noxis-rs ... --no-logs ...
|
/// export NOXIS_NO_LOGS "random-text"
|
||||||
/// ```
|
/// ```
|
||||||
///
|
///
|
||||||
/// `--refresh-logs` - to truncate logs directory
|
/// 3. `NOXIS_REFRESH_LOGS` - to truncate logs directory
|
||||||
|
///
|
||||||
/// ### usage :
|
/// ### usage :
|
||||||
|
/// ``` toml
|
||||||
|
/// ...
|
||||||
|
/// NOXIS_REFRESH_LOGS = "random-text"
|
||||||
|
/// ...
|
||||||
|
/// ```
|
||||||
|
/// or
|
||||||
/// ``` bash
|
/// ``` bash
|
||||||
/// noxis-rs ... --refresh-logs ...
|
/// export NOXIS_REFRESH_LOGS "random-text"
|
||||||
/// ```
|
/// ```
|
||||||
///
|
///
|
||||||
/// `--no-remote-config` - to disable work with Redis as config producer
|
/// 4. `NOXIS_NO_SUB` - to disable Redis subscribtion mechanism
|
||||||
|
///
|
||||||
/// ### usage :
|
/// ### usage :
|
||||||
|
/// ``` toml
|
||||||
|
/// NOXIS_NO_SUB = "random-text"
|
||||||
|
/// ```
|
||||||
|
/// or
|
||||||
/// ``` bash
|
/// ``` bash
|
||||||
/// noxis-rs ... --no-remote-config ...
|
/// export NOXIS_NO_SUB "random-text"
|
||||||
/// ```
|
/// ```
|
||||||
///
|
///
|
||||||
/// `--no-sub` - to disable Redis subscribtion mechanism
|
/// # `Values` vars
|
||||||
|
///
|
||||||
|
/// 1. `NOXIS_HAGENT_SOCKET_PATH` - to set Unix Domain Socket file's directory
|
||||||
|
///
|
||||||
/// ### usage :
|
/// ### usage :
|
||||||
|
/// ``` toml
|
||||||
|
/// NOXIS_HAGENT_SOCKET_PATH = "/var/run/example/hostagent.sock"
|
||||||
|
/// ```
|
||||||
|
/// or
|
||||||
/// ``` bash
|
/// ``` bash
|
||||||
/// noxis-rs ... --no-sub ...
|
/// export NOXIS_HAGENT_SOCKET_PATH "/var/run/example/hostagent.sock"
|
||||||
/// ```
|
/// ```
|
||||||
///
|
///
|
||||||
/// `--socket-path` - to set Unix Domain Socket file's directory
|
/// 2 `NOXIS_LOG_TO` - to set directory for logs
|
||||||
|
///
|
||||||
/// ### usage :
|
/// ### usage :
|
||||||
|
/// ``` toml
|
||||||
|
/// NOXIS_LOG_TO = "/var/log/noxis/noxis.log"
|
||||||
|
/// ```
|
||||||
|
/// or
|
||||||
/// ``` bash
|
/// ``` bash
|
||||||
/// noxis-rs ... --socket-path /var/run/enode/hostagent.sock ...
|
/// export NOXIS_LOG_TO "/var/log/noxis/noxis.log"
|
||||||
/// ```
|
/// ```
|
||||||
///
|
///
|
||||||
/// `--log-to` - to set directory for logs
|
/// 3. `NOXIS_REMOTE_SERVER_URL` - to set Redis Server
|
||||||
|
///
|
||||||
/// ### usage :
|
/// ### usage :
|
||||||
|
/// ``` toml
|
||||||
|
/// NOXIS_REMOTE_SERVER_URL = "ip.ip.ip.ip:port"
|
||||||
|
/// ```
|
||||||
|
/// or
|
||||||
/// ``` bash
|
/// ``` bash
|
||||||
/// noxis-rs ... --log-to /dir/to/logs/ ...
|
/// export NOXIS_REMOTE_SERVER_URL "ip.ip.ip.ip:port"
|
||||||
/// ```
|
/// ```
|
||||||
///
|
///
|
||||||
/// `--remote-server-url` - to set Redis Server
|
/// 4. `NOXIS_CONFIG_PATH` - to set Noxis' config full path
|
||||||
|
///
|
||||||
/// ### usage :
|
/// ### usage :
|
||||||
|
/// ``` toml
|
||||||
|
/// NOXIS_CONFIG_PATH = "./settings.json"
|
||||||
|
/// ```
|
||||||
|
/// or
|
||||||
/// ``` bash
|
/// ``` bash
|
||||||
/// noxis-rs ... --remote-server-url 192.168.28.12 ...
|
/// export NOXIS_CONFIG_PATH "./settings.json"
|
||||||
/// ```
|
/// ```
|
||||||
///
|
///
|
||||||
/// `--config` - to set Noxis' config full path
|
/// 5. `NOXIS_METRICS_MODE` - to set metrics mode
|
||||||
|
///
|
||||||
/// ### usage :
|
/// ### usage :
|
||||||
|
/// ``` toml
|
||||||
|
/// NOXIS_METRICS_MODE = "full"
|
||||||
|
/// ```
|
||||||
|
/// or
|
||||||
/// ``` bash
|
/// ``` bash
|
||||||
/// noxis-rs ... --config /etc/enode/settings.json ...
|
/// export NOXIS_METRICS_MODE "full"
|
||||||
/// ```
|
/// ```
|
||||||
///
|
///
|
||||||
/// `--metrics` - to set metrics mode
|
#[derive(Debug, serde::Serialize, serde::Deserialize)]
|
||||||
/// ### usage :
|
|
||||||
/// ``` bash
|
|
||||||
/// noxis-rs ... --metrics full ...
|
|
||||||
/// ```
|
|
||||||
#[derive(Debug, Parser)]
|
|
||||||
pub struct PrebootParams {
|
pub struct PrebootParams {
|
||||||
// actions
|
// pub no_hostagent : bool,
|
||||||
#[arg(
|
|
||||||
long = "no-hagent",
|
|
||||||
action,
|
|
||||||
conflicts_with="socket_path",
|
|
||||||
help="To disable work with host-agent"
|
|
||||||
)]
|
|
||||||
pub no_hostagent : bool,
|
|
||||||
#[arg(
|
|
||||||
long = "no-logs",
|
|
||||||
action,
|
|
||||||
conflicts_with="log_to",
|
|
||||||
help="To disable logs"
|
|
||||||
)]
|
|
||||||
pub no_logs: bool,
|
pub no_logs: bool,
|
||||||
#[arg(
|
pub refresh_logs: bool,
|
||||||
long = "refresh-logs",
|
pub no_sub: bool,
|
||||||
action,
|
// pub socket_path : PathBuf,
|
||||||
conflicts_with="no_logs",
|
pub log_to: PathBuf,
|
||||||
help="To clear logs directory"
|
pub remote_server_url: String,
|
||||||
)]
|
pub config: PathBuf,
|
||||||
pub refresh_logs : bool,
|
|
||||||
#[arg(
|
|
||||||
long = "no-remote-config",
|
|
||||||
action,
|
|
||||||
help="To disable work with remote config server",
|
|
||||||
conflicts_with="no_sub")]
|
|
||||||
pub no_remote_config : bool,
|
|
||||||
#[arg(
|
|
||||||
long = "no-sub",
|
|
||||||
action,
|
|
||||||
help="To disable subscription mechanism",
|
|
||||||
conflicts_with="no_remote_config")]
|
|
||||||
pub no_sub : bool,
|
|
||||||
|
|
||||||
// params (socket_path, log_to, remote_server_url, config)
|
|
||||||
#[arg(
|
|
||||||
long = "socket-path",
|
|
||||||
default_value="/var/run/enode/hostagent.sock",
|
|
||||||
conflicts_with="no_hostagent",
|
|
||||||
help="To set .sock file's path used in communication with host-agent"
|
|
||||||
)]
|
|
||||||
pub socket_path : PathBuf,
|
|
||||||
#[arg(
|
|
||||||
long = "log-to",
|
|
||||||
default_value="./",
|
|
||||||
conflicts_with="no_logs",
|
|
||||||
help="To set a path to logs directory"
|
|
||||||
)]
|
|
||||||
pub log_to : PathBuf,
|
|
||||||
#[arg(
|
|
||||||
long = "remote-server-url",
|
|
||||||
default_value="localhost",
|
|
||||||
conflicts_with="no_remote_config",
|
|
||||||
help = "To set url of remote config server using in remote config pulling mechanism"
|
|
||||||
)]
|
|
||||||
pub remote_server_url : String,
|
|
||||||
#[arg(
|
|
||||||
long = "config",
|
|
||||||
short,
|
|
||||||
default_value="settings.json",
|
|
||||||
help="To set local config file path"
|
|
||||||
)]
|
|
||||||
pub config : PathBuf,
|
|
||||||
|
|
||||||
// value enum params (metrics)
|
|
||||||
#[arg(
|
|
||||||
long = "metrics",
|
|
||||||
short,
|
|
||||||
default_value_t=MetricsPrebootParams::Full,
|
|
||||||
help="To set metrics grubbing mode"
|
|
||||||
)]
|
|
||||||
pub metrics: MetricsPrebootParams,
|
pub metrics: MetricsPrebootParams,
|
||||||
|
pub self_socket: PathBuf,
|
||||||
|
pub backup_folder: PathBuf,
|
||||||
}
|
}
|
||||||
|
|
||||||
/// # implementation for `MetricsPrebootParams`
|
/// # implementation for `MetricsPrebootParams`
|
||||||
/// ## to enable validation mechanism
|
/// ## to enable validation mechanism
|
||||||
impl PrebootParams {
|
impl PrebootParams {
|
||||||
pub fn validate(mut self) -> Result<Self> {
|
pub fn validate() -> Self {
|
||||||
dotenv().ok();
|
dotenv().ok();
|
||||||
if !self.socket_path.exists() && !self.no_hostagent {
|
Self {
|
||||||
if self.socket_path.to_string_lossy() == SOCKET_PATH {
|
// bool
|
||||||
self.no_hostagent = true;
|
// no_hostagent : {
|
||||||
eprintln!("Warning: Socket-file wasn't found. Working without hostagent module...");
|
// match var("NOXIS_NO_HAGENT") {
|
||||||
|
// Ok(_) => true,
|
||||||
|
// Err(_) => false,
|
||||||
|
// }
|
||||||
|
// },
|
||||||
|
no_logs: {
|
||||||
|
match var("NOXIS_NO_LOGS") {
|
||||||
|
Ok(_) => true,
|
||||||
|
Err(_) => false,
|
||||||
|
}
|
||||||
|
},
|
||||||
|
refresh_logs: {
|
||||||
|
match var("NOXIS_REFRESH_LOGS") {
|
||||||
|
Ok(_) => true,
|
||||||
|
Err(_) => false,
|
||||||
|
}
|
||||||
|
},
|
||||||
|
no_sub: {
|
||||||
|
match var("NOXIS_NO_SUB") {
|
||||||
|
Ok(_) => true,
|
||||||
|
Err(_) => false,
|
||||||
|
}
|
||||||
|
},
|
||||||
|
// vals
|
||||||
|
// socket_path : {
|
||||||
|
// match var("NOXIS_HAGENT_SOCKET_PATH") {
|
||||||
|
// Ok(val) => PathBuf::from(val),
|
||||||
|
// Err(_) => PathBuf::from("/var/run/enode/hostagent.sock"),
|
||||||
|
// }
|
||||||
|
// },
|
||||||
|
log_to: {
|
||||||
|
match var("NOXIS_LOG_TO") {
|
||||||
|
Ok(val) => PathBuf::from(val),
|
||||||
|
Err(_) => PathBuf::from("./"),
|
||||||
|
}
|
||||||
|
},
|
||||||
|
remote_server_url: {
|
||||||
|
match var("NOXIS_REMOTE_SERVER_URL") {
|
||||||
|
Ok(val) => val,
|
||||||
|
Err(_) => String::from("localhost"),
|
||||||
|
}
|
||||||
|
},
|
||||||
|
config: {
|
||||||
|
match var("NOXIS_CONFIG_PATH") {
|
||||||
|
Ok(val) => PathBuf::from(val),
|
||||||
|
Err(_) => PathBuf::from("./settings.json"),
|
||||||
|
}
|
||||||
|
},
|
||||||
|
metrics: {
|
||||||
|
match var("NOXIS_METRICS_MODE") {
|
||||||
|
Ok(val) => MetricsPrebootParams::from_env(&val),
|
||||||
|
Err(_) => MetricsPrebootParams::Full,
|
||||||
|
}
|
||||||
|
},
|
||||||
|
self_socket: {
|
||||||
|
match var("NOXIS_SOCKET_PATH") {
|
||||||
|
Ok(val) => PathBuf::from(val),
|
||||||
|
Err(_) => {
|
||||||
|
let default = std::env::current_dir()
|
||||||
|
.expect("Crushed on getting current_dir path. Check fs state!");
|
||||||
|
warn!(
|
||||||
|
"$NOXIS_SOCKET_PATH wans't set. Default value - {}",
|
||||||
|
default.display()
|
||||||
|
);
|
||||||
|
PathBuf::from(default)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
|
backup_folder: {
|
||||||
|
match var("NOXIS_BACKUP_FOLDER") {
|
||||||
|
Ok(val) => {
|
||||||
|
let path = PathBuf::from(val);
|
||||||
|
if path.exists() && path.is_dir() {
|
||||||
|
path
|
||||||
} else {
|
} else {
|
||||||
eprintln!("Warning: Socket-file wasn't found or Noxis can't read it. Socket-file was set to default");
|
PathBuf::from(std::env::current_dir()
|
||||||
if !PathBuf::from(SOCKET_PATH).exists() {
|
.expect("Crushed on getting current_dir path. Check fs state!")
|
||||||
self.no_hostagent = true;
|
)
|
||||||
eprintln!("Warning: Socket-file wasn't found. Working without hostagent module...");
|
}
|
||||||
} else {
|
},
|
||||||
self.socket_path = PathBuf::from(SOCKET_PATH);
|
Err(_) => {
|
||||||
|
let default = std::env::current_dir()
|
||||||
|
.expect("Crushed on getting current_dir path. Check fs state!");
|
||||||
|
warn!(
|
||||||
|
"$NOXIS_BACKUP_FOLDER wans't set. Default value - {}",
|
||||||
|
default.display()
|
||||||
|
);
|
||||||
|
PathBuf::from(default)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// return Err(Error::msg("Socket-file not found or Noxis can't read it. Cannot start"));
|
},
|
||||||
}
|
}
|
||||||
// existing log dir
|
|
||||||
if !self.log_to.exists() && !self.no_logs {
|
|
||||||
eprintln!("Error: Log-Dir not found or Noxis can't read it. LogDir was set to default");
|
|
||||||
self.log_to = PathBuf::from("./");
|
|
||||||
// return Err(Error::msg("Log Directory Not Found or Noxis can't read it. Cannot start"));
|
|
||||||
}
|
|
||||||
// existing sock file
|
|
||||||
if !self.config.exists() {
|
|
||||||
eprintln!("Error: Invalid character in config file. Config path was set to default");
|
|
||||||
let config = PathBuf::from("/etc/settings.json");
|
|
||||||
if !config.exists() && self.no_remote_config {
|
|
||||||
return Err(Error::msg("Noxis cannot run without config. Create local config or enable remote-config mechanism"));
|
|
||||||
}
|
|
||||||
self.config = PathBuf::from("settings.json");
|
|
||||||
// return Err(Error::msg("Local Config Not Found or Noxis can't read it. Cannot start"));
|
|
||||||
}
|
|
||||||
// redis server check
|
|
||||||
EnvVars::setup(&self);
|
|
||||||
Ok(self)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
// unit tests of preboot params parsing mech
|
// unit tests of preboot params parsing mech
|
||||||
#[cfg(test)]
|
// #[cfg(test)]
|
||||||
mod preboot_unitests{
|
// mod preboot_unitests{
|
||||||
use super::*;
|
// use super::*;
|
||||||
|
|
||||||
#[test]
|
// #[test]
|
||||||
fn parsing_zero_args() {
|
// fn parsing_zero_args() {
|
||||||
assert!(PrebootParams::try_parse_from(vec!["runner-rs"]).is_ok())
|
// assert!(PrebootParams::try_parse_from(vec!["runner-rs"]).is_ok())
|
||||||
}
|
// }
|
||||||
#[test]
|
// #[test]
|
||||||
fn parsing_hagent_valid_args() {
|
// fn parsing_hagent_valid_args() {
|
||||||
assert!(PrebootParams::try_parse_from(vec![
|
// assert!(PrebootParams::try_parse_from(vec![
|
||||||
"runner-rs",
|
// "runner-rs",
|
||||||
"--socket-path", "/path/to/socket"
|
// "--socket-path", "/path/to/socket"
|
||||||
]).is_ok())
|
// ]).is_ok())
|
||||||
}
|
// }
|
||||||
#[test]
|
// #[test]
|
||||||
fn parsing_hagent_invalid_args() {
|
// fn parsing_hagent_invalid_args() {
|
||||||
assert!(PrebootParams::try_parse_from(vec![
|
// assert!(PrebootParams::try_parse_from(vec![
|
||||||
"runner-rs",
|
// "runner-rs",
|
||||||
"--socket-path", "/path/to/socket",
|
// "--socket-path", "/path/to/socket",
|
||||||
"--no-hagent"
|
// "--no-hagent"
|
||||||
]).is_err())
|
// ]).is_err())
|
||||||
}
|
// }
|
||||||
#[test]
|
// #[test]
|
||||||
fn parsing_log_valid_args() {
|
// fn parsing_log_valid_args() {
|
||||||
assert!(PrebootParams::try_parse_from(vec![
|
// assert!(PrebootParams::try_parse_from(vec![
|
||||||
"runner-rs",
|
// "runner-rs",
|
||||||
"--log-to", "/path/to/log/dir"
|
// "--log-to", "/path/to/log/dir"
|
||||||
]).is_ok())
|
// ]).is_ok())
|
||||||
}
|
// }
|
||||||
#[test]
|
// #[test]
|
||||||
fn parsing_log_invalid_args() {
|
// fn parsing_log_invalid_args() {
|
||||||
assert!(PrebootParams::try_parse_from(vec![
|
// assert!(PrebootParams::try_parse_from(vec![
|
||||||
"runner-rs",
|
// "runner-rs",
|
||||||
"--log-to /path/to/log/dir",
|
// "--log-to /path/to/log/dir",
|
||||||
"--no-logs"
|
// "--no-logs"
|
||||||
]).is_err())
|
// ]).is_err())
|
||||||
}
|
// }
|
||||||
#[test]
|
// #[test]
|
||||||
fn parsing_config_valid_args() {
|
// fn parsing_config_valid_args() {
|
||||||
assert!(PrebootParams::try_parse_from(vec![
|
// assert!(PrebootParams::try_parse_from(vec![
|
||||||
"runner-rs",
|
// "runner-rs",
|
||||||
"--no-sub",
|
// "--no-sub",
|
||||||
"--remote-server-url", "redis://127.0.0.1"
|
// "--remote-server-url", "redis://127.0.0.1"
|
||||||
]).is_ok())
|
// ]).is_err())
|
||||||
}
|
// }
|
||||||
#[test]
|
// // #[test]
|
||||||
fn parsing_config_invalid_args_noremote_nosub() {
|
// // fn parsing_config_invalid_args_noremote_nosub() {
|
||||||
assert!(PrebootParams::try_parse_from(vec![
|
// // assert!(PrebootParams::try_parse_from(vec![
|
||||||
"runner-rs",
|
// // "runner-rs",
|
||||||
"--no-remote-config", "--no-sub"
|
// // "--no-remote-config", "--no-sub"
|
||||||
]).is_err())
|
// // ]).is_err())
|
||||||
}
|
// // }
|
||||||
#[test]
|
// #[test]
|
||||||
fn parsing_config_invalid_args_noremote_remoteurl() {
|
// fn parsing_config_invalid_args_noremote_remoteurl() {
|
||||||
assert!(PrebootParams::try_parse_from(vec![
|
// assert!(PrebootParams::try_parse_from(vec![
|
||||||
"runner-rs",
|
// "runner-rs",
|
||||||
"--no-remote-config",
|
// "--no-sub",
|
||||||
"--remote-server-url", "redis://127.0.0.1"
|
// "--remote-server-url", "redis://127.0.0.1"
|
||||||
]).is_err())
|
// ]).is_err())
|
||||||
}
|
// }
|
||||||
#[test]
|
// #[test]
|
||||||
fn parsing_metrics_args_using_value_enum() {
|
// fn parsing_metrics_args_using_value_enum() {
|
||||||
assert!(PrebootParams::try_parse_from(vec!["runner-rs", "--metrics", "full"]).is_ok());
|
// assert!(PrebootParams::try_parse_from(vec!["runner-rs", "--metrics", "full"]).is_ok());
|
||||||
assert!(PrebootParams::try_parse_from(vec!["runner-rs", "--metrics", "system"]).is_ok());
|
// assert!(PrebootParams::try_parse_from(vec!["runner-rs", "--metrics", "system"]).is_ok());
|
||||||
assert!(PrebootParams::try_parse_from(vec!["runner-rs", "--metrics", "processes"]).is_ok());
|
// assert!(PrebootParams::try_parse_from(vec!["runner-rs", "--metrics", "processes"]).is_ok());
|
||||||
assert!(PrebootParams::try_parse_from(vec!["runner-rs", "--metrics", "net"]).is_ok());
|
// assert!(PrebootParams::try_parse_from(vec!["runner-rs", "--metrics", "net"]).is_ok());
|
||||||
assert!(PrebootParams::try_parse_from(vec!["runner-rs", "--metrics", "none"]).is_ok());
|
// assert!(PrebootParams::try_parse_from(vec!["runner-rs", "--metrics", "none"]).is_ok());
|
||||||
assert!(PrebootParams::try_parse_from(vec!["runner-rs", "--metrics", "unusual_value"]).is_err());
|
// assert!(PrebootParams::try_parse_from(vec!["runner-rs", "--metrics", "unusual_value"]).is_err());
|
||||||
}
|
// }
|
||||||
}
|
// }
|
||||||
|
|
|
||||||
|
|
@ -1,4 +1,3 @@
|
||||||
use super::structs::CustomError;
|
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use tokio::io;
|
use tokio::io;
|
||||||
use tokio::sync::mpsc;
|
use tokio::sync::mpsc;
|
||||||
|
|
@ -22,7 +21,7 @@ type SendersVec = Arc<Vec<Arc<mpsc::Sender<u8>>>>;
|
||||||
///
|
///
|
||||||
/// *depends on* : Sig, Signals
|
/// *depends on* : Sig, Signals
|
||||||
///
|
///
|
||||||
pub async fn set_valid_destructor(senders: SendersVec) -> Result<(), CustomError> {
|
pub async fn set_valid_destructor(senders: SendersVec) -> anyhow::Result<()> {
|
||||||
let (mut int, mut term, mut stop) = (
|
let (mut int, mut term, mut stop) = (
|
||||||
Sig::new(Signals::Sigint, senders.clone()),
|
Sig::new(Signals::Sigint, senders.clone()),
|
||||||
Sig::new(Signals::Sigterm, senders.clone()),
|
Sig::new(Signals::Sigterm, senders.clone()),
|
||||||
|
|
|
||||||
|
|
@ -1,8 +1,277 @@
|
||||||
#![allow(dead_code)]
|
#![allow(dead_code)]
|
||||||
|
|
||||||
use std::net::Ipv4Addr;
|
use async_trait::async_trait;
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
|
use std::any::Any;
|
||||||
|
use std::net::Ipv4Addr;
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
|
pub mod bus {
|
||||||
|
use std::fmt::Debug;
|
||||||
|
|
||||||
|
use super::*;
|
||||||
|
use crate::utils::metrics::processes::ProcessesQuery;
|
||||||
|
use crate::utils::metrics::MetricsExportable;
|
||||||
|
use noxis_cli::{metrics_models::MetricsMode, Cli};
|
||||||
|
pub type BusMessageContent = Box<dyn BusContent>;
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum BusMessage {
|
||||||
|
Request(
|
||||||
|
BusMessageDirection,
|
||||||
|
BusMessageContentType,
|
||||||
|
BusMessageContent,
|
||||||
|
),
|
||||||
|
Response(
|
||||||
|
BusMessageDirection,
|
||||||
|
BusMessageContentType,
|
||||||
|
BusMessageContent,
|
||||||
|
),
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum BusMessageDirection {
|
||||||
|
ToCli,
|
||||||
|
ToSupervisor,
|
||||||
|
ToMetrics,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum BusMessageContentType {
|
||||||
|
RawString,
|
||||||
|
Cli,
|
||||||
|
MetricsObj,
|
||||||
|
Result,
|
||||||
|
MetricsModeTransfered,
|
||||||
|
ProcessQuery,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum CLiCommand {
|
||||||
|
Start,
|
||||||
|
Stop,
|
||||||
|
Restart,
|
||||||
|
Freeze,
|
||||||
|
Unfreeze,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub struct InternalCli {
|
||||||
|
pub prc: String,
|
||||||
|
pub cmd: CLiCommand,
|
||||||
|
}
|
||||||
|
|
||||||
|
pub trait BusContent: Send + Sync + 'static + Debug + Any {
|
||||||
|
fn get_bus_type(&self) -> BusMessageContentType;
|
||||||
|
}
|
||||||
|
impl BusContent for anyhow::Result<String> {
|
||||||
|
fn get_bus_type(&self) -> BusMessageContentType {
|
||||||
|
BusMessageContentType::Result
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl BusContent for String {
|
||||||
|
fn get_bus_type(&self) -> BusMessageContentType {
|
||||||
|
BusMessageContentType::RawString
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl BusContent for Cli {
|
||||||
|
fn get_bus_type(&self) -> BusMessageContentType {
|
||||||
|
BusMessageContentType::Cli
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl BusContent for InternalCli {
|
||||||
|
fn get_bus_type(&self) -> BusMessageContentType {
|
||||||
|
BusMessageContentType::Cli
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl BusContent for dyn MetricsExportable {
|
||||||
|
fn get_bus_type(&self) -> BusMessageContentType {
|
||||||
|
BusMessageContentType::MetricsObj
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl BusContent for Box<dyn MetricsExportable> {
|
||||||
|
fn get_bus_type(&self) -> BusMessageContentType {
|
||||||
|
BusMessageContentType::MetricsObj
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl BusContent for MetricsMode {
|
||||||
|
fn get_bus_type(&self) -> BusMessageContentType {
|
||||||
|
BusMessageContentType::MetricsModeTransfered
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl BusContent for ProcessesQuery {
|
||||||
|
fn get_bus_type(&self) -> BusMessageContentType {
|
||||||
|
BusMessageContentType::ProcessQuery
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum DependencyType {
|
||||||
|
File,
|
||||||
|
Service,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Serialize, Clone, Copy)]
|
||||||
|
pub enum ServiceState {
|
||||||
|
Ok,
|
||||||
|
Unavailable,
|
||||||
|
}
|
||||||
|
impl std::fmt::Display for ServiceState {
|
||||||
|
fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result {
|
||||||
|
return match self {
|
||||||
|
ServiceState::Ok => write!(f, "Ok"),
|
||||||
|
ServiceState::Unavailable => write!(f, "Unavailable"),
|
||||||
|
};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct ServiceWaitConfig(u32);
|
||||||
|
|
||||||
|
impl Default for ServiceWaitConfig {
|
||||||
|
fn default() -> Self {
|
||||||
|
Self(5)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub enum FileTriggerType {
|
||||||
|
OnChange,
|
||||||
|
OnDelete,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl std::fmt::Display for FileTriggerType {
|
||||||
|
fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result {
|
||||||
|
return match self {
|
||||||
|
FileTriggerType::OnChange => write!(f, "File was changed"),
|
||||||
|
FileTriggerType::OnDelete => write!(f, "File was moved or deleted"),
|
||||||
|
};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a> FileTriggerType {
|
||||||
|
pub fn event(&self, file_name: Arc<str>, trigger: Arc<str>) -> Events {
|
||||||
|
return match self {
|
||||||
|
FileTriggerType::OnChange => Events::Negative(NegativeOutcomes::FileWasChanged(
|
||||||
|
file_name,
|
||||||
|
DependencyType::File,
|
||||||
|
trigger,
|
||||||
|
)),
|
||||||
|
FileTriggerType::OnDelete => Events::Negative(NegativeOutcomes::FileWasMovedOrDeleted(
|
||||||
|
file_name,
|
||||||
|
DependencyType::File,
|
||||||
|
trigger,
|
||||||
|
)),
|
||||||
|
};
|
||||||
|
}
|
||||||
|
pub fn event_from_file_trigger_controller(
|
||||||
|
&self,
|
||||||
|
file_name: Arc<str>,
|
||||||
|
trigger: &FileTriggersForController,
|
||||||
|
) -> Events {
|
||||||
|
return match self {
|
||||||
|
FileTriggerType::OnChange => Events::Negative(NegativeOutcomes::FileWasChanged(
|
||||||
|
file_name,
|
||||||
|
DependencyType::File,
|
||||||
|
trigger.on_change.clone(),
|
||||||
|
)),
|
||||||
|
FileTriggerType::OnDelete => Events::Negative(NegativeOutcomes::FileWasMovedOrDeleted(
|
||||||
|
file_name,
|
||||||
|
DependencyType::File,
|
||||||
|
trigger.on_delete.clone(),
|
||||||
|
)),
|
||||||
|
};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum Triggers {
|
||||||
|
File {
|
||||||
|
on_change: Arc<str>,
|
||||||
|
on_delete: Arc<str>,
|
||||||
|
},
|
||||||
|
Service {
|
||||||
|
on_lost: Arc<str>,
|
||||||
|
wait: u32,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Triggers {
|
||||||
|
pub fn new_file(on_change: Arc<str>, on_delete: Arc<str>) -> Triggers {
|
||||||
|
Triggers::File {
|
||||||
|
on_change,
|
||||||
|
on_delete,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
pub fn new_service(on_lost: Arc<str>, wait_time: u32) -> Triggers {
|
||||||
|
Triggers::Service {
|
||||||
|
on_lost,
|
||||||
|
wait: wait_time,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
pub fn to_service_negative_event(&self, service_name: Arc<str>) -> Option<Events> {
|
||||||
|
if let Triggers::Service { on_lost, .. } = self {
|
||||||
|
return Some(Events::Negative(NegativeOutcomes::ServiceIsUnreachable(
|
||||||
|
service_name,
|
||||||
|
DependencyType::Service,
|
||||||
|
on_lost.clone(),
|
||||||
|
)));
|
||||||
|
}
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub struct FileTriggersForController {
|
||||||
|
pub on_change: Arc<str>,
|
||||||
|
pub on_delete: Arc<str>,
|
||||||
|
}
|
||||||
|
pub struct ServiceTriggersForController(Arc<str>);
|
||||||
|
|
||||||
|
impl std::fmt::Display for DependencyType {
|
||||||
|
fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result {
|
||||||
|
return match self {
|
||||||
|
DependencyType::File => write!(f, "File"),
|
||||||
|
DependencyType::Service => write!(f, "Service"),
|
||||||
|
};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, serde::Serialize, Clone, Copy)]
|
||||||
|
pub enum ProcessState {
|
||||||
|
Pending,
|
||||||
|
Holding,
|
||||||
|
Stopped,
|
||||||
|
StoppedByCli,
|
||||||
|
HoldingByCli,
|
||||||
|
}
|
||||||
|
impl std::fmt::Display for ProcessState {
|
||||||
|
fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result {
|
||||||
|
return match self {
|
||||||
|
ProcessState::Pending => write!(f, "Running"),
|
||||||
|
ProcessState::Holding => write!(f, "Frozen"),
|
||||||
|
ProcessState::Stopped => write!(f, "Stopped"),
|
||||||
|
ProcessState::StoppedByCli => write!(f, "Stopped by Admin"),
|
||||||
|
ProcessState::HoldingByCli => write!(f, "Frozen by Admin"),
|
||||||
|
};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum Events {
|
||||||
|
Positive(Arc<str>),
|
||||||
|
Negative(NegativeOutcomes),
|
||||||
|
}
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub enum NegativeOutcomes {
|
||||||
|
FileWasChanged(Arc<str>, DependencyType, Arc<str>),
|
||||||
|
FileWasMovedOrDeleted(Arc<str>, DependencyType, Arc<str>),
|
||||||
|
ServiceIsUnreachable(Arc<str>, DependencyType, Arc<str>),
|
||||||
|
}
|
||||||
|
|
||||||
|
#[async_trait]
|
||||||
|
pub trait ProcessUnit {
|
||||||
|
async fn process(&mut self);
|
||||||
|
}
|
||||||
/// # an Error enum (next will be deleted and replaced)
|
/// # an Error enum (next will be deleted and replaced)
|
||||||
pub enum CustomError {
|
pub enum CustomError {
|
||||||
Fatal,
|
Fatal,
|
||||||
|
|
@ -34,12 +303,28 @@ pub struct Processes {
|
||||||
// runner_id: usize,
|
// runner_id: usize,
|
||||||
#[serde(rename = "dateOfCreation")]
|
#[serde(rename = "dateOfCreation")]
|
||||||
pub date_of_creation: String,
|
pub date_of_creation: String,
|
||||||
#[serde(rename = "configServer")]
|
|
||||||
pub config_server: String,
|
|
||||||
#[serde(default)]
|
#[serde(default)]
|
||||||
pub processes: Vec<TrackingProcess>,
|
pub processes: Vec<TrackingProcess>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl Default for Processes {
|
||||||
|
fn default() -> Self {
|
||||||
|
Self {
|
||||||
|
date_of_creation: String::new(),
|
||||||
|
processes: Vec::new(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Processes {
|
||||||
|
pub fn is_default(&self) -> bool {
|
||||||
|
self.date_of_creation.is_empty()
|
||||||
|
}
|
||||||
|
pub fn get_version(&self) -> &str {
|
||||||
|
&self.date_of_creation
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// # Struct for the 2nd level in json conf file
|
/// # Struct for the 2nd level in json conf file
|
||||||
/// ## for each process to contain info, such as name, path and dependencies
|
/// ## for each process to contain info, such as name, path and dependencies
|
||||||
///
|
///
|
||||||
|
|
@ -135,7 +420,7 @@ pub struct Files {
|
||||||
#[derive(Debug, Serialize, Deserialize, Clone)]
|
#[derive(Debug, Serialize, Deserialize, Clone)]
|
||||||
pub struct Services {
|
pub struct Services {
|
||||||
pub hostname: String,
|
pub hostname: String,
|
||||||
pub port: u32,
|
pub port: Option<u32>,
|
||||||
pub triggers: ServiceTriggers,
|
pub triggers: ServiceTriggers,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -159,7 +444,6 @@ pub struct Services {
|
||||||
#[derive(Debug, Serialize, Deserialize, Clone)]
|
#[derive(Debug, Serialize, Deserialize, Clone)]
|
||||||
pub struct ServiceTriggers {
|
pub struct ServiceTriggers {
|
||||||
pub wait: u32,
|
pub wait: u32,
|
||||||
pub delay: u32,
|
|
||||||
#[serde(rename = "onLost")]
|
#[serde(rename = "onLost")]
|
||||||
pub on_lost: String,
|
pub on_lost: String,
|
||||||
}
|
}
|
||||||
|
|
@ -186,6 +470,8 @@ pub struct FileTriggers {
|
||||||
pub on_delete: String,
|
pub on_delete: String,
|
||||||
#[serde(rename = "onChange")]
|
#[serde(rename = "onChange")]
|
||||||
pub on_change: String,
|
pub on_change: String,
|
||||||
|
#[serde(rename = "doRestore")]
|
||||||
|
pub do_restore: bool,
|
||||||
}
|
}
|
||||||
|
|
||||||
/// # Metrics struct
|
/// # Metrics struct
|
||||||
|
|
@ -195,24 +481,23 @@ pub struct FileTriggers {
|
||||||
///
|
///
|
||||||
/// *depends on* : `ContainerMetrics`, `ProcessMetrics`
|
/// *depends on* : `ContainerMetrics`, `ProcessMetrics`
|
||||||
///
|
///
|
||||||
#[derive(Debug, Clone, Serialize,)]
|
#[derive(Debug, Clone, Serialize)]
|
||||||
pub struct Metrics {
|
pub struct Metrics {
|
||||||
pub container_metrics : ContainerMetrics,
|
pub container_metrics: ContainerMetrics,
|
||||||
pub processes_metrics : Vec<ProcessMetrics>,
|
pub processes_metrics: Vec<ProcessMetrics>,
|
||||||
// pub net_metrics : Vec<PacketInfo>,
|
// pub net_metrics : Vec<PacketInfo>,
|
||||||
}
|
}
|
||||||
/// ## Metrics struct's constructor
|
/// ## Metrics struct's constructor
|
||||||
impl Metrics {
|
impl Metrics {
|
||||||
pub fn new(cm: ContainerMetrics, prm: Vec<ProcessMetrics>) -> Self {
|
pub fn new(cm: ContainerMetrics, prm: Vec<ProcessMetrics>) -> Self {
|
||||||
Metrics {
|
Metrics {
|
||||||
container_metrics : cm,
|
container_metrics: cm,
|
||||||
processes_metrics : prm,
|
processes_metrics: prm,
|
||||||
// net_metrics : net,
|
// net_metrics : net,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
/// # Container metrics struct
|
/// # Container metrics struct
|
||||||
/// ## for gathering all container metrics
|
/// ## for gathering all container metrics
|
||||||
///
|
///
|
||||||
|
|
@ -222,20 +507,20 @@ impl Metrics {
|
||||||
///
|
///
|
||||||
#[derive(Debug, Clone, Serialize)]
|
#[derive(Debug, Clone, Serialize)]
|
||||||
pub struct ContainerMetrics {
|
pub struct ContainerMetrics {
|
||||||
container_id : String,
|
container_id: String,
|
||||||
cpu_load : f32,
|
cpu_load: f32,
|
||||||
ram_load : f32,
|
ram_load: f32,
|
||||||
// pub net_activity : ???
|
// pub net_activity : ???
|
||||||
processes : Vec<String>,
|
processes: Vec<String>,
|
||||||
}
|
}
|
||||||
/// ## Container struct's constructor
|
/// ## Container struct's constructor
|
||||||
impl ContainerMetrics {
|
impl ContainerMetrics {
|
||||||
pub fn new(container_id : &str, cpu: f32, ram: f32, subsystems: Vec<String>,) -> Self{
|
pub fn new(container_id: &str, cpu: f32, ram: f32, subsystems: Vec<String>) -> Self {
|
||||||
ContainerMetrics {
|
ContainerMetrics {
|
||||||
container_id : String::from(container_id),
|
container_id: String::from(container_id),
|
||||||
cpu_load : cpu,
|
cpu_load: cpu,
|
||||||
ram_load : ram,
|
ram_load: ram,
|
||||||
processes : subsystems,
|
processes: subsystems,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -249,17 +534,17 @@ impl ContainerMetrics {
|
||||||
///
|
///
|
||||||
#[derive(Debug, Clone, Serialize)]
|
#[derive(Debug, Clone, Serialize)]
|
||||||
pub struct ProcessMetrics {
|
pub struct ProcessMetrics {
|
||||||
pub process_name : String,
|
pub process_name: String,
|
||||||
cpu_load : f32,
|
cpu_load: f32,
|
||||||
ram_load : f32,
|
ram_load: f32,
|
||||||
}
|
}
|
||||||
/// ## Process struct's constructor
|
/// ## Process struct's constructor
|
||||||
impl ProcessMetrics {
|
impl ProcessMetrics {
|
||||||
pub fn new(process_name :&str, cpu: f32, ram: f32) -> Self {
|
pub fn new(process_name: &str, cpu: f32, ram: f32) -> Self {
|
||||||
ProcessMetrics {
|
ProcessMetrics {
|
||||||
process_name : String::from(process_name),
|
process_name: String::from(process_name),
|
||||||
cpu_load : cpu,
|
cpu_load: cpu,
|
||||||
ram_load : ram,
|
ram_load: ram,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -273,19 +558,19 @@ impl ProcessMetrics {
|
||||||
///
|
///
|
||||||
#[derive(Debug, Clone, Serialize)]
|
#[derive(Debug, Clone, Serialize)]
|
||||||
pub struct PacketInfo {
|
pub struct PacketInfo {
|
||||||
protocol : String,
|
protocol: String,
|
||||||
dst_ip : Ipv4Addr,
|
dst_ip: Ipv4Addr,
|
||||||
src_ip : Ipv4Addr,
|
src_ip: Ipv4Addr,
|
||||||
size : usize,
|
size: usize,
|
||||||
}
|
}
|
||||||
/// ## PacketInfo's constructor
|
/// ## PacketInfo's constructor
|
||||||
impl PacketInfo {
|
impl PacketInfo {
|
||||||
pub fn new(prt: String, dest: Ipv4Addr, src: Ipv4Addr, size_of_packet: usize) -> Self {
|
pub fn new(prt: String, dest: Ipv4Addr, src: Ipv4Addr, size_of_packet: usize) -> Self {
|
||||||
PacketInfo {
|
PacketInfo {
|
||||||
protocol : prt,
|
protocol: prt,
|
||||||
dst_ip : dest,
|
dst_ip: dest,
|
||||||
src_ip : src,
|
src_ip: src,
|
||||||
size : size_of_packet,
|
size: size_of_packet,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -1,235 +1,386 @@
|
||||||
|
pub mod bus;
|
||||||
pub mod files;
|
pub mod files;
|
||||||
pub mod hagent;
|
pub mod hagent;
|
||||||
pub mod metrics;
|
pub mod metrics;
|
||||||
pub mod prcs;
|
pub mod prcs;
|
||||||
pub mod services;
|
pub mod services;
|
||||||
|
|
||||||
// TODO : saving current flags state
|
use crate::options::structs::bus::{BusMessage, BusMessageContentType, InternalCli};
|
||||||
|
use crate::options::structs::Processes;
|
||||||
use crate::options::structs::CustomError;
|
use async_trait::async_trait;
|
||||||
use crate::options::structs::TrackingProcess;
|
use files::v2::FilesController;
|
||||||
use files::create_watcher;
|
use lazy_static::lazy_static;
|
||||||
use files::file_handler;
|
use log::{error, info};
|
||||||
use inotify::Inotify;
|
use prcs::v2::ProcessesController;
|
||||||
use log::{error, warn};
|
use services::v2::ServicesController;
|
||||||
use prcs::{
|
|
||||||
freeze_process, is_active, is_frozen, restart_process, start_process, terminate_process,
|
|
||||||
unfreeze_process,
|
|
||||||
};
|
|
||||||
use services::service_handler;
|
|
||||||
use std::process::Command;
|
use std::process::Command;
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use tokio::join;
|
|
||||||
use tokio::sync::mpsc;
|
use tokio::sync::mpsc;
|
||||||
use tokio::time::Duration;
|
use tokio::time::Duration;
|
||||||
|
|
||||||
const GET_ID_CMD: &str = "hostname";
|
lazy_static! {
|
||||||
|
static ref GET_ID_CMD: &'static str = "hostname";
|
||||||
|
}
|
||||||
|
|
||||||
/// # Fn `run_daemons`
|
// const GET_ID_CMD: &str = "hostname";
|
||||||
/// ## async func to run 3 main daemons: process, service and file monitors and manage process state according to given messages into channel
|
|
||||||
///
|
pub mod v2 {
|
||||||
/// *input* : `Arc<TrackingProcess>`, `Arc<mpsc::Sender<u8>>`, `&mut mpsc::Receiver<u8>`,
|
use super::*;
|
||||||
///
|
use crate::options::preboot::PrebootParams;
|
||||||
/// *output* : ()
|
use crate::utils::metrics::processes::{ProcessesAll, ProcessesQuery};
|
||||||
///
|
use crate::{
|
||||||
/// *initiator* : main thread
|
options::structs::{
|
||||||
///
|
bus::CLiCommand, Events, FileTriggersForController, ProcessUnit, Triggers,
|
||||||
/// *managing* : Arc to current process struct, Arc to managing channel writer, mut ref to managing channel reader
|
},
|
||||||
///
|
utils::metrics::processes::deps::{Dependencies, FilesExtended, ServicesExtended},
|
||||||
/// *depends on* : all module `prcs`'s functions, fn `running_handler`, fn `utils::files::create_watcher`
|
};
|
||||||
///
|
use std::any::Any;
|
||||||
/// > *hint* : give mpsc with capacity 1 to jump over potential errors during running process
|
use std::collections::{BTreeMap, HashMap, LinkedList, VecDeque};
|
||||||
///
|
|
||||||
pub async fn run_daemons(
|
type BusReciever = tokio::sync::mpsc::Receiver<BusMessage>;
|
||||||
proc: Arc<TrackingProcess>,
|
type BusSender = Arc<tokio::sync::mpsc::Sender<BusMessage>>;
|
||||||
tx: Arc<mpsc::Sender<u8>>,
|
|
||||||
rx: &mut mpsc::Receiver<u8>,
|
#[derive(Debug)]
|
||||||
) {
|
enum ControllerResult {
|
||||||
// creating watchers + ---buffers---
|
Process(Option<ProcessesController>),
|
||||||
let mut watchers: Vec<Inotify> = vec![];
|
File(Option<FilesController>),
|
||||||
for file in proc.dependencies.files.clone().into_iter() {
|
Service(Option<ServicesController>),
|
||||||
if let Ok(watcher) = create_watcher(&file.filename, &file.src).await {
|
}
|
||||||
watchers.push(watcher);
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
struct Supervisor {
|
||||||
|
prcs: LinkedList<ProcessesController>,
|
||||||
|
files: LinkedList<FilesController>,
|
||||||
|
services: LinkedList<ServicesController>,
|
||||||
|
config: Arc<Processes>,
|
||||||
|
bus: (BusReciever, BusSender),
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Supervisor {
|
||||||
|
pub fn new(bus_reciever: BusReciever, bus_sender: BusSender) -> Supervisor {
|
||||||
|
Supervisor {
|
||||||
|
prcs: LinkedList::new(),
|
||||||
|
files: LinkedList::new(),
|
||||||
|
services: LinkedList::new(),
|
||||||
|
config: Arc::new(Processes::default()),
|
||||||
|
bus: (bus_reciever, bus_sender),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
pub async fn with_config(
|
||||||
|
mut self,
|
||||||
|
config: Processes,
|
||||||
|
preboot : Arc<PrebootParams>
|
||||||
|
) -> Supervisor {
|
||||||
|
self.config = Arc::from(config);
|
||||||
|
let _ = self.config.processes.iter().for_each(|prc| {
|
||||||
|
let (rx, tx) = mpsc::channel::<Events>(100);
|
||||||
|
let temp = ProcessesController::new(&prc.name, tx).with_exe(&prc.path);
|
||||||
|
if !self.prcs.contains(&temp) {
|
||||||
|
self.prcs.push_back(temp);
|
||||||
|
}
|
||||||
|
let rx = Arc::new(rx);
|
||||||
|
let proc_name: Arc<str> = Arc::from(prc.name.clone());
|
||||||
|
|
||||||
|
let _ = prc.dependencies.files.iter().for_each(|file| {
|
||||||
|
let mut hm = HashMap::new();
|
||||||
|
let triggers = FileTriggersForController {
|
||||||
|
on_change: Arc::from(file.triggers.on_change.clone()),
|
||||||
|
on_delete: Arc::from(file.triggers.on_delete.clone()),
|
||||||
|
};
|
||||||
|
hm.insert(proc_name.clone(), (triggers, rx.clone()));
|
||||||
|
|
||||||
|
let backup_file = {
|
||||||
|
if file.triggers.do_restore {
|
||||||
|
use ulid::Ulid;
|
||||||
|
format!("{}{}.bak", {
|
||||||
|
let path = preboot.backup_folder.to_string_lossy();
|
||||||
|
if path.ends_with("/") { path.to_string() }
|
||||||
|
else { format!("{}/", path) }
|
||||||
|
}, Ulid::new())
|
||||||
} else {
|
} else {
|
||||||
let _ = tx.send(121).await;
|
String::new()
|
||||||
}
|
}
|
||||||
// watchers.push(create_watcher(&file.filename, &file.src).await.unwrap());
|
};
|
||||||
}
|
|
||||||
let watchers_clone: Arc<tokio::sync::Mutex<Vec<Inotify>>> =
|
|
||||||
Arc::new(tokio::sync::Mutex::new(watchers));
|
|
||||||
|
|
||||||
|
let tempfile =
|
||||||
|
FilesController::new(&file.filename.as_str(), hm).with_path(&file.src, backup_file);
|
||||||
|
|
||||||
|
if let Ok(file) = tempfile {
|
||||||
|
if let Some(current_file) = self.files.iter_mut().find(|a| &&file == a) {
|
||||||
|
current_file.add_event(file);
|
||||||
|
} else {
|
||||||
|
self.files.push_back(file);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
// servs
|
||||||
|
let _ = prc.dependencies.services.iter().for_each(|serv| {
|
||||||
|
let access_url =
|
||||||
|
ServicesController::get_access_url(&serv.hostname, serv.port.as_ref());
|
||||||
|
// preparations
|
||||||
|
let rx = rx.clone();
|
||||||
|
let serv_cont =
|
||||||
|
ServicesController::new().with_access_name(&serv.hostname, &access_url);
|
||||||
|
// triggers
|
||||||
|
let arc: Arc<str> = Arc::from(serv.triggers.on_lost.clone());
|
||||||
|
let triggers = Triggers::new_service(arc, serv.triggers.wait);
|
||||||
|
|
||||||
|
if let Some(proc) = self.services.iter_mut().find(|a| &&serv_cont == a) {
|
||||||
|
proc.add_process(&prc.name, triggers, rx);
|
||||||
|
} else {
|
||||||
|
// vecdeque for queue
|
||||||
|
let mut vec: VecDeque<Arc<str>> = VecDeque::new();
|
||||||
|
vec.push_back(proc_name.clone());
|
||||||
|
// connection_queue
|
||||||
|
let mut connection_queue: BTreeMap<u32, VecDeque<Arc<str>>> =
|
||||||
|
BTreeMap::new();
|
||||||
|
connection_queue.insert(serv.triggers.wait, vec);
|
||||||
|
// event_reg
|
||||||
|
let mut hm = HashMap::new();
|
||||||
|
hm.insert(proc_name.clone(), (triggers, rx));
|
||||||
|
|
||||||
|
let serv_cont = serv_cont.with_params(connection_queue, hm);
|
||||||
|
self.services.push_back(serv_cont);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
|
self
|
||||||
|
}
|
||||||
|
pub fn get_stats(&self) -> String {
|
||||||
|
format!(
|
||||||
|
"processes: {}, files: {}, services: {}",
|
||||||
|
self.prcs.len(),
|
||||||
|
self.files.len(),
|
||||||
|
self.services.len()
|
||||||
|
)
|
||||||
|
}
|
||||||
|
pub async fn extract_extended_procs(
|
||||||
|
config: Arc<Processes>,
|
||||||
|
prcs_list: &LinkedList<ProcessesController>,
|
||||||
|
files_list: &LinkedList<FilesController>,
|
||||||
|
servs_list: &LinkedList<ServicesController>,
|
||||||
|
) -> Vec<ProcessesAll> {
|
||||||
|
let mut procs = Vec::new();
|
||||||
|
for prc in config.processes.iter() {
|
||||||
|
if let Some(prc_cont) = prcs_list
|
||||||
|
.iter()
|
||||||
|
.find(|&prc_cont| prc.name == *prc_cont.name)
|
||||||
|
{
|
||||||
|
let mut vec_files = Vec::new();
|
||||||
|
let mut vec_services = Vec::new();
|
||||||
|
prc.dependencies
|
||||||
|
.files
|
||||||
|
.iter()
|
||||||
|
.map(|file| (file, format!("{}{}", file.src, file.filename)))
|
||||||
|
.for_each(|(file, code_name)| {
|
||||||
|
if let Some(file_cont) = files_list
|
||||||
|
.iter()
|
||||||
|
.find(|&file_cont| *file_cont.get_code_name() == code_name)
|
||||||
|
{
|
||||||
|
vec_files.push(FilesExtended {
|
||||||
|
name: file.filename.to_string(),
|
||||||
|
path: file.src.to_string(),
|
||||||
|
status: file_cont.get_state(),
|
||||||
|
backup_file : file_cont.get_backup_file(),
|
||||||
|
triggers: file.triggers.to_owned(),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
});
|
||||||
|
prc.dependencies
|
||||||
|
.services
|
||||||
|
.iter()
|
||||||
|
.map(|serv| {
|
||||||
|
(
|
||||||
|
serv,
|
||||||
|
format!("{}{}", serv.hostname, {
|
||||||
|
if let Some(port) = serv.port {
|
||||||
|
format!(":{}", port)
|
||||||
|
} else {
|
||||||
|
String::new()
|
||||||
|
}
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
})
|
||||||
|
.for_each(|(serv, acces_url)| {
|
||||||
|
if let Some(serv_cont) = servs_list
|
||||||
|
.iter()
|
||||||
|
.find(|&serv_cont| *serv_cont.get_arc_access_url() == acces_url)
|
||||||
|
{
|
||||||
|
vec_services.push(ServicesExtended {
|
||||||
|
name: serv.hostname.to_owned(),
|
||||||
|
access_name: (*serv_cont.get_arc_access_url()).to_owned(),
|
||||||
|
status: serv_cont.get_state(),
|
||||||
|
triggers: serv.triggers.to_owned(),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
});
|
||||||
|
procs.push(ProcessesAll {
|
||||||
|
name: prc_cont.name.clone().to_string(),
|
||||||
|
state: prc_cont.get_state(),
|
||||||
|
pid: prc_cont.get_pid(),
|
||||||
|
dependencies: Dependencies {
|
||||||
|
files: vec_files,
|
||||||
|
services: vec_services,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
procs
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[async_trait]
|
||||||
|
impl ProcessUnit for Supervisor {
|
||||||
|
async fn process(&mut self) {
|
||||||
|
info!("Initializing monitoring ...");
|
||||||
loop {
|
loop {
|
||||||
let run_hand = running_handler(proc.clone(), tx.clone(), watchers_clone.clone());
|
|
||||||
tokio::select! {
|
|
||||||
_ = run_hand => continue,
|
|
||||||
_val = rx.recv() => {
|
|
||||||
if process_protocol_symbol(proc.clone(), _val.unwrap()).await.is_err() {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
},
|
|
||||||
}
|
|
||||||
tokio::task::yield_now().await;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn process_protocol_symbol(proc: Arc<TrackingProcess>, val: u8) -> Result<(), CustomError>{
|
|
||||||
match val {
|
|
||||||
// 1 - File-dependency handling error -> terminating (after waiting)
|
|
||||||
1 => {
|
|
||||||
if is_active(&proc.name).await {
|
|
||||||
error!("File-dependency handling error: Terminating {} process ..." , &proc.name);
|
|
||||||
terminate_process(&proc.name).await;
|
|
||||||
tokio::time::sleep(Duration::from_millis(500)).await;
|
|
||||||
}
|
|
||||||
// return;
|
|
||||||
},
|
|
||||||
// 2 - File-dependency handling error -> holding (after waiting)
|
|
||||||
2 => {
|
|
||||||
if !is_frozen(&proc.name).await {
|
|
||||||
error!("File-dependency handling error: Freezing {} process ..." , &proc.name);
|
|
||||||
freeze_process(&proc.name).await;
|
|
||||||
tokio::time::sleep(Duration::from_millis(100)).await;
|
|
||||||
}
|
|
||||||
},
|
|
||||||
// 3 - Running process error
|
|
||||||
3 => {
|
|
||||||
error!("Error due to starting {} process", &proc.name);
|
|
||||||
return Err(CustomError::Fatal)
|
|
||||||
},
|
|
||||||
// 4 - Timeout of waiting service-dependency -> staying (after waiting)
|
|
||||||
4 => {
|
|
||||||
// warn!("Timeout of waiting service-dependency: Ignoring on {} process ..." , &proc.name);
|
|
||||||
tokio::time::sleep(Duration::from_millis(100)).await;
|
|
||||||
},
|
|
||||||
// 5 - Timeout of waiting service-dependency -> terminating (after waiting)
|
|
||||||
5 => {
|
|
||||||
if is_active(&proc.name).await {
|
|
||||||
error!("Timeout of waiting service-dependency: Terminating {} process ..." , &proc.name);
|
|
||||||
terminate_process(&proc.name).await;
|
|
||||||
tokio::time::sleep(Duration::from_millis(500)).await;
|
|
||||||
}
|
|
||||||
},
|
|
||||||
// 6 - Timeout of waiting service-dependency -> holding (after waiting)
|
|
||||||
6 => {
|
|
||||||
// println!("holding {}-{}", proc.name, is_active(&proc.name).await);
|
|
||||||
if !is_frozen(&proc.name).await {
|
|
||||||
error!("Timeout of waiting service-dependency: Freezing {} process ..." , &proc.name);
|
|
||||||
freeze_process(&proc.name).await;
|
|
||||||
tokio::time::sleep(Duration::from_secs(1)).await;
|
|
||||||
}
|
|
||||||
},
|
|
||||||
// // 7 - File-dependency change -> terminating (after check)
|
|
||||||
7 => {
|
|
||||||
error!("File-dependency warning (file changed). Terminating {} process...", &proc.name);
|
|
||||||
terminate_process(&proc.name).await;
|
|
||||||
tokio::time::sleep(Duration::from_millis(100)).await;
|
|
||||||
return Err(CustomError::Fatal)
|
|
||||||
},
|
|
||||||
// // 8 - File-dependency change -> restarting (after check)
|
|
||||||
8 => {
|
|
||||||
warn!("File-dependency warning (file changed). Restarting {} process...", &proc.name);
|
|
||||||
let _ = restart_process(&proc.name, &proc.path).await;
|
|
||||||
tokio::time::sleep(Duration::from_millis(100)).await;
|
|
||||||
},
|
|
||||||
// // 9 - File-dependency change -> staying (after check)
|
|
||||||
9 => {
|
|
||||||
// no need to trash logs
|
|
||||||
warn!("File-dependency warning (file changed). Ignoring event on {} process...", &proc.name);
|
|
||||||
tokio::time::sleep(Duration::from_millis(100)).await;
|
|
||||||
},
|
|
||||||
|
|
||||||
// 10 - Process unfreaze call via file handler (or service handler)
|
|
||||||
10 | 11 => {
|
|
||||||
if is_frozen(&proc.name).await {
|
|
||||||
warn!("Unfreezing process {} call...", &proc.name);
|
|
||||||
unfreeze_process(&proc.name).await;
|
|
||||||
}
|
|
||||||
tokio::time::sleep(Duration::from_millis(100)).await;
|
|
||||||
},
|
|
||||||
// 11 - Process unfreaze call via service handler
|
|
||||||
// 11 => {
|
|
||||||
// if is_frozen(&proc.name).await {
|
|
||||||
// warn!("Unfreezing process {} call...", &proc.name);
|
|
||||||
// unfreeze_process(&proc.name).await;
|
|
||||||
// }
|
|
||||||
// tokio::time::sleep(Duration::from_millis(100)).await;
|
|
||||||
// },
|
|
||||||
// 101 - Impermissible trigger values in JSON
|
|
||||||
101 => {
|
|
||||||
error!("Impermissible trigger values in JSON in {}'s block. Killing thread...", &proc.name);
|
|
||||||
if is_active(&proc.name).await {
|
|
||||||
terminate_process(&proc.name).await;
|
|
||||||
}
|
|
||||||
return Err(CustomError::Fatal)
|
|
||||||
},
|
|
||||||
//
|
//
|
||||||
// 121 - Cannot create valid watcher for file dependency
|
let rec = &mut self.bus.0;
|
||||||
// todo : think about valid situation
|
while let Ok(request) = rec.try_recv() {
|
||||||
121 => {
|
if let BusMessage::Request(_, _, cont) = request {
|
||||||
error!("Cannot create valid watcher for file dependency. Terminating {} process...", &proc.name);
|
let cont: Box<dyn Any + Send> = cont;
|
||||||
let _ = terminate_process(&proc.name).await;
|
match cont.downcast::<InternalCli>() {
|
||||||
return Err(CustomError::Fatal)
|
Ok(cli) => {
|
||||||
},
|
let mut count = 0;
|
||||||
// 111 - global thread termination with killing current child in a face
|
let fut = (&mut self.prcs)
|
||||||
// of a current process
|
.into_iter()
|
||||||
111 => {
|
.find(|prc| prc.name == Arc::from(cli.prc.as_ref()))
|
||||||
warn!("Terminating {}'s child processes...", &proc.name);
|
.map(|prc| async {
|
||||||
match is_active(&proc.name).await {
|
let count = &mut count;
|
||||||
true => {
|
*count += 1;
|
||||||
terminate_process(&proc.name).await;
|
let res = match cli.cmd {
|
||||||
},
|
CLiCommand::Start => prc.start_by_user_call().await,
|
||||||
false => {
|
CLiCommand::Stop => prc.stop_by_user_call().await,
|
||||||
log::info!("Process {} is already terminated!", proc.name);
|
CLiCommand::Restart => prc.restart_by_user_call().await,
|
||||||
},
|
CLiCommand::Freeze => prc.freeze_by_user_call().await,
|
||||||
|
CLiCommand::Unfreeze => {
|
||||||
|
prc.unfreeze_by_user_call().await
|
||||||
|
}
|
||||||
|
};
|
||||||
|
let sender = self.bus.1.clone();
|
||||||
|
let resp_content = match res {
|
||||||
|
Ok(_) => Ok(format!(
|
||||||
|
"Ok on user call abour process {}",
|
||||||
|
prc.name
|
||||||
|
)),
|
||||||
|
Err(er) => Err(anyhow::Error::msg(format!(
|
||||||
|
"Error: User call for process {} failed : {}",
|
||||||
|
prc.name, er
|
||||||
|
))),
|
||||||
|
};
|
||||||
|
let _ = sender.send(BusMessage::Response(
|
||||||
|
crate::options::structs::bus::BusMessageDirection::ToCli,
|
||||||
|
BusMessageContentType::Result,
|
||||||
|
Box::new(resp_content)
|
||||||
|
)).await;
|
||||||
|
1
|
||||||
|
});
|
||||||
|
if let Some(fut) = fut {
|
||||||
|
fut.await;
|
||||||
|
} else {
|
||||||
|
let _ = self.bus.1.clone().send(BusMessage::Response(
|
||||||
|
crate::options::structs::bus::BusMessageDirection::ToCli,
|
||||||
|
BusMessageContentType::RawString,
|
||||||
|
Box::new(
|
||||||
|
Err(anyhow::Error::msg(format!("No process named `{}` was found in controlled scope", cli.prc)))
|
||||||
|
)
|
||||||
|
)).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(boxed) => {
|
||||||
|
if let Ok(query) = boxed.downcast::<ProcessesQuery>() {
|
||||||
|
match *query {
|
||||||
|
ProcessesQuery::QueryAll => {
|
||||||
|
let procs = Self::extract_extended_procs(
|
||||||
|
self.config.clone(),
|
||||||
|
&self.prcs,
|
||||||
|
&self.files,
|
||||||
|
&self.services,
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
let _ = self.bus.1.clone().send(BusMessage::Response(
|
||||||
|
crate::options::structs::bus::BusMessageDirection::ToMetrics,
|
||||||
|
BusMessageContentType::ProcessQuery,
|
||||||
|
Box::new(
|
||||||
|
ProcessesQuery::All(procs)
|
||||||
|
)
|
||||||
|
)).await;
|
||||||
|
}
|
||||||
|
ProcessesQuery::QueryGeneral => {
|
||||||
|
let mut vec = Vec::new();
|
||||||
|
for prc in &self.prcs {
|
||||||
|
vec.push(prc.get_general_info().await);
|
||||||
|
}
|
||||||
|
let _ = self.bus.1.clone().send(BusMessage::Response(
|
||||||
|
crate::options::structs::bus::BusMessageDirection::ToMetrics,
|
||||||
|
BusMessageContentType::ProcessQuery,
|
||||||
|
Box::new(
|
||||||
|
ProcessesQuery::General(vec)
|
||||||
|
)
|
||||||
|
)).await;
|
||||||
|
}
|
||||||
|
_ => {
|
||||||
|
let _ = self.bus.1.clone().send(BusMessage::Response(
|
||||||
|
crate::options::structs::bus::BusMessageDirection::ToCli,
|
||||||
|
BusMessageContentType::RawString,
|
||||||
|
Box::new(
|
||||||
|
Err(anyhow::Error::msg("Unknown request format was send to the Supervisor"))
|
||||||
|
)
|
||||||
|
)).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
let mut tasks: Vec<tokio::task::JoinHandle<ControllerResult>> = vec![];
|
||||||
|
if let Some(mut val) = self.prcs.pop_front() {
|
||||||
|
tasks.push(tokio::spawn(async move {
|
||||||
|
val.process().await;
|
||||||
|
ControllerResult::Process(Some(val))
|
||||||
|
}));
|
||||||
|
}
|
||||||
|
if let Some(mut val) = self.files.pop_front() {
|
||||||
|
tasks.push(tokio::spawn(async move {
|
||||||
|
val.process().await;
|
||||||
|
ControllerResult::File(Some(val))
|
||||||
|
}));
|
||||||
|
}
|
||||||
|
if let Some(mut val) = self.services.pop_front() {
|
||||||
|
tasks.push(tokio::spawn(async move {
|
||||||
|
val.process().await;
|
||||||
|
ControllerResult::Service(Some(val))
|
||||||
|
}));
|
||||||
|
}
|
||||||
|
for task in tasks {
|
||||||
|
match task.await {
|
||||||
|
Ok(ControllerResult::Process(Some(val))) => self.prcs.push_back(val),
|
||||||
|
Ok(ControllerResult::File(Some(val))) => self.files.push_back(val),
|
||||||
|
Ok(ControllerResult::Service(Some(val))) => self.services.push_back(val),
|
||||||
|
Err(er) => error!("Controller task crushed : {er}. Cannot push back to the exec queue ..."),
|
||||||
|
_ => { /* DEAD END (CAN NOT BE EXECUTED) */},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
tokio::time::sleep(Duration::from_millis(100)).await;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
},
|
|
||||||
_ => {},
|
|
||||||
}
|
}
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
// check process status daemon
|
|
||||||
/// # Fn `run_daemons`
|
|
||||||
/// ## func to async exec subjobs of checking process, services and files states
|
|
||||||
///
|
|
||||||
/// *input* : `Arc<TrackingProcess>`, `Arc<mpsc::Sender<u8>>`, `Arc<tokio::sync::Mutex<Vec<Inotify>>>`
|
|
||||||
///
|
|
||||||
/// *output* : ()
|
|
||||||
///
|
|
||||||
/// *initiator* : fn `run_daemons`
|
|
||||||
///
|
|
||||||
/// *managing* : Arc to current process struct, Arc to Mutex to list of file watchers
|
|
||||||
///
|
|
||||||
/// *depends on* : fn `utils::files::file_handler`, fn `utils::services::service_handler`, fn `utils::prcs::{is_active, is_frozen, start_process}`
|
|
||||||
///
|
|
||||||
pub async fn running_handler(
|
|
||||||
prc: Arc<TrackingProcess>,
|
|
||||||
tx: Arc<mpsc::Sender<u8>>,
|
|
||||||
watchers: Arc<tokio::sync::Mutex<Vec<Inotify>>>,
|
|
||||||
) {
|
|
||||||
// services and files check (once)
|
|
||||||
let files_check = file_handler(
|
|
||||||
&prc.name,
|
|
||||||
&prc.dependencies.files,
|
|
||||||
tx.clone(),
|
|
||||||
watchers.clone(),
|
|
||||||
);
|
|
||||||
let services_check = service_handler(&prc.name, &prc.dependencies.services, tx.clone());
|
|
||||||
|
|
||||||
let res = join!(files_check, services_check);
|
pub async fn init_monitoring(
|
||||||
// if inactive -> spawn checks -> active is true
|
config: Processes,
|
||||||
if !is_active(&prc.name).await && res.0.is_ok() && res.1.is_ok() {
|
preboot : Arc<PrebootParams>,
|
||||||
if start_process(&prc.name, &prc.path).await.is_err() {
|
bus_reciever: BusReciever,
|
||||||
tx.send(3).await.unwrap();
|
bus_sender: BusSender,
|
||||||
return;
|
) -> anyhow::Result<()> {
|
||||||
|
let mut supervisor = Supervisor::new(bus_reciever, bus_sender)
|
||||||
|
.with_config(config, preboot)
|
||||||
|
.await;
|
||||||
|
info!("Monitoring: {} ", &supervisor.get_stats());
|
||||||
|
supervisor.process().await;
|
||||||
|
Ok(())
|
||||||
}
|
}
|
||||||
}
|
|
||||||
// if frozen -> spawn checks -> unfreeze is true
|
|
||||||
else if is_frozen(&prc.name).await && res.0.is_ok() && res.1.is_ok() {
|
|
||||||
tx.send(10).await.unwrap();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
// tokio::time::sleep(Duration::from_millis(100)).await;
|
|
||||||
tokio::task::yield_now().await;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// todo: cmd across cat /proc/self/mountinfo | grep "/docker/containers/" | head -1 | awk -F '/' '{print $5}'
|
// todo: cmd across cat /proc/self/mountinfo | grep "/docker/containers/" | head -1 | awk -F '/' '{print $5}'
|
||||||
|
|
@ -247,7 +398,7 @@ pub async fn running_handler(
|
||||||
/// *depends on* : -
|
/// *depends on* : -
|
||||||
///
|
///
|
||||||
pub fn get_container_id() -> Option<String> {
|
pub fn get_container_id() -> Option<String> {
|
||||||
match Command::new(GET_ID_CMD).output() {
|
match Command::new(*GET_ID_CMD).output() {
|
||||||
Ok(output) => {
|
Ok(output) => {
|
||||||
if !output.status.success() {
|
if !output.status.success() {
|
||||||
return None;
|
return None;
|
||||||
|
|
@ -256,7 +407,7 @@ pub fn get_container_id() -> Option<String> {
|
||||||
if id.is_empty() {
|
if id.is_empty() {
|
||||||
return None;
|
return None;
|
||||||
}
|
}
|
||||||
Some(String::from_utf8_lossy(&output.stdout).to_string())
|
Some(String::from_utf8_lossy(&output.stdout).trim().to_string())
|
||||||
}
|
}
|
||||||
Err(_) => None,
|
Err(_) => None,
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,97 @@
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
|
use crate::options::structs::bus::{BusMessage, BusMessageDirection};
|
||||||
|
use crate::options::structs::ProcessUnit;
|
||||||
|
use log::{error, trace};
|
||||||
|
use tokio::sync::mpsc::{Receiver, Sender};
|
||||||
|
|
||||||
|
type Inner = Receiver<BusMessage>;
|
||||||
|
type Outter = Arc<Sender<BusMessage>>;
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub struct Highway {
|
||||||
|
to_cli: Outter,
|
||||||
|
to_supervisor: Outter,
|
||||||
|
to_metrics: Outter,
|
||||||
|
}
|
||||||
|
impl Highway {
|
||||||
|
fn new(to_cli: Outter, to_supervisor: Outter, to_metrics: Outter) -> Self {
|
||||||
|
Self {
|
||||||
|
to_cli,
|
||||||
|
to_supervisor,
|
||||||
|
to_metrics,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
async fn send(&self, msg: BusMessage) -> anyhow::Result<()> {
|
||||||
|
let dir = match &msg {
|
||||||
|
BusMessage::Request(dir, ..) | BusMessage::Response(dir, ..) => {
|
||||||
|
trace!("redirecting message to {:?} ...", dir);
|
||||||
|
dir
|
||||||
|
}
|
||||||
|
};
|
||||||
|
match dir {
|
||||||
|
BusMessageDirection::ToCli => self.send_cli(msg).await,
|
||||||
|
BusMessageDirection::ToSupervisor => self.send_supervisor(msg).await,
|
||||||
|
BusMessageDirection::ToMetrics => self.send_metrics(msg).await,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
async fn send_cli(&self, msg: BusMessage) -> anyhow::Result<()> {
|
||||||
|
self.to_cli.send(msg).await?;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
async fn send_supervisor(&self, msg: BusMessage) -> anyhow::Result<()> {
|
||||||
|
self.to_supervisor.send(msg).await?;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
async fn send_metrics(&self, msg: BusMessage) -> anyhow::Result<()> {
|
||||||
|
self.to_metrics.send(msg).await?;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct Bus {
|
||||||
|
inner: Inner,
|
||||||
|
highway: Highway,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Bus {
|
||||||
|
pub fn new(inner: Inner, to_cli: Outter, to_supervisor: Outter, to_metrics: Outter) -> Self {
|
||||||
|
Self {
|
||||||
|
inner,
|
||||||
|
highway: Highway::new(to_cli, to_supervisor, to_metrics),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[async_trait::async_trait]
|
||||||
|
impl ProcessUnit for Bus {
|
||||||
|
async fn process(&mut self) {
|
||||||
|
loop {
|
||||||
|
while let Ok(content) = self.inner.try_recv() {
|
||||||
|
// debug!("new message to the Bus : {:?}", &content);
|
||||||
|
let msg = match content {
|
||||||
|
BusMessage::Request(direction, content_type, content) => {
|
||||||
|
trace!(
|
||||||
|
"bus has got a new Request with direction {:?} and type {:?}",
|
||||||
|
&direction,
|
||||||
|
&content_type
|
||||||
|
);
|
||||||
|
BusMessage::Request(direction, content_type, content)
|
||||||
|
}
|
||||||
|
BusMessage::Response(direction, content_type, content) => {
|
||||||
|
trace!(
|
||||||
|
"bus has got a new Response with direction {:?} and type {:?}",
|
||||||
|
&direction,
|
||||||
|
&content_type
|
||||||
|
);
|
||||||
|
BusMessage::Response(direction, content_type, content)
|
||||||
|
}
|
||||||
|
};
|
||||||
|
if let Err(er) = self.highway.send(msg).await {
|
||||||
|
error!("Cannot redirect message : {}", er);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
tokio::time::sleep(std::time::Duration::from_millis(20)).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
@ -1,11 +1,220 @@
|
||||||
use crate::options::structs::{CustomError, Files};
|
use crate::options::structs::CustomError;
|
||||||
use super::prcs::{is_active, is_frozen};
|
use crate::options::structs::Events;
|
||||||
|
use async_trait::async_trait;
|
||||||
use inotify::{EventMask, Inotify, WatchMask};
|
use inotify::{EventMask, Inotify, WatchMask};
|
||||||
use std::borrow::BorrowMut;
|
|
||||||
use std::path::Path;
|
use std::path::Path;
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use tokio::sync::mpsc;
|
use tokio::sync::mpsc::Sender;
|
||||||
use tokio::time::Duration;
|
|
||||||
|
pub mod v2 {
|
||||||
|
use super::*;
|
||||||
|
use crate::options::structs::{
|
||||||
|
FileTriggerType, FileTriggersForController as Triggers, ProcessUnit,
|
||||||
|
};
|
||||||
|
use log::{error, info, warn};
|
||||||
|
use serde::Serialize;
|
||||||
|
use std::{collections::HashMap, path::Path};
|
||||||
|
|
||||||
|
type MpscSender = Arc<Sender<Events>>;
|
||||||
|
type EventHandlers = HashMap<Arc<str>, (Triggers, MpscSender)>;
|
||||||
|
|
||||||
|
#[derive(Debug, Serialize, Clone, Copy)]
|
||||||
|
pub enum FileState {
|
||||||
|
Ok,
|
||||||
|
NotFound,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub struct FilesController {
|
||||||
|
name: Arc<str>,
|
||||||
|
path: String,
|
||||||
|
code_name: Arc<str>,
|
||||||
|
backup_file : String,
|
||||||
|
state: FileState,
|
||||||
|
watcher: Option<Inotify>,
|
||||||
|
triggers: EventHandlers,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl PartialEq for FilesController {
|
||||||
|
fn eq(&self, other: &Self) -> bool {
|
||||||
|
self.code_name == other.code_name
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl FilesController {
|
||||||
|
#[inline(always)]
|
||||||
|
pub fn new(name: &str, triggers: EventHandlers) -> FilesController {
|
||||||
|
let name: Arc<str> = Arc::from(name);
|
||||||
|
Self {
|
||||||
|
name: name.clone(),
|
||||||
|
path: String::new(),
|
||||||
|
state: FileState::Ok,
|
||||||
|
watcher: None,
|
||||||
|
triggers,
|
||||||
|
code_name: name.clone(),
|
||||||
|
backup_file: String::new(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#[inline(always)]
|
||||||
|
pub fn with_path(mut self, path: impl AsRef<Path>, backup : String) -> anyhow::Result<FilesController> {
|
||||||
|
self.path = path.as_ref().to_string_lossy().into_owned();
|
||||||
|
self.watcher = {
|
||||||
|
match create_watcher(&self.name, &self.path) {
|
||||||
|
Ok(val) => Some(val),
|
||||||
|
Err(er) => {
|
||||||
|
error!(
|
||||||
|
"Cannot create watcher for {} ({}) due to {}",
|
||||||
|
self.name, &self.path, er
|
||||||
|
);
|
||||||
|
return Err(er);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
self.code_name = Arc::from(format!("{}{}", &self.path, &self.code_name));
|
||||||
|
self.backup_file = backup;
|
||||||
|
match create_backup(&self.code_name, &self.backup_file) {
|
||||||
|
Ok(_) => info!("Backup file for {} was created ({})", &self.code_name, &self.backup_file),
|
||||||
|
Err(er) => warn!("{}. Ignoring ...", er),
|
||||||
|
}
|
||||||
|
Ok(self)
|
||||||
|
}
|
||||||
|
pub fn add_event(&mut self, file_controller: FilesController) {
|
||||||
|
for (k, v) in file_controller.triggers {
|
||||||
|
self.triggers.entry(k).or_insert(v);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
async fn trigger_on(&mut self, trigger_type: Option<FileTriggerType>) {
|
||||||
|
for (prc_name, (triggers, channel)) in &self.triggers {
|
||||||
|
let msg = match &trigger_type {
|
||||||
|
None => Events::Positive(self.code_name.clone()),
|
||||||
|
Some(event) => {
|
||||||
|
info!(
|
||||||
|
"Event on file {} ({}) : {}. Notifying `{}` ...",
|
||||||
|
&self.name, &self.path, event, &prc_name
|
||||||
|
);
|
||||||
|
event.event_from_file_trigger_controller(self.code_name.clone(), &triggers)
|
||||||
|
}
|
||||||
|
};
|
||||||
|
let _ = channel.send(msg).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
pub fn get_state(&self) -> FileState {
|
||||||
|
self.state
|
||||||
|
}
|
||||||
|
pub fn get_code_name(&self) -> Arc<str> {
|
||||||
|
self.code_name.clone()
|
||||||
|
}
|
||||||
|
pub fn get_backup_file(&self) -> String {
|
||||||
|
self.backup_file.to_string()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#[async_trait]
|
||||||
|
impl ProcessUnit for FilesController {
|
||||||
|
async fn process(&mut self) {
|
||||||
|
if let Ok(_) = check_file(&self.name, &self.path).await {
|
||||||
|
if let FileState::NotFound = self.state {
|
||||||
|
info!(
|
||||||
|
"File {} ({}) was found in determined scope. Notifying ...",
|
||||||
|
self.name, self.code_name
|
||||||
|
);
|
||||||
|
self.state = FileState::Ok;
|
||||||
|
self.trigger_on(None).await;
|
||||||
|
}
|
||||||
|
match &mut self.watcher {
|
||||||
|
Some(notify) => {
|
||||||
|
let mut buffer = [0; 128];
|
||||||
|
if let Ok(notif_events) = notify.read_events(&mut buffer) {
|
||||||
|
let (need_to_recreate, was_modifired) =
|
||||||
|
notif_events.fold((false, false), |(a, b), mask| {
|
||||||
|
(
|
||||||
|
a || mask.mask == EventMask::DELETE_SELF,
|
||||||
|
b || mask.mask == EventMask::MODIFY,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
);
|
||||||
|
if self.backup_file.is_empty() {
|
||||||
|
|
||||||
|
} else {
|
||||||
|
|
||||||
|
}
|
||||||
|
if let (mut recreate_watcher, true) = (need_to_recreate, was_modifired) {
|
||||||
|
if self.backup_file.is_empty() {
|
||||||
|
warn!("File {} ({}) was changed", self.name, &self.path);
|
||||||
|
self.trigger_on(Some(FileTriggerType::OnChange)).await;
|
||||||
|
} else {
|
||||||
|
recreate_watcher = true;
|
||||||
|
match restore_file(&self.code_name, &self.backup_file).await {
|
||||||
|
Ok(_) => info!("File {} was successfully restored", &self.code_name),
|
||||||
|
Err(er) => error!("Cannot restore file {} : {}", &self.code_name, er),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if recreate_watcher {
|
||||||
|
self.watcher = match create_watcher(&self.name, &self.path) {
|
||||||
|
Ok(notifier) => Some(notifier),
|
||||||
|
Err(er) => {
|
||||||
|
error!(
|
||||||
|
"Failed to recreate watcher for {} ({}) due to {}",
|
||||||
|
self.name, &self.path, er
|
||||||
|
);
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
None => return,
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if let FileState::Ok = self.state {
|
||||||
|
if self.backup_file.is_empty() {
|
||||||
|
warn!(
|
||||||
|
"File {} ({}) was not found in determined scope",
|
||||||
|
self.name, &self.path
|
||||||
|
);
|
||||||
|
self.state = FileState::NotFound;
|
||||||
|
self.trigger_on(Some(FileTriggerType::OnDelete)).await;
|
||||||
|
} else {
|
||||||
|
warn!(
|
||||||
|
"File {} ({}) was not found in determined scope. Restoring from backup-file ...",
|
||||||
|
self.name, &self.path
|
||||||
|
);
|
||||||
|
match restore_file(&self.code_name, &self.backup_file).await {
|
||||||
|
Err(er) => error!("Cannot restore file {} : {}", &self.code_name, er),
|
||||||
|
Ok(_) => {
|
||||||
|
info!("File {} was successfully restored", &self.code_name);
|
||||||
|
self.watcher = match create_watcher(&self.name, &self.path) {
|
||||||
|
Ok(notifier) => Some(notifier),
|
||||||
|
Err(er) => {
|
||||||
|
error!(
|
||||||
|
"Failed to recreate watcher for {} ({}) : {}",
|
||||||
|
self.name, &self.path, er
|
||||||
|
);
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
self.trigger_on(None).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn create_backup(target: &str, backup: &str) -> anyhow::Result<u64> {
|
||||||
|
return if !backup.is_empty() {
|
||||||
|
Ok(std::fs::copy(target, backup)?)
|
||||||
|
} else {
|
||||||
|
Err(anyhow::Error::msg(format!("No need to create backup-file for {}", target)))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn restore_file(target: &str, backup: &str) -> anyhow::Result<u64> {
|
||||||
|
Ok(tokio::fs::copy(backup, target).await?)
|
||||||
|
}
|
||||||
|
|
||||||
/// # Fn `create_watcher`
|
/// # Fn `create_watcher`
|
||||||
/// ## for creating watcher on file's delete | update events
|
/// ## for creating watcher on file's delete | update events
|
||||||
|
|
@ -20,112 +229,13 @@ use tokio::time::Duration;
|
||||||
///
|
///
|
||||||
/// *depends on* : -
|
/// *depends on* : -
|
||||||
///
|
///
|
||||||
pub async fn create_watcher(filename: &str, path: &str) -> Result<Inotify, std::io::Error> {
|
pub fn create_watcher(filename: &str, path: &str) -> anyhow::Result<Inotify> {
|
||||||
let src = format!("{}{}", path, filename);
|
let src = format!("{}{}", path, filename);
|
||||||
let inotify: Inotify = Inotify::init()?;
|
let inotify: Inotify = Inotify::init()?;
|
||||||
inotify.watches().add(&src, WatchMask::ALL_EVENTS)?;
|
inotify.watches().add(&src, WatchMask::ALL_EVENTS)?;
|
||||||
Ok(inotify)
|
Ok(inotify)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// # Fn `create_watcher`
|
|
||||||
/// ## for managing processes by checking dep files' states
|
|
||||||
///
|
|
||||||
/// *input* : `&str`, `&[Files]`, `Arc<mpsc::Sender<u8>>`, `Arc<tokio::sync::Mutex<Vec<Inotify>>>`
|
|
||||||
///
|
|
||||||
/// *output* : `Err` if something with dep file is wrong | `Ok(())` on successfull dep file check
|
|
||||||
///
|
|
||||||
/// *initiator* : fn `utils::running_handler`
|
|
||||||
///
|
|
||||||
/// *managing* : current process's name: &str, list of dep files : `&[Files]`, atomic ref counter on sender main channel for current process `Arc<mpsc::Sender<u8>>`, mut list of file watchers`Arc<tokio::sync::Mutex<Vec<Inotify>>>`
|
|
||||||
///
|
|
||||||
/// *depends on* : Files
|
|
||||||
///
|
|
||||||
pub async fn file_handler(
|
|
||||||
name: &str,
|
|
||||||
files: &[Files],
|
|
||||||
tx: Arc<mpsc::Sender<u8>>,
|
|
||||||
watchers: Arc<tokio::sync::Mutex<Vec<Inotify>>>,
|
|
||||||
) -> Result<(), CustomError> {
|
|
||||||
for (i, file) in files.iter().enumerate() {
|
|
||||||
// let src = format!("{}{}", file.src, file.filename);
|
|
||||||
if check_file(&file.filename, &file.src).await.is_err() {
|
|
||||||
if !is_active(name).await || is_frozen(name).await {
|
|
||||||
return Err(CustomError::Fatal);
|
|
||||||
}
|
|
||||||
match file.triggers.on_delete.as_str() {
|
|
||||||
"stay" => {
|
|
||||||
tx.send(9).await.unwrap();
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
"stop" => {
|
|
||||||
if is_active(name).await {
|
|
||||||
tx.send(1).await.unwrap();
|
|
||||||
}
|
|
||||||
return Err(CustomError::Fatal);
|
|
||||||
}
|
|
||||||
"hold" => {
|
|
||||||
if is_active(name).await {
|
|
||||||
tx.send(2).await.unwrap();
|
|
||||||
return Err(CustomError::Fatal);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
_ => {
|
|
||||||
tokio::time::sleep(Duration::from_millis(50)).await;
|
|
||||||
tx.send(101).await.unwrap();
|
|
||||||
return Err(CustomError::Fatal);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} else if is_active(name).await && !is_frozen(name).await {
|
|
||||||
let watchers = watchers.clone();
|
|
||||||
// println!("mutex: {:?}", watchers);
|
|
||||||
let mut buffer = [0; 128];
|
|
||||||
let mut mutex_guard = watchers.lock().await;
|
|
||||||
if let Some(notify) = mutex_guard.get_mut(i) {
|
|
||||||
let events = notify.read_events(&mut buffer);
|
|
||||||
// println!("{:?}", events);
|
|
||||||
if events.is_ok() {
|
|
||||||
let events: Vec<EventMask> = events
|
|
||||||
.unwrap()
|
|
||||||
.map(|mask| mask.mask)
|
|
||||||
.filter(|mask| {
|
|
||||||
*mask == EventMask::MODIFY || *mask == EventMask::DELETE_SELF
|
|
||||||
})
|
|
||||||
.collect();
|
|
||||||
for event in events {
|
|
||||||
if let EventMask::DELETE_SELF = event {
|
|
||||||
// ! warning (DELETE_SELF event) !
|
|
||||||
// println!("! warning (DELETE_SELF event) !");
|
|
||||||
// * watcher recreation after dealing with file recreation mechanism in text editors
|
|
||||||
let mutex = notify.borrow_mut();
|
|
||||||
|
|
||||||
// *mutex = create_watcher(&file.filename, &file.src).await.unwrap();
|
|
||||||
if let Ok(watcher) = create_watcher(&file.filename, &file.src).await {
|
|
||||||
*mutex = watcher;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
match file.triggers.on_change.as_str() {
|
|
||||||
"stop" => {
|
|
||||||
let _ = tx.send(7).await;
|
|
||||||
}
|
|
||||||
"restart" => {
|
|
||||||
let _ = tx.send(8).await;
|
|
||||||
}
|
|
||||||
"stay" => {
|
|
||||||
let _ = tx.send(9).await;
|
|
||||||
}
|
|
||||||
_ => {
|
|
||||||
let _ = tx.send(101).await;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
tokio::task::yield_now().await;
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Fn `check_file`
|
/// # Fn `check_file`
|
||||||
/// ## for checking existance of current file
|
/// ## for checking existance of current file
|
||||||
///
|
///
|
||||||
|
|
@ -162,12 +272,12 @@ mod files_unittests {
|
||||||
use super::*;
|
use super::*;
|
||||||
#[tokio::test]
|
#[tokio::test]
|
||||||
async fn try_to_create_watcher() {
|
async fn try_to_create_watcher() {
|
||||||
let res = create_watcher("dep-file", "./tests/examples/").await;
|
let res = create_watcher("dep-file", "./tests/examples/");
|
||||||
assert!(res.is_ok());
|
assert!(res.is_ok());
|
||||||
}
|
}
|
||||||
#[tokio::test]
|
#[tokio::test]
|
||||||
async fn try_to_create_invalid_watcher() {
|
async fn try_to_create_invalid_watcher() {
|
||||||
let res = create_watcher("invalid-file", "/path/to/the/no/dir").await;
|
let res = create_watcher("invalid-file", "/path/to/the/no/dir");
|
||||||
assert!(res.is_err());
|
assert!(res.is_err());
|
||||||
}
|
}
|
||||||
#[tokio::test]
|
#[tokio::test]
|
||||||
|
|
|
||||||
|
|
@ -1,8 +1,8 @@
|
||||||
//
|
//
|
||||||
// module needed to check host-agent health condition and to communicate with it
|
// module needed to check host-agent health condition and to communicate with it
|
||||||
//
|
//
|
||||||
|
use anyhow::{Error, Ok, Result};
|
||||||
use tokio::{io::Interest, net::UnixStream};
|
use tokio::{io::Interest, net::UnixStream};
|
||||||
use anyhow::{Ok, Result, Error};
|
|
||||||
// to kill lint bug
|
// to kill lint bug
|
||||||
#[allow(unused_imports)]
|
#[allow(unused_imports)]
|
||||||
use tokio::net::UnixListener;
|
use tokio::net::UnixListener;
|
||||||
|
|
@ -61,7 +61,7 @@ async fn ha_healthcheck(socket: &UnixStream) -> Result<(), Error> {
|
||||||
/// *depends on* : -
|
/// *depends on* : -
|
||||||
///
|
///
|
||||||
#[allow(dead_code)]
|
#[allow(dead_code)]
|
||||||
async fn ha_send_data(socket: &UnixStream, data: &str) -> Result<(), Error > {
|
async fn ha_send_data(socket: &UnixStream, data: &str) -> Result<(), Error> {
|
||||||
socket.ready(Interest::WRITABLE).await?;
|
socket.ready(Interest::WRITABLE).await?;
|
||||||
socket.writable().await?;
|
socket.writable().await?;
|
||||||
socket.try_write(data.as_bytes())?;
|
socket.try_write(data.as_bytes())?;
|
||||||
|
|
@ -91,7 +91,7 @@ mod hagent_unittets {
|
||||||
// --Result<maybe Response>
|
// --Result<maybe Response>
|
||||||
// one-shot func
|
// one-shot func
|
||||||
async fn hagent_communication_test() {
|
async fn hagent_communication_test() {
|
||||||
use crate::options::structs::{ProcessMetrics, ContainerMetrics, Metrics};
|
use crate::options::structs::{ContainerMetrics, Metrics, ProcessMetrics};
|
||||||
|
|
||||||
let procm = ProcessMetrics::new("test-prc", 15.0, 5.0);
|
let procm = ProcessMetrics::new("test-prc", 15.0, 5.0);
|
||||||
let contm = ContainerMetrics::new("test", 32.0, 12.0, vec![procm.process_name.clone()]);
|
let contm = ContainerMetrics::new("test", 32.0, 12.0, vec![procm.process_name.clone()]);
|
||||||
|
|
@ -105,10 +105,11 @@ mod hagent_unittets {
|
||||||
let sock = sock.unwrap();
|
let sock = sock.unwrap();
|
||||||
assert!(ha_healthcheck(&sock).await.is_ok());
|
assert!(ha_healthcheck(&sock).await.is_ok());
|
||||||
assert!(ha_send_data(&sock, &metrics).await.is_ok());
|
assert!(ha_send_data(&sock, &metrics).await.is_ok());
|
||||||
|
|
||||||
}
|
}
|
||||||
#[tokio::test]
|
#[tokio::test]
|
||||||
async fn open_unixsocket_test() {
|
async fn open_unixsocket_test() {
|
||||||
assert!(open_unix_socket("non/valid/socket/file.sock").await.is_err());
|
assert!(open_unix_socket("non/valid/socket/file.sock")
|
||||||
|
.await
|
||||||
|
.is_err());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -1,18 +1,27 @@
|
||||||
// submodule needed to get metrics such as
|
///! Submodule needed to get metrics such as
|
||||||
// cpu load, ram/rom load and net activity
|
///! cpu load, ram/rom load and net activity
|
||||||
|
|
||||||
// use std::sync::Mutex;
|
use crate::{
|
||||||
use std::sync::Arc;
|
options::structs::ProcessState,
|
||||||
use crate::options::structs::TrackingProcess;
|
utils::metrics::processes::{ProcessesGeneral, ProcessesQuery},
|
||||||
use sysinfo::{Process, System};
|
};
|
||||||
use tokio::join;
|
use log::warn;
|
||||||
use crate::options::structs::{ProcessMetrics, ContainerMetrics};
|
use noxis_cli::metrics_models::MetricsMode;
|
||||||
use super::get_container_id;
|
use std::{any::Any, collections::BTreeMap, sync::Arc};
|
||||||
// use pcap::{Device, Capture, Active};
|
// use chrono::Duration;
|
||||||
// use std::net::Ipv4Addr;
|
use super::prcs::v2::Pid;
|
||||||
// use anyhow::{Result, Ok};
|
use crate::options::structs::bus::{BusMessage, BusMessageContentType, BusMessageDirection};
|
||||||
|
use serde::Serialize;
|
||||||
|
use std::fmt::Debug;
|
||||||
|
use sysinfo::{Disks as DisksList, Networks, System};
|
||||||
|
// use noxis_cli::metrics_models::MetricsMode;
|
||||||
|
|
||||||
// type PacketBuffer = Arc<Mutex<Vec<PacketInfo>>>;
|
pub type MetricProcesses = Vec<ProcessExtended>;
|
||||||
|
type CoreUsage = BTreeMap<usize, CoreInfo>;
|
||||||
|
type Disks = Vec<Disk>;
|
||||||
|
type Ifaces = Vec<Network>;
|
||||||
|
type BusReciever = tokio::sync::mpsc::Receiver<BusMessage>;
|
||||||
|
type BusSender = Arc<tokio::sync::mpsc::Sender<BusMessage>>;
|
||||||
|
|
||||||
/// # Fn `init_metrics_grubber`
|
/// # Fn `init_metrics_grubber`
|
||||||
/// ## for initializing process of unstoppable grubbing metrics.
|
/// ## for initializing process of unstoppable grubbing metrics.
|
||||||
|
|
@ -27,204 +36,536 @@ use super::get_container_id;
|
||||||
///
|
///
|
||||||
/// *depends on* : -
|
/// *depends on* : -
|
||||||
///
|
///
|
||||||
#[allow(dead_code)]
|
pub async fn init_metrics_grubber(
|
||||||
pub async fn init_metrics_grubber() {
|
/* BROADCSAT LISTENER TO GET `PROCESSES` OBJ */
|
||||||
|
bus_sender: BusSender,
|
||||||
|
bus_reciever: BusReciever,
|
||||||
|
) -> anyhow::Result<()> {
|
||||||
let mut system = System::new();
|
let mut system = System::new();
|
||||||
// let mut buffer: Vec<PacketInfo> = vec![];
|
let mut disks = DisksList::new_with_refreshed_list();
|
||||||
// let shared_buf: PacketBuffer = Arc::new(Mutex::new(buffer));
|
let mut networks = Networks::new_with_refreshed_list();
|
||||||
|
// get_all_metrics(&mut system).await;
|
||||||
|
/* TODO */
|
||||||
|
let mut bus_reciever = bus_reciever;
|
||||||
|
loop {
|
||||||
|
let msg = bus_reciever.try_recv();
|
||||||
|
if let Ok(BusMessage::Request(_, _, cont)) = msg {
|
||||||
system.refresh_all();
|
system.refresh_all();
|
||||||
// let temp = String::from_utf8(get_pid("systemd").await.unwrap().stdout).unwrap();
|
disks.refresh_list();
|
||||||
// let prc = system.process(Pid::from_str(&temp).unwrap()).unwrap();
|
networks.refresh_list();
|
||||||
// prc.
|
let cont: Box<dyn Any + Send> = cont;
|
||||||
// let _ = capture_packets(shared_buf.clone()).await;
|
match cont.downcast::<MetricsMode>() {
|
||||||
|
Err(_) => {
|
||||||
|
warn!("Unrecognized Metric mode was given");
|
||||||
|
let _ = bus_sender
|
||||||
|
.send(BusMessage::Response(
|
||||||
|
BusMessageDirection::ToCli,
|
||||||
|
BusMessageContentType::Result,
|
||||||
|
Box::new(Err(anyhow::Error::msg(format!(
|
||||||
|
"Unrecognized Metric mode was given"
|
||||||
|
)))),
|
||||||
|
))
|
||||||
|
.await;
|
||||||
|
}
|
||||||
|
Ok(mode) => {
|
||||||
|
tokio::time::sleep(tokio::time::Duration::from_millis(200)).await;
|
||||||
|
let metric: Box<dyn MetricsExportable> = match *mode {
|
||||||
|
MetricsMode::Full => {
|
||||||
|
let mut refs =
|
||||||
|
get_all_metrics(&mut system, bus_sender.clone(), &disks, &networks)
|
||||||
|
.await;
|
||||||
|
if let Some(prcs) = bus_reciever.recv().await {
|
||||||
|
if let BusMessage::Response(_, _, cont) = prcs {
|
||||||
|
let cont: Box<dyn Any> = cont;
|
||||||
|
if let Ok(cont) = cont.downcast::<ProcessesQuery>() {
|
||||||
|
if let ProcessesQuery::General(info) = *cont {
|
||||||
|
refs.processes = info;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Box::new(refs)
|
||||||
|
}
|
||||||
|
MetricsMode::Host => {
|
||||||
|
Box::new(get_global_host_info(&mut system, &disks, &networks).await)
|
||||||
|
}
|
||||||
|
MetricsMode::Cpu => Box::new(get_cpu_metrics(&mut system).await),
|
||||||
|
MetricsMode::Ram => Box::new(get_ram_metrics(&mut system).await),
|
||||||
|
MetricsMode::Rom => Box::new(get_all_disks_metrics(&disks).await),
|
||||||
|
MetricsMode::Network => Box::new(get_all_ifaces_metrics(&networks).await),
|
||||||
|
// inspect processes
|
||||||
|
MetricsMode::Processes => {
|
||||||
|
todo!();
|
||||||
|
}
|
||||||
|
};
|
||||||
|
// let metric: Box<dyn BusContent> = Box::new(metric);
|
||||||
|
let metric = metric.serialze_into_output();
|
||||||
|
|
||||||
|
let _ = bus_sender
|
||||||
|
.send(BusMessage::Response(
|
||||||
|
BusMessageDirection::ToCli,
|
||||||
|
BusMessageContentType::MetricsObj,
|
||||||
|
Box::new(metric),
|
||||||
|
))
|
||||||
|
.await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} else if let Ok(BusMessage::Response(_, _, cont)) = msg {
|
||||||
|
let cont: Box<dyn Any + Send> = cont;
|
||||||
|
if let Ok(info) = cont.downcast::<ProcessesQuery>() {
|
||||||
|
if let ProcessesQuery::All(info) = *info {
|
||||||
|
let procs: Vec<_> = info
|
||||||
|
.into_iter()
|
||||||
|
.map(|prc| ProcessExtended::from_process_query_all(&mut system, prc))
|
||||||
|
.collect();
|
||||||
|
let _ = bus_sender
|
||||||
|
.send(BusMessage::Response(
|
||||||
|
BusMessageDirection::ToCli,
|
||||||
|
BusMessageContentType::Result,
|
||||||
|
Box::<anyhow::Result<String>>::new(Ok(serde_json::to_string_pretty(
|
||||||
|
&procs,
|
||||||
|
)?)),
|
||||||
|
))
|
||||||
|
.await;
|
||||||
|
} else {
|
||||||
|
let _ = bus_sender
|
||||||
|
.send(BusMessage::Response(
|
||||||
|
BusMessageDirection::ToCli,
|
||||||
|
BusMessageContentType::Result,
|
||||||
|
Box::new(Err(anyhow::Error::msg(format!(
|
||||||
|
"Unknown type was send by the Supervisor"
|
||||||
|
)))),
|
||||||
|
))
|
||||||
|
.await;
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
let _ = bus_sender
|
||||||
|
.send(BusMessage::Response(
|
||||||
|
BusMessageDirection::ToCli,
|
||||||
|
BusMessageContentType::Result,
|
||||||
|
Box::new(Err(anyhow::Error::msg(format!(
|
||||||
|
"Unknown type was send by the Supervisor"
|
||||||
|
)))),
|
||||||
|
))
|
||||||
|
.await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
tokio::time::sleep(std::time::Duration::from_millis(100)).await;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[allow(dead_code)]
|
async fn get_all_metrics(
|
||||||
#[allow(unused_variables)]
|
system: &mut System,
|
||||||
async fn gather_metrics(proc: Arc<Process>) {
|
sender: BusSender,
|
||||||
|
disks: &DisksList,
|
||||||
|
networks: &Networks,
|
||||||
|
) -> FullMetrics {
|
||||||
|
let host = get_host_info().await;
|
||||||
|
let cpu = get_cpu_metrics(system).await;
|
||||||
|
let ram = get_ram_metrics(system).await;
|
||||||
|
let disks = get_all_disks_metrics(&disks).await;
|
||||||
|
let ifaces = get_all_ifaces_metrics(&networks).await;
|
||||||
|
let prcs: Vec<ProcessesGeneral> = Vec::new();
|
||||||
|
let _ = sender
|
||||||
|
.send(BusMessage::Request(
|
||||||
|
BusMessageDirection::ToSupervisor,
|
||||||
|
BusMessageContentType::ProcessQuery,
|
||||||
|
Box::new(ProcessesQuery::QueryGeneral),
|
||||||
|
))
|
||||||
|
.await;
|
||||||
|
FullMetrics::create(host, cpu, ram, disks, ifaces, prcs)
|
||||||
}
|
}
|
||||||
|
|
||||||
// DEPRECATED : for net monitoring
|
async fn get_global_host_info(
|
||||||
// async fn capture_packets(buffer: PacketBuffer) -> Result<()> {
|
system: &mut System,
|
||||||
// let mut cap = Capture::from_device(Device::lookup()?.unwrap())?
|
disks: &DisksList,
|
||||||
// .promisc(true)
|
networks: &Networks,
|
||||||
// .open()?;
|
) -> HostGeneral {
|
||||||
|
HostGeneral {
|
||||||
|
hostname: System::host_name().unwrap_or_default(),
|
||||||
|
os: System::long_os_version().unwrap_or_default(),
|
||||||
|
kernel: System::kernel_version().unwrap_or_default(),
|
||||||
|
cpu_percentage: system.global_cpu_usage(),
|
||||||
|
ram_available: system.total_memory() - system.free_memory(),
|
||||||
|
disk_percentage: {
|
||||||
|
let total = disks
|
||||||
|
.iter()
|
||||||
|
.map(|disk| disk.available_space() * 100 / disk.total_space())
|
||||||
|
.collect::<Vec<u64>>();
|
||||||
|
total.iter().sum::<u64>() / (total.len() as u64)
|
||||||
|
},
|
||||||
|
net_stat: {
|
||||||
|
let total = networks
|
||||||
|
.iter()
|
||||||
|
.map(|(_, iface_data)| iface_data.received() + iface_data.transmitted())
|
||||||
|
.collect::<Vec<u64>>();
|
||||||
|
total.iter().sum::<u64>() / ((total.len() * 2) as u64)
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// cap.filter("not broadcast and not multicast", true)?;
|
async fn get_host_info() -> HostInfo {
|
||||||
|
HostInfo {
|
||||||
|
hostname: System::host_name().unwrap_or_default(),
|
||||||
|
os: System::long_os_version().unwrap_or_default(),
|
||||||
|
kernel: System::kernel_version().unwrap_or_default(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// while let core::result::Result::Ok(packet) = cap.next_packet() {
|
async fn get_cpu_metrics(system: &mut System) -> Cpu {
|
||||||
// if let Some((src, dst, prot)) = get_packet_info(&packet.data).await {
|
let mut buffer = CoreUsage::new();
|
||||||
// let packet_info = PacketInfo::new(String::from(prot), dst, src, packet.header.len as usize);
|
let global_usage = system.global_cpu_usage();
|
||||||
// let mut locked_buffer = buffer.lock().unwrap();
|
|
||||||
// println!("{:?}", &packet_info);
|
|
||||||
// locked_buffer.push(packet_info);
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
// Ok(())
|
|
||||||
// }
|
|
||||||
// async fn get_packet_info(data: &[u8]) -> Option<(Ipv4Addr, Ipv4Addr, &str)> {
|
|
||||||
// if data.len() >= 20 {
|
|
||||||
// let src_ip = Ipv4Addr::new(data[12], data[13], data[14], data[15]);
|
|
||||||
// let dst_ip = Ipv4Addr::new(data[16], data[17], data[18], data[19]);
|
|
||||||
// let protocol = match data[9] {
|
|
||||||
// 1 => "ICMP",
|
|
||||||
// 6 => "TCP",
|
|
||||||
// 17 => "UDP",
|
|
||||||
// _ => "Unknown",
|
|
||||||
// };
|
|
||||||
|
|
||||||
// Some((src_ip, dst_ip, protocol))
|
system.cpus().iter().enumerate().for_each(|(id, cpu)| {
|
||||||
// } else {
|
let core_info = CoreInfo {
|
||||||
// None
|
// id,
|
||||||
// }
|
brand: cpu.brand().to_string(),
|
||||||
// }
|
name: cpu.name().to_string(),
|
||||||
|
frequency: cpu.frequency(),
|
||||||
|
vendor_id: cpu.vendor_id().to_string(),
|
||||||
|
usage: cpu.cpu_usage(),
|
||||||
|
};
|
||||||
|
buffer.entry(id).or_insert(core_info);
|
||||||
|
});
|
||||||
|
|
||||||
|
Cpu {
|
||||||
|
global_usage,
|
||||||
|
usage: buffer,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// # Fn `get_all_container_metrics`
|
async fn get_ram_metrics(system: &mut System) -> Ram {
|
||||||
/// ## for gathering all container (whole system metrics)
|
Ram {
|
||||||
///
|
free_mem: system.free_memory(),
|
||||||
/// *input* : `Arc<System>`, `Arc<Vec<TrackingProcess>>`
|
free_swap: system.free_swap(),
|
||||||
///
|
total_mem: system.total_memory(),
|
||||||
/// *output* : `ContainerMetrics`
|
total_swap: system.total_swap(),
|
||||||
///
|
}
|
||||||
/// *initiator* : main thread ??
|
}
|
||||||
///
|
|
||||||
/// *managing* : ref counter to `System` object, ref counter to list of processes
|
async fn get_all_disks_metrics(disks: &DisksList) -> Disks {
|
||||||
///
|
// let disks = DisksList::new_with_refreshed_list();
|
||||||
/// *depends on* : `TrackingProcess`
|
let mut buffer = Disks::new();
|
||||||
///
|
disks.list().iter().for_each(|disk| {
|
||||||
#[allow(dead_code)]
|
let disk = Disk {
|
||||||
async fn get_all_container_metrics(sys: Arc<System>, prcs: Arc<Vec<TrackingProcess>>) -> ContainerMetrics {
|
name: disk.name().to_string_lossy().into_owned(),
|
||||||
let metrics = join!(
|
kind: disk.kind().to_string(),
|
||||||
get_cpu_metrics_container(sys.clone()),
|
fs: disk.file_system().to_string_lossy().into_owned(),
|
||||||
get_ram_metrics_container(sys.clone()),
|
mount_point: disk.mount_point().to_string_lossy().into_owned(),
|
||||||
get_subsystems(prcs.clone())
|
total_space: disk.total_space(),
|
||||||
);
|
available_space: disk.available_space(),
|
||||||
ContainerMetrics::new(
|
is_removable: disk.is_removable(),
|
||||||
&get_container_id().unwrap_or(String::from("unknown")),
|
is_readonly: disk.is_read_only(),
|
||||||
metrics.0,
|
};
|
||||||
metrics.1,
|
buffer.push(disk);
|
||||||
metrics.2
|
});
|
||||||
|
buffer
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn get_all_ifaces_metrics(networks: &Networks) -> Ifaces {
|
||||||
|
let mut ifaces = Ifaces::new();
|
||||||
|
networks.iter().for_each(|(iface_name, data)| {
|
||||||
|
let mac = data.mac_address().to_string();
|
||||||
|
let ip_addrs = data
|
||||||
|
.ip_networks()
|
||||||
|
.iter()
|
||||||
|
.map(|ipaddr| format!("{}/{}", ipaddr.addr, ipaddr.prefix))
|
||||||
|
.collect::<Vec<String>>();
|
||||||
|
|
||||||
|
let iface = Network {
|
||||||
|
iname: iface_name.to_owned(),
|
||||||
|
mac: mac,
|
||||||
|
ip_addresses: ip_addrs,
|
||||||
|
recieved: data.received(),
|
||||||
|
transmitted: data.transmitted(),
|
||||||
|
total_recieved_bytes: data.total_received(),
|
||||||
|
total_transmitted_bytes: data.total_transmitted(),
|
||||||
|
total_recieved_packets: data.total_packets_received(),
|
||||||
|
total_transmitted_packets: data.total_packets_transmitted(),
|
||||||
|
errors_on_recieved: data.errors_on_received(),
|
||||||
|
errors_on_transmitted: data.errors_on_transmitted(),
|
||||||
|
};
|
||||||
|
ifaces.push(iface);
|
||||||
|
});
|
||||||
|
ifaces
|
||||||
|
}
|
||||||
|
|
||||||
|
pub mod processes {
|
||||||
|
use crate::options::structs::ProcessState;
|
||||||
|
use crate::utils::prcs::v2::Pid;
|
||||||
|
|
||||||
|
#[derive(Debug, serde::Serialize)]
|
||||||
|
pub enum ProcessesQuery {
|
||||||
|
General(Vec<ProcessesGeneral>),
|
||||||
|
All(Vec<ProcessesAll>),
|
||||||
|
QueryGeneral,
|
||||||
|
QueryAll,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, serde::Serialize)]
|
||||||
|
pub struct ProcessesGeneral {
|
||||||
|
pub name: String,
|
||||||
|
pub state: ProcessState,
|
||||||
|
pub pid: Pid,
|
||||||
|
}
|
||||||
|
#[derive(Debug, serde::Serialize)]
|
||||||
|
pub struct ProcessesAll {
|
||||||
|
pub name: String,
|
||||||
|
pub state: ProcessState,
|
||||||
|
pub pid: Pid,
|
||||||
|
pub dependencies: deps::Dependencies,
|
||||||
|
}
|
||||||
|
|
||||||
|
pub mod deps {
|
||||||
|
use crate::options::structs::{FileTriggers, ServiceState, ServiceTriggers};
|
||||||
|
use crate::utils::files::v2::FileState;
|
||||||
|
|
||||||
|
// use super::*;
|
||||||
|
#[derive(Debug, serde::Serialize)]
|
||||||
|
pub struct FilesExtended {
|
||||||
|
pub name: String,
|
||||||
|
pub path: String,
|
||||||
|
pub status: FileState,
|
||||||
|
pub backup_file : String,
|
||||||
|
pub triggers: FileTriggers,
|
||||||
|
}
|
||||||
|
#[derive(Debug, serde::Serialize)]
|
||||||
|
pub struct ServicesExtended {
|
||||||
|
pub name: String,
|
||||||
|
pub access_name: String,
|
||||||
|
pub status: ServiceState,
|
||||||
|
pub triggers: ServiceTriggers,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, serde::Serialize)]
|
||||||
|
pub struct Dependencies {
|
||||||
|
pub files: Vec<FilesExtended>,
|
||||||
|
pub services: Vec<ServicesExtended>,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub trait MetricsExportable: Send + Sync + 'static + Debug + Any {
|
||||||
|
fn serialze_into_output(&self) -> anyhow::Result<String>;
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Debug)]
|
||||||
|
struct FullMetrics {
|
||||||
|
hostname: String,
|
||||||
|
os: String,
|
||||||
|
kernel: String,
|
||||||
|
cpu: Cpu,
|
||||||
|
ram: Ram,
|
||||||
|
disks: Disks,
|
||||||
|
networks: Ifaces,
|
||||||
|
pub processes: Vec<ProcessesGeneral>,
|
||||||
|
}
|
||||||
|
impl FullMetrics {
|
||||||
|
fn create(
|
||||||
|
host: HostInfo,
|
||||||
|
cpu: Cpu,
|
||||||
|
ram: Ram,
|
||||||
|
disks: Disks,
|
||||||
|
ifaces: Ifaces,
|
||||||
|
processes: Vec<ProcessesGeneral>,
|
||||||
|
) -> Self {
|
||||||
|
Self {
|
||||||
|
hostname: host.hostname,
|
||||||
|
os: host.os,
|
||||||
|
kernel: host.kernel,
|
||||||
|
cpu,
|
||||||
|
ram,
|
||||||
|
disks,
|
||||||
|
networks: ifaces,
|
||||||
|
processes,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl MetricsExportable for FullMetrics {
|
||||||
|
fn serialze_into_output(&self) -> anyhow::Result<String> {
|
||||||
|
Ok(serde_json::to_string_pretty(self)?)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Serialize)]
|
||||||
|
struct HostInfo {
|
||||||
|
hostname: String,
|
||||||
|
os: String,
|
||||||
|
kernel: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl MetricsExportable for HostInfo {
|
||||||
|
fn serialze_into_output(&self) -> anyhow::Result<String> {
|
||||||
|
Ok(serde_json::to_string_pretty(self)?)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Serialize)]
|
||||||
|
struct HostGeneral {
|
||||||
|
hostname: String,
|
||||||
|
os: String,
|
||||||
|
kernel: String,
|
||||||
|
cpu_percentage: f32,
|
||||||
|
ram_available: u64,
|
||||||
|
disk_percentage: u64,
|
||||||
|
net_stat: u64,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl MetricsExportable for HostGeneral {
|
||||||
|
fn serialze_into_output(&self) -> anyhow::Result<String> {
|
||||||
|
Ok(serde_json::to_string_pretty(self)?)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Debug)]
|
||||||
|
struct Cpu {
|
||||||
|
global_usage: f32,
|
||||||
|
usage: CoreUsage,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl MetricsExportable for Cpu {
|
||||||
|
fn serialze_into_output(&self) -> anyhow::Result<String> {
|
||||||
|
Ok(serde_json::to_string_pretty(self)?)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Debug)]
|
||||||
|
struct CoreInfo {
|
||||||
|
name: String,
|
||||||
|
brand: String,
|
||||||
|
frequency: u64,
|
||||||
|
vendor_id: String,
|
||||||
|
usage: f32,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Debug)]
|
||||||
|
struct Ram {
|
||||||
|
free_mem: u64,
|
||||||
|
free_swap: u64,
|
||||||
|
total_mem: u64,
|
||||||
|
total_swap: u64,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl MetricsExportable for Ram {
|
||||||
|
fn serialze_into_output(&self) -> anyhow::Result<String> {
|
||||||
|
Ok(serde_json::to_string_pretty(self)?)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Debug)]
|
||||||
|
struct Disk {
|
||||||
|
name: String,
|
||||||
|
kind: String,
|
||||||
|
fs: String,
|
||||||
|
mount_point: String,
|
||||||
|
total_space: u64,
|
||||||
|
available_space: u64,
|
||||||
|
is_removable: bool,
|
||||||
|
is_readonly: bool,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl MetricsExportable for Disks {
|
||||||
|
fn serialze_into_output(&self) -> anyhow::Result<String> {
|
||||||
|
Ok(serde_json::to_string_pretty(self)?)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// vec<Network>
|
||||||
|
#[derive(Serialize, Debug)]
|
||||||
|
struct Network {
|
||||||
|
iname: String,
|
||||||
|
mac: String,
|
||||||
|
ip_addresses: Vec<String>,
|
||||||
|
recieved: u64,
|
||||||
|
transmitted: u64,
|
||||||
|
total_recieved_bytes: u64,
|
||||||
|
total_transmitted_bytes: u64,
|
||||||
|
total_recieved_packets: u64,
|
||||||
|
total_transmitted_packets: u64,
|
||||||
|
errors_on_recieved: u64,
|
||||||
|
errors_on_transmitted: u64,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl MetricsExportable for Ifaces {
|
||||||
|
fn serialze_into_output(&self) -> anyhow::Result<String> {
|
||||||
|
Ok(serde_json::to_string_pretty(self)?)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Debug)]
|
||||||
|
pub struct ProcessExtended {
|
||||||
|
name: String,
|
||||||
|
status: ProcessState,
|
||||||
|
pid: Pid,
|
||||||
|
start_time: String,
|
||||||
|
duration: String,
|
||||||
|
dependencies: processes::deps::Dependencies,
|
||||||
|
cpu_usage: f32,
|
||||||
|
ram_usage: u64,
|
||||||
|
virtual_mem_usage: u64,
|
||||||
|
disks_usage_read_bytes: u64,
|
||||||
|
disks_usage_write_bytes: u64,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl ProcessExtended {
|
||||||
|
pub fn from_process_query_all(system: &mut System, proc: processes::ProcessesAll) -> Self {
|
||||||
|
system.refresh_processes(sysinfo::ProcessesToUpdate::All, true);
|
||||||
|
return if let Some(prc) = system.process(proc.pid.new_sysinfo_pid()) {
|
||||||
|
let disk_usage = prc.disk_usage();
|
||||||
|
let duration = chrono::Duration::new(prc.run_time() as i64, 0);
|
||||||
|
let start_time = chrono::DateTime::from_timestamp(prc.start_time() as i64, 0);
|
||||||
|
Self {
|
||||||
|
name: proc.name,
|
||||||
|
status: proc.state,
|
||||||
|
pid: proc.pid,
|
||||||
|
start_time : {
|
||||||
|
match start_time {
|
||||||
|
Some(date) => date.to_string(),
|
||||||
|
None => String::new()
|
||||||
|
}
|
||||||
|
},
|
||||||
|
duration: {
|
||||||
|
match duration {
|
||||||
|
Some(duration) => {
|
||||||
|
format!("{}:{}:{}:{}",
|
||||||
|
duration.num_days(),
|
||||||
|
duration.num_hours() % 24,
|
||||||
|
duration.num_minutes() % 60,
|
||||||
|
duration.num_seconds() % 60
|
||||||
)
|
)
|
||||||
|
},
|
||||||
|
None => String::new()
|
||||||
|
}
|
||||||
|
},
|
||||||
|
dependencies: proc.dependencies,
|
||||||
|
cpu_usage: prc.cpu_usage(),
|
||||||
|
ram_usage: prc.memory(),
|
||||||
|
virtual_mem_usage: prc.virtual_memory(),
|
||||||
|
disks_usage_read_bytes: disk_usage.read_bytes,
|
||||||
|
disks_usage_write_bytes: disk_usage.written_bytes,
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
Self {
|
||||||
|
name: proc.name,
|
||||||
|
status: proc.state,
|
||||||
|
pid: proc.pid,
|
||||||
|
start_time : String::new(),
|
||||||
|
duration: String::new(),
|
||||||
|
dependencies: proc.dependencies,
|
||||||
|
cpu_usage: 0.0,
|
||||||
|
ram_usage: 0,
|
||||||
|
virtual_mem_usage: 0,
|
||||||
|
disks_usage_read_bytes: 0,
|
||||||
|
disks_usage_write_bytes: 0,
|
||||||
|
}
|
||||||
|
};
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// # Fn `get_cpu_metrics_container`
|
impl MetricsExportable for MetricProcesses {
|
||||||
/// ## for gathering container cpu metrics
|
fn serialze_into_output(&self) -> anyhow::Result<String> {
|
||||||
///
|
Ok(serde_json::to_string_pretty(self)?)
|
||||||
/// *input* : `Arc<System>`
|
}
|
||||||
///
|
|
||||||
/// *output* : `f32`
|
|
||||||
///
|
|
||||||
/// *initiator* : main thread ??
|
|
||||||
///
|
|
||||||
/// *managing* : ref counter to `System` object
|
|
||||||
///
|
|
||||||
/// *depends on* : -
|
|
||||||
///
|
|
||||||
#[allow(dead_code)]
|
|
||||||
async fn get_cpu_metrics_container(sys: Arc<System>) -> f32 {
|
|
||||||
sys.global_cpu_usage()
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Fn `get_ram_metrics_container`
|
|
||||||
/// ## for gathering container ram metrics
|
|
||||||
///
|
|
||||||
/// *input* : `Arc<System>`
|
|
||||||
///
|
|
||||||
/// *output* : `f32`
|
|
||||||
///
|
|
||||||
/// *initiator* : main thread ??
|
|
||||||
///
|
|
||||||
/// *managing* : ref counter to `System` object
|
|
||||||
///
|
|
||||||
/// *depends on* : -
|
|
||||||
///
|
|
||||||
#[allow(dead_code)]
|
|
||||||
async fn get_ram_metrics_container(sys: Arc<System>) -> f32 {
|
|
||||||
(sys.used_memory() / sys.total_memory()) as f32 * 100.0
|
|
||||||
}
|
|
||||||
// async fn get_mem_metrics_container(sys: Arc<System>) -> f32 {
|
|
||||||
// sys.
|
|
||||||
// }
|
|
||||||
|
|
||||||
/// # Fn `get_subsystems`
|
|
||||||
/// ## for gathering info about container subsystems (processes)
|
|
||||||
///
|
|
||||||
/// *input* : `Arc<Vec<TrackingProcess>>`
|
|
||||||
///
|
|
||||||
/// *output* : `Vec<String>`
|
|
||||||
///
|
|
||||||
/// *initiator* : main thread ??
|
|
||||||
///
|
|
||||||
/// *managing* : ref counter to list of `TrackingProcess`
|
|
||||||
///
|
|
||||||
/// *depends on* : `TrackingProcess`
|
|
||||||
///
|
|
||||||
#[allow(dead_code)]
|
|
||||||
async fn get_subsystems(prcs: Arc<Vec<TrackingProcess>>) -> Vec<String> {
|
|
||||||
prcs.iter().map(|process| process.name.clone()).collect()
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Fn `get_all_metrics_process`
|
|
||||||
/// ## for gathering all process' metrics
|
|
||||||
///
|
|
||||||
/// *input* : `Arc<Process>`, `Arc<System>`
|
|
||||||
///
|
|
||||||
/// *output* : `ProcessMetrics`
|
|
||||||
///
|
|
||||||
/// *initiator* : main thread ??
|
|
||||||
///
|
|
||||||
/// *managing* : two ref counters to `Process` and `System`
|
|
||||||
///
|
|
||||||
/// *depends on* : -
|
|
||||||
///
|
|
||||||
#[allow(dead_code)]
|
|
||||||
async fn get_all_metrics_process(proc: Arc<Process>, sys: Arc<System>) -> ProcessMetrics {
|
|
||||||
let metrics = join!(
|
|
||||||
get_cpu_metrics_process(proc.clone()),
|
|
||||||
get_ram_metrics_process(proc.clone(), sys.clone())
|
|
||||||
);
|
|
||||||
ProcessMetrics::new(
|
|
||||||
proc.name().to_str().unwrap_or("unknown"),
|
|
||||||
metrics.0,
|
|
||||||
metrics.1
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Fn `get_cpu_metrics_process`
|
|
||||||
/// ## for gathering process cpu metrics
|
|
||||||
///
|
|
||||||
/// *input* : `Arc<Process>`
|
|
||||||
///
|
|
||||||
/// *output* : `f32`
|
|
||||||
///
|
|
||||||
/// *initiator* : main thread ??
|
|
||||||
///
|
|
||||||
/// *managing* : ref counter to `Process` object
|
|
||||||
///
|
|
||||||
/// *depends on* : -
|
|
||||||
///
|
|
||||||
async fn get_cpu_metrics_process(proc: Arc<Process>) -> f32 {
|
|
||||||
proc.cpu_usage()
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Fn `get_ram_metrics_process`
|
|
||||||
/// ## for gathering process ram metrics
|
|
||||||
///
|
|
||||||
/// *input* : `Arc<Process>`
|
|
||||||
///
|
|
||||||
/// *output* : `f32`
|
|
||||||
///
|
|
||||||
/// *initiator* : main thread ??
|
|
||||||
///
|
|
||||||
/// *managing* : ref counter to `Process` object
|
|
||||||
///
|
|
||||||
/// *depends on* : -
|
|
||||||
///
|
|
||||||
async fn get_ram_metrics_process(proc: Arc<Process>, sys: Arc<System>) -> f32 {
|
|
||||||
(proc.memory() as f64 / sys.total_memory() as f64) as f32 * 100.0 as f32
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
|
|
|
||||||
|
|
@ -1,9 +1,336 @@
|
||||||
use crate::options::structs::CustomError;
|
use crate::options::structs::{Events, NegativeOutcomes, ProcessState, ProcessUnit};
|
||||||
|
use async_trait::async_trait;
|
||||||
use log::{error, warn};
|
use log::{error, warn};
|
||||||
|
use serde::Serialize;
|
||||||
|
use std::collections::HashSet;
|
||||||
use std::process::{Command, Output};
|
use std::process::{Command, Output};
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
|
use tokio::sync::mpsc::Receiver as MpscReciever;
|
||||||
use tokio::time::Duration;
|
use tokio::time::Duration;
|
||||||
|
|
||||||
|
pub mod v2 {
|
||||||
|
use crate::options::structs::DependencyType;
|
||||||
|
use crate::utils::metrics::processes::ProcessesGeneral;
|
||||||
|
use log::info;
|
||||||
|
use std::path::Path;
|
||||||
|
use tokio::time::sleep;
|
||||||
|
|
||||||
|
use super::*;
|
||||||
|
|
||||||
|
#[derive(Debug, Serialize, Clone, Copy)]
|
||||||
|
pub struct Pid(u32);
|
||||||
|
|
||||||
|
impl std::fmt::Display for Pid {
|
||||||
|
fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result {
|
||||||
|
return write!(f, "{}", self.0);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Pid {
|
||||||
|
fn new() -> Self {
|
||||||
|
Pid(0)
|
||||||
|
}
|
||||||
|
#[allow(unused)]
|
||||||
|
pub fn new_sysinfo_pid(&self) -> sysinfo::Pid {
|
||||||
|
sysinfo::Pid::from_u32(self.0 as u32)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub struct ProcessesController {
|
||||||
|
pub name: Arc<str>,
|
||||||
|
pid: Pid,
|
||||||
|
bin: String,
|
||||||
|
state: ProcessState,
|
||||||
|
event_reader: MpscReciever<Events>,
|
||||||
|
negative_events: HashSet<Arc<str>>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl PartialEq for ProcessesController {
|
||||||
|
fn eq(&self, other: &Self) -> bool {
|
||||||
|
self.bin == other.bin
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl ProcessesController {
|
||||||
|
#[inline(always)]
|
||||||
|
pub fn new(name: &str, event_reader: MpscReciever<Events>) -> ProcessesController {
|
||||||
|
ProcessesController {
|
||||||
|
name: Arc::from(name),
|
||||||
|
pid: Pid::new(),
|
||||||
|
bin: String::new(),
|
||||||
|
state: ProcessState::Stopped,
|
||||||
|
event_reader,
|
||||||
|
negative_events: HashSet::new(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#[inline(always)]
|
||||||
|
pub fn with_exe(mut self, bin: impl AsRef<Path>) -> ProcessesController {
|
||||||
|
self.bin = bin.as_ref().to_string_lossy().into_owned();
|
||||||
|
self
|
||||||
|
}
|
||||||
|
pub fn get_pid(&self) -> Pid {
|
||||||
|
self.pid
|
||||||
|
}
|
||||||
|
pub fn get_state(&self) -> ProcessState {
|
||||||
|
self.state
|
||||||
|
}
|
||||||
|
async fn trigger_on(&mut self, dep_name: &str, trigger: &str, dep_type: DependencyType) {
|
||||||
|
match trigger {
|
||||||
|
"stay" => {
|
||||||
|
info!(
|
||||||
|
"Event on {} `{}` for {}. Ignoring ...",
|
||||||
|
dep_type, dep_name, self.name
|
||||||
|
);
|
||||||
|
}
|
||||||
|
"stop" => {
|
||||||
|
if is_active(&self.name).await {
|
||||||
|
info!(
|
||||||
|
"Event on {} `{}` for {}. Stopping ...",
|
||||||
|
dep_type, dep_name, self.name
|
||||||
|
);
|
||||||
|
match terminate_process(&self.name).await {
|
||||||
|
Ok(_) => {
|
||||||
|
info!("Process {} was stopped ...", &self.name);
|
||||||
|
self.state = ProcessState::Stopped;
|
||||||
|
self.pid = Pid::new();
|
||||||
|
}
|
||||||
|
Err(er) => {
|
||||||
|
error!("Cannot stop process {} : {}", self.name, er);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
"user-stop" => {
|
||||||
|
if is_active(&self.name).await {
|
||||||
|
info!(
|
||||||
|
"Event on {} `{}` for {}. Stopping ...",
|
||||||
|
dep_type, "User Stop Call", self.name
|
||||||
|
);
|
||||||
|
match terminate_process(&self.name).await {
|
||||||
|
Ok(_) => {
|
||||||
|
info!("Process {} was forcefully stopped ...", &self.name);
|
||||||
|
self.state = ProcessState::StoppedByCli;
|
||||||
|
self.pid = Pid::new();
|
||||||
|
}
|
||||||
|
Err(er) => {
|
||||||
|
error!("Cannot forcefully stop process {} : {}", self.name, er);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
"user-hold" => {
|
||||||
|
if is_active(&self.name).await {
|
||||||
|
info!(
|
||||||
|
"Event on {} `{}` for {}. Stopping ...",
|
||||||
|
dep_type, "User Hold Call", self.name
|
||||||
|
);
|
||||||
|
match freeze_process(&self.name).await {
|
||||||
|
Ok(_) => {
|
||||||
|
info!("Process {} was forcefully frozen ...", &self.name);
|
||||||
|
self.state = ProcessState::HoldingByCli;
|
||||||
|
}
|
||||||
|
Err(er) => {
|
||||||
|
error!("Cannot forcefully freeze process {} : {}", self.name, er);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
"hold" => {
|
||||||
|
if !is_frozen(&self.name).await {
|
||||||
|
info!(
|
||||||
|
"Event on {} `{}` for {}. Freezing ...",
|
||||||
|
dep_type, dep_name, self.name
|
||||||
|
);
|
||||||
|
match freeze_process(&self.name).await {
|
||||||
|
Ok(_) => {
|
||||||
|
info!("Process {} was frozen ...", &self.name);
|
||||||
|
self.state = ProcessState::Holding;
|
||||||
|
}
|
||||||
|
Err(er) => {
|
||||||
|
error!("Cannot freeze process {} : {}", self.name, er);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
"restart" => {
|
||||||
|
info!(
|
||||||
|
"Event on {} `{}` for {}. Restarting ...",
|
||||||
|
dep_type, dep_name, self.name
|
||||||
|
);
|
||||||
|
let pid = restart_process(&self.name, &self.bin).await;
|
||||||
|
sleep(Duration::from_millis(100)).await;
|
||||||
|
if let Ok(pid) = pid {
|
||||||
|
self.pid = Pid(pid);
|
||||||
|
info!("{}: New PID - {}", self.name, self.pid);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
_ => error!(
|
||||||
|
"Impermissible trigger in file-trigger for {}. Ignoring event ...",
|
||||||
|
self.name
|
||||||
|
),
|
||||||
|
}
|
||||||
|
tokio::time::sleep(Duration::from_micros(100)).await;
|
||||||
|
}
|
||||||
|
#[allow(unused)]
|
||||||
|
pub async fn stop_by_user_call(&mut self) -> anyhow::Result<()> {
|
||||||
|
terminate_process(&self.name).await?;
|
||||||
|
warn!("Process {} was stopped by user call ...", self.name);
|
||||||
|
self.state = ProcessState::StoppedByCli;
|
||||||
|
self.pid = Pid::new();
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
#[allow(unused)]
|
||||||
|
pub async fn freeze_by_user_call(&mut self) -> anyhow::Result<()> {
|
||||||
|
freeze_process(&self.name).await?;
|
||||||
|
warn!("Process {} was frozen by user call ...", self.name);
|
||||||
|
self.state = ProcessState::HoldingByCli;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
#[allow(unused)]
|
||||||
|
pub async fn start_by_user_call(&mut self) -> anyhow::Result<()> {
|
||||||
|
if self.negative_events.is_empty() {
|
||||||
|
let pid = start_process(&self.name, &self.bin).await?;
|
||||||
|
warn!("Process {} was started by user call ...", self.name);
|
||||||
|
self.state = ProcessState::Pending;
|
||||||
|
self.pid = Pid(pid);
|
||||||
|
return Ok(());
|
||||||
|
} else {
|
||||||
|
warn!("Attempt to start process {} by user call was stopped due to existance of negative incidents ...", self.name);
|
||||||
|
return Err(anyhow::Error::msg(
|
||||||
|
format!("Attempt to start process {} by user call was stopped due to existance of negative incidents ...", self.name)
|
||||||
|
));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#[allow(unused)]
|
||||||
|
pub async fn unfreeze_by_user_call(&mut self) -> anyhow::Result<()> {
|
||||||
|
if self.negative_events.is_empty() {
|
||||||
|
unfreeze_process(&self.name).await?;
|
||||||
|
warn!("Process {} was unfrozen by user call ...", self.name);
|
||||||
|
self.state = ProcessState::Pending;
|
||||||
|
Ok(())
|
||||||
|
} else {
|
||||||
|
warn!("Attempt to unfreeze process {} by user call was stopped due to existance of negative incidents ...", self.name);
|
||||||
|
return Err(anyhow::Error::msg(
|
||||||
|
format!("Attempt to unfreeze process {} by user call was stopped due to existance of negative incidents ...", self.name)
|
||||||
|
));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#[allow(unused)]
|
||||||
|
pub async fn restart_by_user_call(&mut self) -> anyhow::Result<()> {
|
||||||
|
let pid = restart_process(&self.name, &self.bin).await?;
|
||||||
|
warn!("Process {} was restarted by user call ...", self.name);
|
||||||
|
self.pid = Pid(pid);
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn get_general_info(&self) -> ProcessesGeneral {
|
||||||
|
ProcessesGeneral {
|
||||||
|
name: self.name.to_string(),
|
||||||
|
state: self.state,
|
||||||
|
pid: self.pid,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[async_trait]
|
||||||
|
impl ProcessUnit for ProcessesController {
|
||||||
|
async fn process(&mut self) {
|
||||||
|
if self.negative_events.len() == 0 {
|
||||||
|
let conditions = (is_active(&self.name).await, is_frozen(&self.name).await);
|
||||||
|
let state = &self.state;
|
||||||
|
match (state, conditions) {
|
||||||
|
(ProcessState::Holding, (_, _)) => {
|
||||||
|
info!(
|
||||||
|
"No negative dependecies events on {} frozen process. Unfreezing ...",
|
||||||
|
self.name
|
||||||
|
);
|
||||||
|
if let Err(er) = unfreeze_process(&self.name).await {
|
||||||
|
if er.to_string().contains("already") {
|
||||||
|
self.state = ProcessState::Pending;
|
||||||
|
} else {
|
||||||
|
error!("Cannot unfreeze process {} : {}", self.name, er);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
self.state = ProcessState::Pending;
|
||||||
|
info!("Process {} was unfreezed", &self.name);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
(ProcessState::Stopped, (_, _)) => {
|
||||||
|
info!(
|
||||||
|
"No negative dependecies events on stopped {} process. Starting ...",
|
||||||
|
self.name
|
||||||
|
);
|
||||||
|
match start_process(&self.name, &self.bin).await {
|
||||||
|
Ok(pid) => {
|
||||||
|
self.state = ProcessState::Pending;
|
||||||
|
self.pid = Pid(pid);
|
||||||
|
info!("{}: New PID - {}", self.name, self.pid);
|
||||||
|
}
|
||||||
|
Err(er) => {
|
||||||
|
if er.to_string().contains("already") {
|
||||||
|
self.state = ProcessState::Pending;
|
||||||
|
} else {
|
||||||
|
error!("Cannot start process {} : {}", self.name, er);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
(ProcessState::Pending, (false, false)) => {
|
||||||
|
info!(
|
||||||
|
"{} process was impermissibly stopped. Starting ...",
|
||||||
|
self.name
|
||||||
|
);
|
||||||
|
match start_process(&self.name, &self.bin).await {
|
||||||
|
Ok(pid) => {
|
||||||
|
self.state = ProcessState::Pending;
|
||||||
|
self.pid = Pid(pid);
|
||||||
|
info!("{}: New PID - {}", self.name, self.pid);
|
||||||
|
}
|
||||||
|
Err(er) => {
|
||||||
|
error!("Cannot start process {} : {}", self.name, er);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
(ProcessState::Pending, (true, true)) => {
|
||||||
|
info!(
|
||||||
|
"No negative dependecies events on {} process. Unfreezing ...",
|
||||||
|
self.name
|
||||||
|
);
|
||||||
|
if let Err(er) = unfreeze_process(&self.name).await {
|
||||||
|
error!("Cannot unfreeze process {} : {}", self.name, er);
|
||||||
|
} else {
|
||||||
|
self.state = ProcessState::Pending;
|
||||||
|
info!("Process {} was unfreezed", &self.name);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
_ => {}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
while let Ok(event) = self.event_reader.try_recv() {
|
||||||
|
match event {
|
||||||
|
Events::Positive(target) => {
|
||||||
|
if self.negative_events.contains(&target) {
|
||||||
|
self.negative_events.remove(&target);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Events::Negative(event) => match event {
|
||||||
|
NegativeOutcomes::FileWasChanged(target, dep_type, trigger)
|
||||||
|
| NegativeOutcomes::FileWasMovedOrDeleted(target, dep_type, trigger)
|
||||||
|
| NegativeOutcomes::ServiceIsUnreachable(target, dep_type, trigger) => {
|
||||||
|
if !self.negative_events.contains(&target) {
|
||||||
|
self.negative_events.insert(target.clone());
|
||||||
|
|
||||||
|
self.trigger_on(&target, &trigger, dep_type).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// # Fn `get_pid`
|
/// # Fn `get_pid`
|
||||||
/// ## for initializing process of unstoppable grubbing metrics.
|
/// ## for initializing process of unstoppable grubbing metrics.
|
||||||
///
|
///
|
||||||
|
|
@ -116,14 +443,15 @@ pub async fn is_frozen(name: &str) -> bool {
|
||||||
///
|
///
|
||||||
/// *depends on* : -
|
/// *depends on* : -
|
||||||
///
|
///
|
||||||
pub async fn terminate_process(name: &str) {
|
pub async fn terminate_process(name: &str) -> anyhow::Result<()> {
|
||||||
let _ = Command::new("pkill")
|
if !is_active(name).await {
|
||||||
.arg(name)
|
return Err(anyhow::Error::msg(format!(
|
||||||
.output()
|
"Process {} is already stopped",
|
||||||
.unwrap_or_else(|_| {
|
name
|
||||||
error!("Failed to execute command 'pkill'");
|
)));
|
||||||
std::process::exit(101);
|
}
|
||||||
});
|
let _ = Command::new("pkill").arg(name).output()?;
|
||||||
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
/// # Fn `terminate_process`
|
/// # Fn `terminate_process`
|
||||||
|
|
@ -139,14 +467,9 @@ pub async fn terminate_process(name: &str) {
|
||||||
///
|
///
|
||||||
/// *depends on* : -
|
/// *depends on* : -
|
||||||
///
|
///
|
||||||
pub async fn freeze_process(name: &str) {
|
pub async fn freeze_process(name: &str) -> anyhow::Result<()> {
|
||||||
let _ = Command::new("pkill")
|
let _ = Command::new("pkill").args(["-STOP", name]).output()?;
|
||||||
.args(["-STOP", name])
|
Ok(())
|
||||||
.output()
|
|
||||||
.unwrap_or_else(|_| {
|
|
||||||
error!("Failed to freeze process");
|
|
||||||
std::process::exit(101);
|
|
||||||
});
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// # Fn `unfreeze_process`
|
/// # Fn `unfreeze_process`
|
||||||
|
|
@ -162,14 +485,9 @@ pub async fn freeze_process(name: &str) {
|
||||||
///
|
///
|
||||||
/// *depends on* : -
|
/// *depends on* : -
|
||||||
///
|
///
|
||||||
pub async fn unfreeze_process(name: &str) {
|
pub async fn unfreeze_process(name: &str) -> anyhow::Result<()> {
|
||||||
let _ = Command::new("pkill")
|
let _ = Command::new("pkill").args(["-CONT", name]).output()?;
|
||||||
.args(["-CONT", name])
|
Ok(())
|
||||||
.output()
|
|
||||||
.unwrap_or_else(|_| {
|
|
||||||
error!("Failed to unfreeze process");
|
|
||||||
std::process::exit(101);
|
|
||||||
});
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// # Fn `restart_process`
|
/// # Fn `restart_process`
|
||||||
|
|
@ -185,8 +503,8 @@ pub async fn unfreeze_process(name: &str) {
|
||||||
///
|
///
|
||||||
/// *depends on* : fn `start_process`, fn `terminate_process`
|
/// *depends on* : fn `start_process`, fn `terminate_process`
|
||||||
///
|
///
|
||||||
pub async fn restart_process(name: &str, path: &str) -> Result<(), CustomError> {
|
pub async fn restart_process(name: &str, path: &str) -> anyhow::Result<u32> {
|
||||||
terminate_process(name).await;
|
terminate_process(name).await?;
|
||||||
tokio::time::sleep(Duration::from_millis(100)).await;
|
tokio::time::sleep(Duration::from_millis(100)).await;
|
||||||
start_process(name, path).await
|
start_process(name, path).await
|
||||||
}
|
}
|
||||||
|
|
@ -204,20 +522,26 @@ pub async fn restart_process(name: &str, path: &str) -> Result<(), CustomError>
|
||||||
///
|
///
|
||||||
/// *depends on* : -
|
/// *depends on* : -
|
||||||
///
|
///
|
||||||
pub async fn start_process(name: &str, path: &str) -> Result<(), CustomError> {
|
pub async fn start_process(name: &str, path: &str) -> anyhow::Result<u32> {
|
||||||
// let runsh = format!("{} {}", "exec", path);
|
if is_active(name).await {
|
||||||
|
return Err(anyhow::Error::msg(format!(
|
||||||
|
"Process {} is already running",
|
||||||
|
name
|
||||||
|
)));
|
||||||
|
}
|
||||||
let mut command = Command::new(path);
|
let mut command = Command::new(path);
|
||||||
// command.arg(path);
|
// command.arg(path);
|
||||||
|
|
||||||
match command.spawn() {
|
match command.spawn() {
|
||||||
Ok(_) => {
|
Ok(child) => {
|
||||||
|
let pid = child.id();
|
||||||
warn!("Process {} is running now!", name);
|
warn!("Process {} is running now!", name);
|
||||||
Ok(())
|
Ok(pid)
|
||||||
}
|
|
||||||
Err(er) => {
|
|
||||||
println!("{:?}", er);
|
|
||||||
Err(CustomError::Fatal)
|
|
||||||
}
|
}
|
||||||
|
Err(er) => Err(anyhow::Error::msg(format!(
|
||||||
|
"Cannot start process {} : {}",
|
||||||
|
name, er
|
||||||
|
))),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -236,8 +560,7 @@ mod process_unittests {
|
||||||
// let _ = std::io::stdout().write_all(b"");
|
// let _ = std::io::stdout().write_all(b"");
|
||||||
let res1 = start_process("restart-prc", "./tests/examples/restart-prc").await;
|
let res1 = start_process("restart-prc", "./tests/examples/restart-prc").await;
|
||||||
assert!(res1.is_ok());
|
assert!(res1.is_ok());
|
||||||
let res2 =
|
let res2 = restart_process("restart-prc", "./tests/examples/restart-prc").await;
|
||||||
restart_process("restart-prc", "./tests/examples/restart-prc").await;
|
|
||||||
assert!(res2.is_ok());
|
assert!(res2.is_ok());
|
||||||
let _ = terminate_process("restart-prc").await;
|
let _ = terminate_process("restart-prc").await;
|
||||||
let res3 = is_active("restart-prc").await;
|
let res3 = is_active("restart-prc").await;
|
||||||
|
|
@ -264,6 +587,7 @@ mod process_unittests {
|
||||||
let res1 = start_process("freeze-check", "./tests/examples/freeze-check").await;
|
let res1 = start_process("freeze-check", "./tests/examples/freeze-check").await;
|
||||||
assert!(res1.is_ok());
|
assert!(res1.is_ok());
|
||||||
assert!(!is_frozen("freeze-check").await);
|
assert!(!is_frozen("freeze-check").await);
|
||||||
|
let _ = terminate_process("freeze-check").await;
|
||||||
}
|
}
|
||||||
#[tokio::test]
|
#[tokio::test]
|
||||||
async fn pidof_active_process() {
|
async fn pidof_active_process() {
|
||||||
|
|
|
||||||
|
|
@ -1,173 +1,291 @@
|
||||||
use crate::options::structs::{CustomError, Services};
|
use async_trait::async_trait;
|
||||||
use super::prcs::{is_active, is_frozen};
|
use futures::future::Future;
|
||||||
use log::{error, warn};
|
use log::{error, warn};
|
||||||
use std::net::{TcpStream, ToSocketAddrs};
|
use std::net::ToSocketAddrs;
|
||||||
|
use std::pin::Pin;
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use tokio::sync::mpsc;
|
use tokio::sync::mpsc::Sender;
|
||||||
use tokio::time::{Duration, Instant};
|
use tokio::time::Duration;
|
||||||
|
|
||||||
/// # Fn `service_handler`
|
pub mod v2 {
|
||||||
/// ## function to realize mechanism of current process' dep services monitoring
|
use futures::FutureExt;
|
||||||
///
|
use log::info;
|
||||||
/// *input* : `&str`, `&Vec<Services>`, `Arc<mpsc::Sender<u8>>`
|
|
||||||
///
|
|
||||||
/// *output* : ()
|
|
||||||
///
|
|
||||||
/// *initiator* : fn `utils::running_handler`
|
|
||||||
///
|
|
||||||
/// *managing* : process name, ref of vec of dep services, ref counter to managing channel writer
|
|
||||||
///
|
|
||||||
/// *depends on* : fn `check_service`, fn `utils::prcs::is_active`, fn `utils::prcs::is_frozen`, fn `looped_service_connecting`
|
|
||||||
///
|
|
||||||
pub async fn service_handler(
|
|
||||||
name: &str,
|
|
||||||
services: &Vec<Services>,
|
|
||||||
tx: Arc<mpsc::Sender<u8>>,
|
|
||||||
) -> Result<(), CustomError> {
|
|
||||||
// println!("service daemon on {}", name);
|
|
||||||
for serv in services {
|
|
||||||
if check_service(&serv.hostname, &serv.port).await.is_err() {
|
|
||||||
if !is_active(name).await || is_frozen(name).await {
|
|
||||||
return Err(CustomError::Fatal);
|
|
||||||
}
|
|
||||||
error!(
|
|
||||||
"Service {}:{} is unreachable for process {}",
|
|
||||||
&serv.hostname, &serv.port, &name
|
|
||||||
);
|
|
||||||
match serv.triggers.on_lost.as_str() {
|
|
||||||
"stay" => {
|
|
||||||
tx.send(4).await.unwrap();
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
"stop" => {
|
|
||||||
if looped_service_connecting(name, serv).await.is_err() {
|
|
||||||
tx.send(5).await.unwrap();
|
|
||||||
tokio::task::yield_now().await;
|
|
||||||
return Err(CustomError::Fatal);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
"hold" => {
|
|
||||||
// if is_frozen(name).await {
|
|
||||||
// return Err(CustomError::Fatal);
|
|
||||||
// }
|
|
||||||
if looped_service_connecting(name, serv).await.is_err() {
|
|
||||||
tx.send(6).await.unwrap();
|
|
||||||
tokio::task::yield_now().await;
|
|
||||||
return Err(CustomError::Fatal);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
_ => {
|
|
||||||
tx.send(101).await.unwrap();
|
|
||||||
return Err(CustomError::Fatal);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
tokio::time::sleep(Duration::from_millis(100)).await;
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Fn `looped_service_connecting`
|
use crate::options::structs::{Events, ProcessUnit, ServiceState, Triggers};
|
||||||
/// ## for service's state check in loop (with delay and restriction of attempts)
|
|
||||||
///
|
use super::*;
|
||||||
/// *input* : `&str`, `&Services`
|
use std::collections::{BTreeMap, HashMap, VecDeque};
|
||||||
///
|
|
||||||
/// *output* : Ok(()) if service now available | Err(er) if still not
|
type MpscSender = Arc<Sender<Events>>;
|
||||||
///
|
// type EventHandlers<'a> = Vec<MpscSender<Events<'a>>>;
|
||||||
/// *initiator* : fn `service_handler`
|
type EventHandlers = HashMap<Arc<str>, (Triggers, MpscSender)>;
|
||||||
///
|
// type wrapper for service wait queue
|
||||||
/// *managing* : process name, current service struct
|
type ConnectionQueue = BTreeMap<u32, VecDeque<Arc<str>>>;
|
||||||
///
|
|
||||||
/// *depends on* : fn `check_service`
|
#[derive(Debug)]
|
||||||
///
|
pub struct ServicesController {
|
||||||
async fn looped_service_connecting(name: &str, serv: &Services) -> Result<(), CustomError> {
|
// i.e. yandex.ru
|
||||||
if serv.triggers.wait == 0 {
|
#[allow(unused)]
|
||||||
loop {
|
name: String,
|
||||||
tokio::time::sleep(Duration::from_secs(serv.triggers.delay.into())).await;
|
// i.e. yandex.ru:443
|
||||||
warn!(
|
access_url: Arc<str>,
|
||||||
"Attempting to connect from {} process to {}:{}",
|
// "OK" or "Unavailable"
|
||||||
&name, &serv.hostname, &serv.port
|
state: ServiceState,
|
||||||
);
|
// btree map with key as max wait time and it's key to hashmap
|
||||||
match check_service(&serv.hostname, &serv.port).await {
|
config: ConnectionQueue,
|
||||||
Ok(_) => {
|
// Map of processes with their (trigger and mpsc sender)
|
||||||
log::info!(
|
event_registrator: EventHandlers,
|
||||||
"Successfully connected to {} from {} process!",
|
}
|
||||||
&serv.hostname,
|
|
||||||
&name
|
impl PartialEq for ServicesController {
|
||||||
);
|
fn eq(&self, other: &Self) -> bool {
|
||||||
|
self.access_url == other.access_url
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl ServicesController {
|
||||||
|
#[inline(always)]
|
||||||
|
pub fn new() -> ServicesController {
|
||||||
|
ServicesController {
|
||||||
|
name: String::new(),
|
||||||
|
access_url: Arc::from(String::new()),
|
||||||
|
state: ServiceState::Ok,
|
||||||
|
config: ConnectionQueue::new(),
|
||||||
|
event_registrator: EventHandlers::new(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#[inline(always)]
|
||||||
|
pub fn with_access_name(mut self, hostname: &str, access_url: &str) -> ServicesController {
|
||||||
|
self.name = hostname.to_string();
|
||||||
|
self.access_url = Arc::from(access_url);
|
||||||
|
self
|
||||||
|
}
|
||||||
|
#[inline(always)]
|
||||||
|
pub fn with_params(
|
||||||
|
mut self,
|
||||||
|
conn_queue: ConnectionQueue,
|
||||||
|
event_reg: EventHandlers,
|
||||||
|
) -> ServicesController {
|
||||||
|
self.config = conn_queue;
|
||||||
|
self.event_registrator = event_reg;
|
||||||
|
self
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn get_access_url(hostname: &str, port: Option<&u32>) -> String {
|
||||||
|
format!(
|
||||||
|
"{}{}",
|
||||||
|
hostname,
|
||||||
|
port.map_or_else(|| "".to_string(), |p| format!(":{}", p))
|
||||||
|
)
|
||||||
|
}
|
||||||
|
pub fn get_state(&self) -> ServiceState {
|
||||||
|
self.state
|
||||||
|
}
|
||||||
|
pub fn add_process(&mut self, proc_name: &str, trigger: Triggers, sender: MpscSender) {
|
||||||
|
let proc_name: Arc<str> = Arc::from(proc_name);
|
||||||
|
// queue add
|
||||||
|
if let Triggers::Service { wait, .. } = trigger {
|
||||||
|
self.config
|
||||||
|
.entry(wait)
|
||||||
|
.and_modify(|el| el.push_back(proc_name.clone()))
|
||||||
|
.or_insert({
|
||||||
|
let mut temp = VecDeque::new();
|
||||||
|
temp.push_back(proc_name.clone());
|
||||||
|
temp
|
||||||
|
});
|
||||||
|
}
|
||||||
|
// event add
|
||||||
|
self.event_registrator
|
||||||
|
.entry(proc_name)
|
||||||
|
.or_insert((trigger, sender));
|
||||||
|
}
|
||||||
|
async fn check_state(&self) -> anyhow::Result<()> {
|
||||||
|
let url = self.access_url.clone();
|
||||||
|
let resolve_future = tokio::task::spawn_blocking(move || url.to_socket_addrs());
|
||||||
|
let addrs: Vec<_> =
|
||||||
|
match tokio::time::timeout(Duration::from_secs(1), resolve_future).await {
|
||||||
|
Ok(Ok(addrs)) => addrs?.collect(),
|
||||||
|
Ok(Err(er)) => return Err(er.into()),
|
||||||
|
Err(_) => return Err(anyhow::Error::msg("DNS resolution timeout")),
|
||||||
|
};
|
||||||
|
|
||||||
|
if addrs.is_empty() {
|
||||||
|
return Err(anyhow::Error::msg("No addresses resolved"));
|
||||||
|
}
|
||||||
|
|
||||||
|
let tasks: Vec<_> = addrs
|
||||||
|
.into_iter()
|
||||||
|
.map(|addr| async move {
|
||||||
|
match tokio::time::timeout(
|
||||||
|
Duration::from_secs(2),
|
||||||
|
tokio::net::TcpStream::connect(&addr),
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
{
|
||||||
|
Ok(Ok(_)) => Some(addr),
|
||||||
|
_ => None,
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.collect();
|
||||||
|
let mut any_success = false;
|
||||||
|
for task in futures::future::join_all(tasks).await {
|
||||||
|
if task.is_some() {
|
||||||
|
any_success = true;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
Err(_) => {
|
|
||||||
tokio::task::yield_now().await;
|
|
||||||
}
|
}
|
||||||
|
if !any_success {
|
||||||
|
return Err(anyhow::Error::msg(format!(
|
||||||
|
"No access to service `{}`",
|
||||||
|
&self.access_url
|
||||||
|
)));
|
||||||
}
|
}
|
||||||
}
|
|
||||||
Ok(())
|
|
||||||
} else {
|
|
||||||
let start = Instant::now();
|
|
||||||
while start.elapsed().as_secs() < serv.triggers.wait.into() {
|
|
||||||
tokio::time::sleep(Duration::from_secs(serv.triggers.delay.into())).await;
|
|
||||||
warn!(
|
|
||||||
"Attempting to connect from {} process to {}:{}",
|
|
||||||
&name, &serv.hostname, &serv.port
|
|
||||||
);
|
|
||||||
match check_service(&serv.hostname, &serv.port).await {
|
|
||||||
Ok(_) => {
|
|
||||||
log::info!(
|
|
||||||
"Successfully connected to {} from {} process!",
|
|
||||||
&serv.hostname,
|
|
||||||
&name
|
|
||||||
);
|
|
||||||
return Ok(());
|
|
||||||
}
|
|
||||||
Err(_) => {
|
|
||||||
tokio::task::yield_now().await;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Err(CustomError::Fatal)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Fn `check_service`
|
|
||||||
/// ## for check current service's availiability
|
|
||||||
///
|
|
||||||
/// *input* : `&str`, `&u32`
|
|
||||||
///
|
|
||||||
/// *output* : Ok(()) if service now available | Err(er) if still not
|
|
||||||
///
|
|
||||||
/// *initiator* : fn `service_handler`, fn `looped_service_connecting`
|
|
||||||
///
|
|
||||||
/// *managing* : hostname, port
|
|
||||||
///
|
|
||||||
/// *depends on* : -
|
|
||||||
///
|
|
||||||
// ! have to be rewritten
|
|
||||||
// todo: rewrite use
|
|
||||||
async fn check_service(hostname: &str, port: &u32) -> Result<(), CustomError> {
|
|
||||||
let addr = format!("{}:{}", hostname, port);
|
|
||||||
|
|
||||||
match addr.to_socket_addrs() {
|
|
||||||
Ok(mut addrs) => {
|
|
||||||
if addrs.any(|a| TcpStream::connect_timeout(&a, Duration::new(1, 0)).is_ok()) {
|
|
||||||
Ok(())
|
Ok(())
|
||||||
|
}
|
||||||
|
async fn trigger_on(&mut self) {
|
||||||
|
match self.state {
|
||||||
|
ServiceState::Ok => {
|
||||||
|
let futures: Vec<Pin<Box<dyn Future<Output = ()> + Send>>> = self
|
||||||
|
.event_registrator
|
||||||
|
.iter()
|
||||||
|
.map(|(prc, (_, sender_opt))| (prc, (self.access_url.clone(), sender_opt)))
|
||||||
|
.map(|(prc, (serv, sender_opt))| async move {
|
||||||
|
info!("Notifying process {} ...", prc);
|
||||||
|
let _ = sender_opt.send(Events::Positive(serv.clone())).await;
|
||||||
|
})
|
||||||
|
.map(|fut| fut.boxed())
|
||||||
|
.collect();
|
||||||
|
|
||||||
|
futures::future::join_all(futures).await;
|
||||||
|
}
|
||||||
|
ServiceState::Unavailable => {
|
||||||
|
// looped check and notifying
|
||||||
|
self.looped_check().await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
async fn looped_check(self: &mut Self) {
|
||||||
|
let longest = self.config.last_entry().unwrap();
|
||||||
|
let longest = longest.key();
|
||||||
|
let mut interapter = tokio::time::interval(tokio::time::Duration::from_secs(1));
|
||||||
|
let timer = tokio::time::Instant::now();
|
||||||
|
let mut attempt: u32 = 1;
|
||||||
|
let access_url = Arc::new(self.access_url.clone());
|
||||||
|
|
||||||
|
if let Err(_) = tokio::time::timeout(
|
||||||
|
tokio::time::Duration::from_secs((longest + 1) as u64),
|
||||||
|
async {
|
||||||
|
// let access_url = access_url.clone();
|
||||||
|
loop {
|
||||||
|
interapter.tick().await;
|
||||||
|
info!(
|
||||||
|
"Trying to connect to {} (attempt: {}) ...",
|
||||||
|
&access_url, attempt
|
||||||
|
);
|
||||||
|
attempt += 1;
|
||||||
|
|
||||||
|
let state_check_result = self.check_state().await;
|
||||||
|
|
||||||
|
if state_check_result.is_ok() {
|
||||||
|
info!("Connection to {} is `OK` now", &access_url);
|
||||||
|
self.state = ServiceState::Ok;
|
||||||
|
let futures: Vec<Pin<Box<dyn Future<Output = ()> + Send>>> = self
|
||||||
|
.event_registrator
|
||||||
|
.iter()
|
||||||
|
.map(|(prc, (_, sender_opt))| {
|
||||||
|
(prc, (self.access_url.clone(), sender_opt))
|
||||||
|
})
|
||||||
|
.map(|(prc, (serv, sender_opt))| async move {
|
||||||
|
info!("Notifying process {} ...", prc);
|
||||||
|
let _ = sender_opt.send(Events::Positive(serv.clone()));
|
||||||
|
})
|
||||||
|
.map(|fut| fut.boxed())
|
||||||
|
.collect();
|
||||||
|
|
||||||
|
futures::future::join_all(futures).await;
|
||||||
|
break;
|
||||||
} else {
|
} else {
|
||||||
Err(CustomError::Fatal)
|
let now = timer.elapsed();
|
||||||
|
|
||||||
|
let iterator = self
|
||||||
|
.config
|
||||||
|
.iter()
|
||||||
|
.filter(|(&wait, _)| {
|
||||||
|
tokio::time::Duration::from_secs(wait as u64) <= now
|
||||||
|
})
|
||||||
|
.flat_map(|(_, a)| a.iter().cloned())
|
||||||
|
.collect::<VecDeque<Arc<str>>>();
|
||||||
|
|
||||||
|
for name in iterator {
|
||||||
|
let proc_name = name.to_string();
|
||||||
|
info!("Trying to notify process `{}` ...", &proc_name);
|
||||||
|
let sender_opt =
|
||||||
|
self.event_registrator.get(&name).map(|(trigger, sender)| {
|
||||||
|
(
|
||||||
|
trigger
|
||||||
|
.to_service_negative_event(self.access_url.clone()),
|
||||||
|
sender,
|
||||||
|
)
|
||||||
|
});
|
||||||
|
|
||||||
|
if let Some((tr, tx)) = sender_opt {
|
||||||
|
let _ = tx.send(tr.unwrap()).await;
|
||||||
|
} else {
|
||||||
|
error!(
|
||||||
|
"Cannot find {} channel sender in {} service",
|
||||||
|
name.clone(),
|
||||||
|
&self.access_url
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
{
|
||||||
|
info!("Timeout of establishing connection to {}. ", &access_url);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
pub fn get_arc_access_url(&self) -> Arc<str> {
|
||||||
|
self.access_url.clone()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#[async_trait]
|
||||||
|
impl ProcessUnit for ServicesController {
|
||||||
|
async fn process(&mut self) {
|
||||||
|
// check_service(hostname, port)
|
||||||
|
let current_state = self.check_state().await;
|
||||||
|
match (&self.state, current_state) {
|
||||||
|
(ServiceState::Unavailable, Ok(_)) => {
|
||||||
|
warn!("Connection with `{}` service was established. Notifying {} process(es) ...", &self.access_url, &self.event_registrator.len());
|
||||||
|
self.state = ServiceState::Ok;
|
||||||
|
self.trigger_on().await;
|
||||||
|
}
|
||||||
|
(ServiceState::Ok, Err(_)) => {
|
||||||
|
warn!("Unreachable for connection service `{}`. Initializing reconnect mechanism ...", &self.access_url);
|
||||||
|
self.state = ServiceState::Unavailable;
|
||||||
|
self.trigger_on().await;
|
||||||
|
}
|
||||||
|
(ServiceState::Unavailable, Err(_)) => {
|
||||||
|
warn!("Service {} is still unreachable", &self.access_url)
|
||||||
|
}
|
||||||
|
_ => { /* DEAD END WITH NO INTEREST */ }
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
Err(_) => Err(CustomError::Fatal),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod service_unittests {
|
mod service_unittests {
|
||||||
use super::check_service;
|
// use super::check_service;
|
||||||
#[tokio::test]
|
// #[tokio::test]
|
||||||
async fn check_available_service() {
|
// async fn check_available_service() {
|
||||||
assert!(check_service("ya.ru", &443).await.is_ok());
|
// assert!(check_service("ya.ru", &443).await.is_ok());
|
||||||
}
|
// }
|
||||||
#[tokio::test]
|
// #[tokio::test]
|
||||||
async fn check_unavailable_service() {
|
// async fn check_unavailable_service() {
|
||||||
assert!(check_service("unavailable.service", &1111).await.is_err());
|
// assert!(check_service("unavailable.service", &1111).await.is_err());
|
||||||
}
|
// }
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -1,5 +1,7 @@
|
||||||
{
|
{
|
||||||
"dateOfCreation": "1",
|
"dateOfCreation": "1",
|
||||||
"configServer": "",
|
"processes": []
|
||||||
|
}
|
||||||
|
,
|
||||||
"processes": []
|
"processes": []
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -1,5 +1,7 @@
|
||||||
{
|
{
|
||||||
"dateOfCreation": "1",
|
"dateOfCreation": "1",
|
||||||
"configServer": "",
|
"processes": []
|
||||||
|
}
|
||||||
|
,
|
||||||
"processes": []
|
"processes": []
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -1,6 +1,5 @@
|
||||||
{
|
{
|
||||||
"dateOfCreation": "1721381809103",
|
"dateOfCreation": "1721381809103",
|
||||||
"configServer" : "localhost",
|
|
||||||
"processes": [
|
"processes": [
|
||||||
{
|
{
|
||||||
"name": "temp-process",
|
"name": "temp-process",
|
||||||
|
|
@ -12,7 +11,8 @@
|
||||||
"src": "/home/vladislav/web/runner-rs/examples/",
|
"src": "/home/vladislav/web/runner-rs/examples/",
|
||||||
"triggers": {
|
"triggers": {
|
||||||
"onDelete": "hold",
|
"onDelete": "hold",
|
||||||
"onChange": "stop"
|
"onChange": "stop",
|
||||||
|
"doRestore" : true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
],
|
],
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue