Compare commits
58 Commits
better-ind
...
0395d11e98
| Author | SHA1 | Date | |
|---|---|---|---|
| 0395d11e98 | |||
| 05e7b8075c | |||
| b857412151 | |||
| 7bb3602ab8 | |||
| 78b80c2169 | |||
| 0876f4e4f0 | |||
| 6ac0e095a3 | |||
| ff2efc0a66 | |||
|
|
f180cc4c80 | ||
| 8cc7adf196 | |||
| a1ab5d40fb | |||
| 6c92dd24f7 | |||
| c805d7e018 | |||
| b33615b969 | |||
| 0f59f29ac4 | |||
| 361f240762 | |||
| 57c3b01e66 | |||
| 94ddf027dd | |||
| 06a2be4496 | |||
| e2a09efdee | |||
| d36c574590 | |||
| 2618441de3 | |||
| da6610c625 | |||
| e956772593 | |||
| 27c51e0ec5 | |||
| bfca9cf163 | |||
| 597dcbc848 | |||
| cd3ea6fc10 | |||
| a53e8552e9 | |||
| 89eb88d10e | |||
| 72fb05b5cc | |||
| 6685b05cc5 | |||
| 07116eb8a6 | |||
| 3f34f868eb | |||
| bc6f7336d2 | |||
| 01da8631da | |||
| 67b5c2df07 | |||
| 1eaf63417b | |||
| 5e7803d2ba | |||
| 9a610661c7 | |||
| 70a65ed5d0 | |||
| 26e8e386b9 | |||
| 19cb7f73bc | |||
| 84f38974b1 | |||
| 7d027bcfc4 | |||
| d1a274b705 | |||
| b43ca7c740 | |||
| 2a6a233fb2 | |||
|
|
610ce84280 | ||
|
|
8bb4a9d3f6 | ||
|
|
67f3a23071 | ||
| d86970f81b | |||
| 623a3f019b | |||
| fd8f643a8f | |||
|
|
bd214f8fb8 | ||
| f0ed548755 | |||
| 1de96027a1 | |||
| 0812937a67 |
@@ -9,7 +9,7 @@ jobs:
|
|||||||
check:
|
check:
|
||||||
runs-on: docker
|
runs-on: docker
|
||||||
container:
|
container:
|
||||||
image: hub.nationtech.io/harmony/harmony_composer:latest@sha256:eb0406fcb95c63df9b7c4b19bc50ad7914dd8232ce98e9c9abef628e07c69386
|
image: hub.nationtech.io/harmony/harmony_composer:latest
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout code
|
- name: Checkout code
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ on:
|
|||||||
jobs:
|
jobs:
|
||||||
package_harmony_composer:
|
package_harmony_composer:
|
||||||
container:
|
container:
|
||||||
image: hub.nationtech.io/harmony/harmony_composer:latest@sha256:eb0406fcb95c63df9b7c4b19bc50ad7914dd8232ce98e9c9abef628e07c69386
|
image: hub.nationtech.io/harmony/harmony_composer:latest
|
||||||
runs-on: dind
|
runs-on: dind
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout code
|
- name: Checkout code
|
||||||
|
|||||||
1071
Cargo.lock
generated
1071
Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
11
Cargo.toml
11
Cargo.toml
@@ -12,6 +12,9 @@ members = [
|
|||||||
"harmony_cli",
|
"harmony_cli",
|
||||||
"k3d",
|
"k3d",
|
||||||
"harmony_composer",
|
"harmony_composer",
|
||||||
|
"harmony_inventory_agent",
|
||||||
|
"harmony_secret_derive",
|
||||||
|
"harmony_secret", "adr/agent_discovery/mdns",
|
||||||
]
|
]
|
||||||
|
|
||||||
[workspace.package]
|
[workspace.package]
|
||||||
@@ -20,7 +23,7 @@ readme = "README.md"
|
|||||||
license = "GNU AGPL v3"
|
license = "GNU AGPL v3"
|
||||||
|
|
||||||
[workspace.dependencies]
|
[workspace.dependencies]
|
||||||
log = "0.4"
|
log = { version = "0.4", features = ["kv"] }
|
||||||
env_logger = "0.11"
|
env_logger = "0.11"
|
||||||
derive-new = "0.7"
|
derive-new = "0.7"
|
||||||
async-trait = "0.1"
|
async-trait = "0.1"
|
||||||
@@ -53,6 +56,12 @@ chrono = "0.4"
|
|||||||
similar = "2"
|
similar = "2"
|
||||||
uuid = { version = "1.11", features = ["v4", "fast-rng", "macro-diagnostics"] }
|
uuid = { version = "1.11", features = ["v4", "fast-rng", "macro-diagnostics"] }
|
||||||
pretty_assertions = "1.4.1"
|
pretty_assertions = "1.4.1"
|
||||||
|
tempfile = "3.20.0"
|
||||||
bollard = "0.19.1"
|
bollard = "0.19.1"
|
||||||
base64 = "0.22.1"
|
base64 = "0.22.1"
|
||||||
tar = "0.4.44"
|
tar = "0.4.44"
|
||||||
|
lazy_static = "1.5.0"
|
||||||
|
directories = "6.0.0"
|
||||||
|
thiserror = "2.0.14"
|
||||||
|
serde = { version = "1.0.209", features = ["derive", "rc"] }
|
||||||
|
serde_json = "1.0.127"
|
||||||
|
|||||||
17
adr/agent_discovery/mdns/Cargo.toml
Normal file
17
adr/agent_discovery/mdns/Cargo.toml
Normal file
@@ -0,0 +1,17 @@
|
|||||||
|
[package]
|
||||||
|
name = "mdns"
|
||||||
|
edition = "2024"
|
||||||
|
version.workspace = true
|
||||||
|
readme.workspace = true
|
||||||
|
license.workspace = true
|
||||||
|
|
||||||
|
[dependencies]
|
||||||
|
mdns-sd = "0.14"
|
||||||
|
tokio = { version = "1", features = ["full"] }
|
||||||
|
futures = "0.3"
|
||||||
|
dmidecode = "0.2" # For getting the motherboard ID on the agent
|
||||||
|
log.workspace=true
|
||||||
|
env_logger.workspace=true
|
||||||
|
clap = { version = "4.5.46", features = ["derive"] }
|
||||||
|
get_if_addrs = "0.5.3"
|
||||||
|
local-ip-address = "0.6.5"
|
||||||
60
adr/agent_discovery/mdns/src/advertise.rs
Normal file
60
adr/agent_discovery/mdns/src/advertise.rs
Normal file
@@ -0,0 +1,60 @@
|
|||||||
|
// harmony-agent/src/main.rs
|
||||||
|
|
||||||
|
use log::info;
|
||||||
|
use mdns_sd::{ServiceDaemon, ServiceInfo};
|
||||||
|
use std::collections::HashMap;
|
||||||
|
|
||||||
|
use crate::SERVICE_TYPE;
|
||||||
|
|
||||||
|
// The service we are advertising.
|
||||||
|
const SERVICE_PORT: u16 = 43210; // A port for the service. It needs one, even if unused.
|
||||||
|
|
||||||
|
pub async fn advertise() {
|
||||||
|
info!("Starting Harmony Agent...");
|
||||||
|
|
||||||
|
// Get a unique ID for this machine.
|
||||||
|
let motherboard_id = "some motherboard id";
|
||||||
|
let instance_name = format!("harmony-agent-{}", motherboard_id);
|
||||||
|
info!("This agent's instance name: {}", instance_name);
|
||||||
|
info!("Advertising with ID: {}", motherboard_id);
|
||||||
|
|
||||||
|
// Create a new mDNS daemon.
|
||||||
|
let mdns = ServiceDaemon::new().expect("Failed to create mDNS daemon");
|
||||||
|
|
||||||
|
// Create a TXT record HashMap to hold our metadata.
|
||||||
|
let mut properties = HashMap::new();
|
||||||
|
properties.insert("id".to_string(), motherboard_id.to_string());
|
||||||
|
properties.insert("version".to_string(), "1.0".to_string());
|
||||||
|
|
||||||
|
// Create the service information.
|
||||||
|
// The instance name should be unique on the network.
|
||||||
|
let local_ip = local_ip_address::local_ip().unwrap();
|
||||||
|
let service_info = ServiceInfo::new(
|
||||||
|
SERVICE_TYPE,
|
||||||
|
&instance_name,
|
||||||
|
"harmony-host.local.", // A hostname for the service
|
||||||
|
local_ip,
|
||||||
|
// "0.0.0.0",
|
||||||
|
SERVICE_PORT,
|
||||||
|
Some(properties),
|
||||||
|
)
|
||||||
|
.expect("Failed to create service info");
|
||||||
|
|
||||||
|
// Register our service with the daemon.
|
||||||
|
mdns.register(service_info)
|
||||||
|
.expect("Failed to register service");
|
||||||
|
|
||||||
|
info!(
|
||||||
|
"Service '{}' registered and now being advertised.",
|
||||||
|
instance_name
|
||||||
|
);
|
||||||
|
info!("Agent is running. Press Ctrl+C to exit.");
|
||||||
|
|
||||||
|
for iface in get_if_addrs::get_if_addrs().unwrap() {
|
||||||
|
println!("{:#?}", iface);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Keep the agent running indefinitely.
|
||||||
|
tokio::signal::ctrl_c().await.unwrap();
|
||||||
|
info!("Shutting down agent.");
|
||||||
|
}
|
||||||
110
adr/agent_discovery/mdns/src/discover.rs
Normal file
110
adr/agent_discovery/mdns/src/discover.rs
Normal file
@@ -0,0 +1,110 @@
|
|||||||
|
use log::debug;
|
||||||
|
use mdns_sd::{ServiceDaemon, ServiceEvent};
|
||||||
|
|
||||||
|
use crate::SERVICE_TYPE;
|
||||||
|
|
||||||
|
pub async fn discover() {
|
||||||
|
println!("Starting Harmony Master and browsing for agents...");
|
||||||
|
|
||||||
|
// Create a new mDNS daemon.
|
||||||
|
let mdns = ServiceDaemon::new().expect("Failed to create mDNS daemon");
|
||||||
|
|
||||||
|
// Start browsing for the service type.
|
||||||
|
// The receiver will be a stream of events.
|
||||||
|
let receiver = mdns.browse(SERVICE_TYPE).expect("Failed to browse");
|
||||||
|
|
||||||
|
println!(
|
||||||
|
"Listening for mDNS events for '{}'. Press Ctrl+C to exit.",
|
||||||
|
SERVICE_TYPE
|
||||||
|
);
|
||||||
|
|
||||||
|
std::thread::spawn(move || {
|
||||||
|
while let Ok(event) = receiver.recv() {
|
||||||
|
match event {
|
||||||
|
ServiceEvent::ServiceData(resolved) => {
|
||||||
|
println!("Resolved a new service: {}", resolved.fullname);
|
||||||
|
}
|
||||||
|
other_event => {
|
||||||
|
println!("Received other event: {:?}", &other_event);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
// Gracefully shutdown the daemon.
|
||||||
|
std::thread::sleep(std::time::Duration::from_secs(1000000));
|
||||||
|
mdns.shutdown().unwrap();
|
||||||
|
|
||||||
|
// Process events as they come in.
|
||||||
|
// while let Ok(event) = receiver.recv_async().await {
|
||||||
|
// debug!("Received event {event:?}");
|
||||||
|
// // match event {
|
||||||
|
// // ServiceEvent::ServiceFound(svc_type, fullname) => {
|
||||||
|
// // println!("\n--- Agent Discovered ---");
|
||||||
|
// // println!(" Service Name: {}", fullname());
|
||||||
|
// // // You can now resolve this service to get its IP, port, and TXT records
|
||||||
|
// // // The resolve operation is a separate network call.
|
||||||
|
// // let receiver = mdns.browse(info.get_fullname()).unwrap();
|
||||||
|
// // if let Ok(resolve_event) = receiver.recv_timeout(Duration::from_secs(2)) {
|
||||||
|
// // if let ServiceEvent::ServiceResolved(info) = resolve_event {
|
||||||
|
// // let ip = info.get_addresses().iter().next().unwrap();
|
||||||
|
// // let port = info.get_port();
|
||||||
|
// // let motherboard_id = info.get_property("id").map_or("N/A", |v| v.val_str());
|
||||||
|
// //
|
||||||
|
// // println!(" IP: {}:{}", ip, port);
|
||||||
|
// // println!(" Motherboard ID: {}", motherboard_id);
|
||||||
|
// // println!("------------------------");
|
||||||
|
// //
|
||||||
|
// // // TODO: Add this agent to your central list of discovered hosts.
|
||||||
|
// // }
|
||||||
|
// // } else {
|
||||||
|
// // println!("Could not resolve service '{}' in time.", info.get_fullname());
|
||||||
|
// // }
|
||||||
|
// // }
|
||||||
|
// // ServiceEvent::ServiceRemoved(info) => {
|
||||||
|
// // println!("\n--- Agent Removed ---");
|
||||||
|
// // println!(" Service Name: {}", info.get_fullname());
|
||||||
|
// // println!("---------------------");
|
||||||
|
// // // TODO: Remove this agent from your list.
|
||||||
|
// // }
|
||||||
|
// // _ => {
|
||||||
|
// // // We don't care about other event types for this example
|
||||||
|
// // }
|
||||||
|
// // }
|
||||||
|
// }
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn discover_example() {
|
||||||
|
use mdns_sd::{ServiceDaemon, ServiceEvent};
|
||||||
|
|
||||||
|
// Create a daemon
|
||||||
|
let mdns = ServiceDaemon::new().expect("Failed to create daemon");
|
||||||
|
|
||||||
|
// Use recently added `ServiceEvent::ServiceData`.
|
||||||
|
mdns.use_service_data(true)
|
||||||
|
.expect("Failed to use ServiceData");
|
||||||
|
|
||||||
|
// Browse for a service type.
|
||||||
|
let service_type = "_mdns-sd-my-test._udp.local.";
|
||||||
|
let receiver = mdns.browse(service_type).expect("Failed to browse");
|
||||||
|
|
||||||
|
// Receive the browse events in sync or async. Here is
|
||||||
|
// an example of using a thread. Users can call `receiver.recv_async().await`
|
||||||
|
// if running in async environment.
|
||||||
|
std::thread::spawn(move || {
|
||||||
|
while let Ok(event) = receiver.recv() {
|
||||||
|
match event {
|
||||||
|
ServiceEvent::ServiceData(resolved) => {
|
||||||
|
println!("Resolved a new service: {}", resolved.fullname);
|
||||||
|
}
|
||||||
|
other_event => {
|
||||||
|
println!("Received other event: {:?}", &other_event);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
// Gracefully shutdown the daemon.
|
||||||
|
std::thread::sleep(std::time::Duration::from_secs(1));
|
||||||
|
mdns.shutdown().unwrap();
|
||||||
|
}
|
||||||
31
adr/agent_discovery/mdns/src/main.rs
Normal file
31
adr/agent_discovery/mdns/src/main.rs
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
use clap::{Parser, ValueEnum};
|
||||||
|
|
||||||
|
mod advertise;
|
||||||
|
mod discover;
|
||||||
|
|
||||||
|
#[derive(Parser, Debug)]
|
||||||
|
#[command(version, about, long_about = None)]
|
||||||
|
struct Args {
|
||||||
|
#[arg(value_enum)]
|
||||||
|
profile: Profiles,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Copy, Clone, Debug, PartialEq, Eq, PartialOrd, Ord, ValueEnum)]
|
||||||
|
enum Profiles {
|
||||||
|
Advertise,
|
||||||
|
Discover,
|
||||||
|
}
|
||||||
|
|
||||||
|
// The service type we are looking for.
|
||||||
|
const SERVICE_TYPE: &str = "_harmony._tcp.local.";
|
||||||
|
|
||||||
|
#[tokio::main]
|
||||||
|
async fn main() {
|
||||||
|
env_logger::init();
|
||||||
|
let args = Args::parse();
|
||||||
|
|
||||||
|
match args.profile {
|
||||||
|
Profiles::Advertise => advertise::advertise().await,
|
||||||
|
Profiles::Discover => discover::discover().await,
|
||||||
|
}
|
||||||
|
}
|
||||||
1
check.sh
1
check.sh
@@ -1,6 +1,7 @@
|
|||||||
#!/bin/sh
|
#!/bin/sh
|
||||||
set -e
|
set -e
|
||||||
|
|
||||||
|
rustc --version
|
||||||
cargo check --all-targets --all-features --keep-going
|
cargo check --all-targets --all-features --keep-going
|
||||||
cargo fmt --check
|
cargo fmt --check
|
||||||
cargo clippy
|
cargo clippy
|
||||||
|
|||||||
8
data/pxe/okd/README.md
Normal file
8
data/pxe/okd/README.md
Normal file
@@ -0,0 +1,8 @@
|
|||||||
|
Here lies all the data files required for an OKD cluster PXE boot setup.
|
||||||
|
|
||||||
|
This inclues ISO files, binary boot files, ipxe, etc.
|
||||||
|
|
||||||
|
TODO as of august 2025 :
|
||||||
|
|
||||||
|
- `harmony_inventory_agent` should be downloaded from official releases, this embedded version is practical for now though
|
||||||
|
- The cluster ssh key should be generated and handled by harmony with the private key saved in a secret store
|
||||||
9
data/pxe/okd/http_files/.gitattributes
vendored
Normal file
9
data/pxe/okd/http_files/.gitattributes
vendored
Normal file
@@ -0,0 +1,9 @@
|
|||||||
|
harmony_inventory_agent filter=lfs diff=lfs merge=lfs -text
|
||||||
|
os filter=lfs diff=lfs merge=lfs -text
|
||||||
|
os/centos-stream-9 filter=lfs diff=lfs merge=lfs -text
|
||||||
|
os/centos-stream-9/images filter=lfs diff=lfs merge=lfs -text
|
||||||
|
os/centos-stream-9/initrd.img filter=lfs diff=lfs merge=lfs -text
|
||||||
|
os/centos-stream-9/vmlinuz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
os/centos-stream-9/images/efiboot.img filter=lfs diff=lfs merge=lfs -text
|
||||||
|
os/centos-stream-9/images/install.img filter=lfs diff=lfs merge=lfs -text
|
||||||
|
os/centos-stream-9/images/pxeboot filter=lfs diff=lfs merge=lfs -text
|
||||||
1
data/pxe/okd/http_files/cluster_ssh_key.pub
Normal file
1
data/pxe/okd/http_files/cluster_ssh_key.pub
Normal file
@@ -0,0 +1 @@
|
|||||||
|
ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBx6bDylvC68cVpjKfEFtLQJ/dOFi6PVS2vsIOqPDJIc jeangab@liliane2
|
||||||
BIN
data/pxe/okd/http_files/harmony_inventory_agent
(Stored with Git LFS)
Executable file
BIN
data/pxe/okd/http_files/harmony_inventory_agent
(Stored with Git LFS)
Executable file
Binary file not shown.
BIN
data/pxe/okd/http_files/os/centos-stream-9/images/efiboot.img
(Stored with Git LFS)
Normal file
BIN
data/pxe/okd/http_files/os/centos-stream-9/images/efiboot.img
(Stored with Git LFS)
Normal file
Binary file not shown.
BIN
data/pxe/okd/http_files/os/centos-stream-9/images/install.img
(Stored with Git LFS)
Normal file
BIN
data/pxe/okd/http_files/os/centos-stream-9/images/install.img
(Stored with Git LFS)
Normal file
Binary file not shown.
Binary file not shown.
BIN
data/pxe/okd/http_files/os/centos-stream-9/images/pxeboot/vmlinuz
Executable file
BIN
data/pxe/okd/http_files/os/centos-stream-9/images/pxeboot/vmlinuz
Executable file
Binary file not shown.
BIN
data/pxe/okd/http_files/os/centos-stream-9/initrd.img
(Stored with Git LFS)
Normal file
BIN
data/pxe/okd/http_files/os/centos-stream-9/initrd.img
(Stored with Git LFS)
Normal file
Binary file not shown.
BIN
data/pxe/okd/http_files/os/centos-stream-9/vmlinuz
(Stored with Git LFS)
Executable file
BIN
data/pxe/okd/http_files/os/centos-stream-9/vmlinuz
(Stored with Git LFS)
Executable file
Binary file not shown.
BIN
data/pxe/okd/tftpboot/ipxe.efi
Normal file
BIN
data/pxe/okd/tftpboot/ipxe.efi
Normal file
Binary file not shown.
BIN
data/pxe/okd/tftpboot/undionly.kpxe
Normal file
BIN
data/pxe/okd/tftpboot/undionly.kpxe
Normal file
Binary file not shown.
108
docs/pxe_test/README.md
Normal file
108
docs/pxe_test/README.md
Normal file
@@ -0,0 +1,108 @@
|
|||||||
|
# OPNsense PXE Lab Environment
|
||||||
|
|
||||||
|
This project contains a script to automatically set up a virtual lab environment for testing PXE boot services managed by an OPNsense firewall.
|
||||||
|
|
||||||
|
## Overview
|
||||||
|
|
||||||
|
The `pxe_vm_lab_setup.sh` script will create the following resources using libvirt/KVM:
|
||||||
|
|
||||||
|
1. **A Virtual Network**: An isolated network named `harmonylan` (`virbr1`) for the lab.
|
||||||
|
2. **Two Virtual Machines**:
|
||||||
|
* `opnsense-pxe`: A firewall VM that will act as the gateway and PXE server.
|
||||||
|
* `pxe-node-1`: A client VM configured to boot from the network.
|
||||||
|
|
||||||
|
## Prerequisites
|
||||||
|
|
||||||
|
Ensure you have the following software installed on your Arch Linux host:
|
||||||
|
|
||||||
|
* `libvirt`
|
||||||
|
* `qemu`
|
||||||
|
* `virt-install` (from the `virt-install` package)
|
||||||
|
* `curl`
|
||||||
|
* `bzip2`
|
||||||
|
|
||||||
|
## Usage
|
||||||
|
|
||||||
|
### 1. Create the Environment
|
||||||
|
|
||||||
|
Run the `up` command to download the necessary images and create the network and VMs.
|
||||||
|
|
||||||
|
```bash
|
||||||
|
sudo ./pxe_vm_lab_setup.sh up
|
||||||
|
```
|
||||||
|
|
||||||
|
### 2. Install and Configure OPNsense
|
||||||
|
|
||||||
|
The OPNsense VM is created but the OS needs to be installed manually via the console.
|
||||||
|
|
||||||
|
1. **Connect to the VM console**:
|
||||||
|
```bash
|
||||||
|
sudo virsh console opnsense-pxe
|
||||||
|
```
|
||||||
|
|
||||||
|
2. **Log in as the installer**:
|
||||||
|
* Username: `installer`
|
||||||
|
* Password: `opnsense`
|
||||||
|
|
||||||
|
3. **Follow the on-screen installation wizard**. When prompted to assign network interfaces (`WAN` and `LAN`):
|
||||||
|
* Find the MAC address for the `harmonylan` interface by running this command in another terminal:
|
||||||
|
```bash
|
||||||
|
virsh domiflist opnsense-pxe
|
||||||
|
# Example output:
|
||||||
|
# Interface Type Source Model MAC
|
||||||
|
# ---------------------------------------------------------
|
||||||
|
# vnet18 network default virtio 52:54:00:b5:c4:6d
|
||||||
|
# vnet19 network harmonylan virtio 52:54:00:21:f9:ba
|
||||||
|
```
|
||||||
|
* Assign the interface connected to `harmonylan` (e.g., `vtnet1` with MAC `52:54:00:21:f9:ba`) as your **LAN**.
|
||||||
|
* Assign the other interface as your **WAN**.
|
||||||
|
|
||||||
|
4. After the installation is complete, **shut down** the VM from the console menu.
|
||||||
|
|
||||||
|
5. **Detach the installation media** by editing the VM's configuration:
|
||||||
|
```bash
|
||||||
|
sudo virsh edit opnsense-pxe
|
||||||
|
```
|
||||||
|
Find and **delete** the entire `<disk>` block corresponding to the `.img` file (the one with `<target ... bus='usb'/>`).
|
||||||
|
|
||||||
|
6. **Start the VM** to boot into the newly installed system:
|
||||||
|
```bash
|
||||||
|
sudo virsh start opnsense-pxe
|
||||||
|
```
|
||||||
|
|
||||||
|
### 3. Connect to OPNsense from Your Host
|
||||||
|
|
||||||
|
To configure OPNsense, you need to connect your host to the `harmonylan` network.
|
||||||
|
|
||||||
|
1. By default, OPNsense configures its LAN interface with the IP `192.168.1.1`.
|
||||||
|
2. Assign a compatible IP address to your host's `virbr1` bridge interface:
|
||||||
|
```bash
|
||||||
|
sudo ip addr add 192.168.1.5/24 dev virbr1
|
||||||
|
```
|
||||||
|
3. You can now access the OPNsense VM from your host:
|
||||||
|
* **SSH**: `ssh root@192.168.1.1` (password: `opnsense`)
|
||||||
|
* **Web UI**: `https://192.168.1.1`
|
||||||
|
|
||||||
|
### 4. Configure PXE Services with Harmony
|
||||||
|
|
||||||
|
With connectivity established, you can now use Harmony to configure the OPNsense firewall for PXE booting. Point your Harmony OPNsense scores to the firewall using these details:
|
||||||
|
|
||||||
|
* **Hostname/IP**: `192.168.1.1`
|
||||||
|
* **Credentials**: `root` / `opnsense`
|
||||||
|
|
||||||
|
### 5. Boot the PXE Client
|
||||||
|
|
||||||
|
Once your Harmony configuration has been applied and OPNsense is serving DHCP/TFTP, start the client VM. It will automatically attempt to boot from the network.
|
||||||
|
|
||||||
|
```bash
|
||||||
|
sudo virsh start pxe-node-1
|
||||||
|
sudo virsh console pxe-node-1
|
||||||
|
```
|
||||||
|
|
||||||
|
## Cleanup
|
||||||
|
|
||||||
|
To destroy all VMs and networks created by the script, run the `clean` command:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
sudo ./pxe_vm_lab_setup.sh clean
|
||||||
|
```
|
||||||
191
docs/pxe_test/pxe_vm_lab_setup.sh
Executable file
191
docs/pxe_test/pxe_vm_lab_setup.sh
Executable file
@@ -0,0 +1,191 @@
|
|||||||
|
#!/usr/bin/env bash
|
||||||
|
set -euo pipefail
|
||||||
|
|
||||||
|
# --- Configuration ---
|
||||||
|
LAB_DIR="/var/lib/harmony_pxe_test"
|
||||||
|
IMG_DIR="${LAB_DIR}/images"
|
||||||
|
STATE_DIR="${LAB_DIR}/state"
|
||||||
|
VM_OPN="opnsense-pxe"
|
||||||
|
VM_PXE="pxe-node-1"
|
||||||
|
NET_HARMONYLAN="harmonylan"
|
||||||
|
|
||||||
|
# Network settings for the isolated LAN
|
||||||
|
VLAN_CIDR="192.168.150.0/24"
|
||||||
|
VLAN_GW="192.168.150.1"
|
||||||
|
VLAN_MASK="255.255.255.0"
|
||||||
|
|
||||||
|
# VM Specifications
|
||||||
|
RAM_OPN="2048"
|
||||||
|
VCPUS_OPN="2"
|
||||||
|
DISK_OPN_GB="10"
|
||||||
|
OS_VARIANT_OPN="freebsd14.0" # Updated to a more recent FreeBSD variant
|
||||||
|
|
||||||
|
RAM_PXE="4096"
|
||||||
|
VCPUS_PXE="2"
|
||||||
|
DISK_PXE_GB="40"
|
||||||
|
OS_VARIANT_LINUX="centos-stream9"
|
||||||
|
|
||||||
|
OPN_IMG_URL="https://mirror.ams1.nl.leaseweb.net/opnsense/releases/25.7/OPNsense-25.7-serial-amd64.img.bz2"
|
||||||
|
OPN_IMG_PATH="${IMG_DIR}/OPNsense-25.7-serial-amd64.img"
|
||||||
|
CENTOS_ISO_URL="https://mirror.stream.centos.org/9-stream/BaseOS/x86_64/os/images/boot.iso"
|
||||||
|
CENTOS_ISO_PATH="${IMG_DIR}/CentOS-Stream-9-latest-boot.iso"
|
||||||
|
|
||||||
|
CONNECT_URI="qemu:///system"
|
||||||
|
|
||||||
|
download_if_missing() {
|
||||||
|
local url="$1"
|
||||||
|
local dest="$2"
|
||||||
|
if [[ ! -f "$dest" ]]; then
|
||||||
|
echo "Downloading $url to $dest"
|
||||||
|
mkdir -p "$(dirname "$dest")"
|
||||||
|
local tmp
|
||||||
|
tmp="$(mktemp)"
|
||||||
|
curl -L --progress-bar "$url" -o "$tmp"
|
||||||
|
case "$url" in
|
||||||
|
*.bz2) bunzip2 -c "$tmp" > "$dest" && rm -f "$tmp" ;;
|
||||||
|
*) mv "$tmp" "$dest" ;;
|
||||||
|
esac
|
||||||
|
else
|
||||||
|
echo "Already present: $dest"
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
# Ensures a libvirt network is defined and active
|
||||||
|
ensure_network() {
|
||||||
|
local net_name="$1"
|
||||||
|
local net_xml_path="$2"
|
||||||
|
if virsh --connect "${CONNECT_URI}" net-info "${net_name}" >/dev/null 2>&1; then
|
||||||
|
echo "Network ${net_name} already exists."
|
||||||
|
else
|
||||||
|
echo "Defining network ${net_name} from ${net_xml_path}"
|
||||||
|
virsh --connect "${CONNECT_URI}" net-define "${net_xml_path}"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if ! virsh --connect "${CONNECT_URI}" net-info "${net_name}" | grep "Active: *yes"; then
|
||||||
|
echo "Starting network ${net_name}..."
|
||||||
|
virsh --connect "${CONNECT_URI}" net-start "${net_name}"
|
||||||
|
virsh --connect "${CONNECT_URI}" net-autostart "${net_name}"
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
# Destroys a VM completely
|
||||||
|
destroy_vm() {
|
||||||
|
local vm_name="$1"
|
||||||
|
if virsh --connect "${CONNECT_URI}" dominfo "$vm_name" >/dev/null 2>&1; then
|
||||||
|
echo "Destroying and undefining VM: ${vm_name}"
|
||||||
|
virsh --connect "${CONNECT_URI}" destroy "$vm_name" || true
|
||||||
|
virsh --connect "${CONNECT_URI}" undefine "$vm_name" --nvram
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
# Destroys a libvirt network
|
||||||
|
destroy_network() {
|
||||||
|
local net_name="$1"
|
||||||
|
if virsh --connect "${CONNECT_URI}" net-info "$net_name" >/dev/null 2>&1; then
|
||||||
|
echo "Destroying and undefining network: ${net_name}"
|
||||||
|
virsh --connect "${CONNECT_URI}" net-destroy "$net_name" || true
|
||||||
|
virsh --connect "${CONNECT_URI}" net-undefine "$net_name"
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
# --- Main Logic ---
|
||||||
|
create_lab_environment() {
|
||||||
|
# Create network definition files
|
||||||
|
cat > "${STATE_DIR}/default.xml" <<EOF
|
||||||
|
<network>
|
||||||
|
<name>default</name>
|
||||||
|
<forward mode='nat'/>
|
||||||
|
<bridge name='virbr0' stp='on' delay='0'/>
|
||||||
|
<ip address='192.168.122.1' netmask='255.255.255.0'>
|
||||||
|
<dhcp>
|
||||||
|
<range start='192.168.122.100' end='192.168.122.200'/>
|
||||||
|
</dhcp>
|
||||||
|
</ip>
|
||||||
|
</network>
|
||||||
|
EOF
|
||||||
|
|
||||||
|
cat > "${STATE_DIR}/${NET_HARMONYLAN}.xml" <<EOF
|
||||||
|
<network>
|
||||||
|
<name>${NET_HARMONYLAN}</name>
|
||||||
|
<bridge name='virbr1' stp='on' delay='0'/>
|
||||||
|
</network>
|
||||||
|
EOF
|
||||||
|
|
||||||
|
# Ensure both networks exist and are active
|
||||||
|
ensure_network "default" "${STATE_DIR}/default.xml"
|
||||||
|
ensure_network "${NET_HARMONYLAN}" "${STATE_DIR}/${NET_HARMONYLAN}.xml"
|
||||||
|
|
||||||
|
# --- Create OPNsense VM (MODIFIED SECTION) ---
|
||||||
|
local disk_opn="${IMG_DIR}/${VM_OPN}.qcow2"
|
||||||
|
if [[ ! -f "$disk_opn" ]]; then
|
||||||
|
qemu-img create -f qcow2 "$disk_opn" "${DISK_OPN_GB}G"
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "Creating OPNsense VM using serial image..."
|
||||||
|
virt-install \
|
||||||
|
--connect "${CONNECT_URI}" \
|
||||||
|
--name "${VM_OPN}" \
|
||||||
|
--ram "${RAM_OPN}" \
|
||||||
|
--vcpus "${VCPUS_OPN}" \
|
||||||
|
--cpu host-passthrough \
|
||||||
|
--os-variant "${OS_VARIANT_OPN}" \
|
||||||
|
--graphics none \
|
||||||
|
--noautoconsole \
|
||||||
|
--disk path="${disk_opn}",device=disk,bus=virtio,boot.order=1 \
|
||||||
|
--disk path="${OPN_IMG_PATH}",device=disk,bus=usb,readonly=on,boot.order=2 \
|
||||||
|
--network network=default,model=virtio \
|
||||||
|
--network network="${NET_HARMONYLAN}",model=virtio \
|
||||||
|
--boot uefi,menu=on
|
||||||
|
|
||||||
|
echo "OPNsense VM created. Connect with: sudo virsh console ${VM_OPN}"
|
||||||
|
echo "The VM will boot from the serial installation image."
|
||||||
|
echo "Login with user 'installer' and password 'opnsense' to start the installation."
|
||||||
|
echo "Install onto the VirtIO disk (vtbd0)."
|
||||||
|
echo "After installation, shutdown the VM, then run 'sudo virsh edit ${VM_OPN}' and remove the USB disk block to boot from the installed system."
|
||||||
|
|
||||||
|
# --- Create PXE Client VM ---
|
||||||
|
local disk_pxe="${IMG_DIR}/${VM_PXE}.qcow2"
|
||||||
|
if [[ ! -f "$disk_pxe" ]]; then
|
||||||
|
qemu-img create -f qcow2 "$disk_pxe" "${DISK_PXE_GB}G"
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "Creating PXE client VM..."
|
||||||
|
virt-install \
|
||||||
|
--connect "${CONNECT_URI}" \
|
||||||
|
--name "${VM_PXE}" \
|
||||||
|
--ram "${RAM_PXE}" \
|
||||||
|
--vcpus "${VCPUS_PXE}" \
|
||||||
|
--cpu host-passthrough \
|
||||||
|
--os-variant "${OS_VARIANT_LINUX}" \
|
||||||
|
--graphics none \
|
||||||
|
--noautoconsole \
|
||||||
|
--disk path="${disk_pxe}",format=qcow2,bus=virtio \
|
||||||
|
--network network="${NET_HARMONYLAN}",model=virtio \
|
||||||
|
--pxe \
|
||||||
|
--boot uefi,menu=on
|
||||||
|
|
||||||
|
echo "PXE VM created. It will attempt to netboot on ${NET_HARMONYLAN}."
|
||||||
|
}
|
||||||
|
|
||||||
|
# --- Script Entrypoint ---
|
||||||
|
case "${1:-}" in
|
||||||
|
up)
|
||||||
|
mkdir -p "${IMG_DIR}" "${STATE_DIR}"
|
||||||
|
download_if_missing "$OPN_IMG_URL" "$OPN_IMG_PATH"
|
||||||
|
download_if_missing "$CENTOS_ISO_URL" "$CENTOS_ISO_PATH"
|
||||||
|
create_lab_environment
|
||||||
|
echo "Lab setup complete. Use 'sudo virsh list --all' to see VMs."
|
||||||
|
;;
|
||||||
|
clean)
|
||||||
|
destroy_vm "${VM_PXE}"
|
||||||
|
destroy_vm "${VM_OPN}"
|
||||||
|
destroy_network "${NET_HARMONYLAN}"
|
||||||
|
# Optionally destroy the default network if you want a full reset
|
||||||
|
# destroy_network "default"
|
||||||
|
echo "Cleanup complete."
|
||||||
|
;;
|
||||||
|
*)
|
||||||
|
echo "Usage: sudo $0 {up|clean}"
|
||||||
|
exit 1
|
||||||
|
;;
|
||||||
|
esac
|
||||||
BIN
examples/application_monitoring_with_tenant/harmony
Executable file
BIN
examples/application_monitoring_with_tenant/harmony
Executable file
Binary file not shown.
@@ -1,6 +1,9 @@
|
|||||||
use harmony::{
|
use harmony::{
|
||||||
inventory::Inventory,
|
inventory::Inventory,
|
||||||
modules::dummy::{ErrorScore, PanicScore, SuccessScore},
|
modules::{
|
||||||
|
dummy::{ErrorScore, PanicScore, SuccessScore},
|
||||||
|
inventory::DiscoverInventoryAgentScore,
|
||||||
|
},
|
||||||
topology::LocalhostTopology,
|
topology::LocalhostTopology,
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -13,6 +16,9 @@ async fn main() {
|
|||||||
Box::new(SuccessScore {}),
|
Box::new(SuccessScore {}),
|
||||||
Box::new(ErrorScore {}),
|
Box::new(ErrorScore {}),
|
||||||
Box::new(PanicScore {}),
|
Box::new(PanicScore {}),
|
||||||
|
Box::new(DiscoverInventoryAgentScore {
|
||||||
|
discovery_timeout: Some(10),
|
||||||
|
}),
|
||||||
],
|
],
|
||||||
None,
|
None,
|
||||||
)
|
)
|
||||||
|
|||||||
@@ -8,7 +8,6 @@ use harmony::{
|
|||||||
hardware::{FirewallGroup, HostCategory, Location, PhysicalHost, SwitchGroup},
|
hardware::{FirewallGroup, HostCategory, Location, PhysicalHost, SwitchGroup},
|
||||||
infra::opnsense::OPNSenseManagementInterface,
|
infra::opnsense::OPNSenseManagementInterface,
|
||||||
inventory::Inventory,
|
inventory::Inventory,
|
||||||
maestro::Maestro,
|
|
||||||
modules::{
|
modules::{
|
||||||
http::StaticFilesHttpScore,
|
http::StaticFilesHttpScore,
|
||||||
ipxe::IpxeScore,
|
ipxe::IpxeScore,
|
||||||
@@ -126,12 +125,18 @@ async fn main() {
|
|||||||
harmony::modules::okd::load_balancer::OKDLoadBalancerScore::new(&topology);
|
harmony::modules::okd::load_balancer::OKDLoadBalancerScore::new(&topology);
|
||||||
|
|
||||||
let tftp_score = TftpScore::new(Url::LocalFolder("./data/watchguard/tftpboot".to_string()));
|
let tftp_score = TftpScore::new(Url::LocalFolder("./data/watchguard/tftpboot".to_string()));
|
||||||
let http_score = StaticFilesHttpScore::new(Url::LocalFolder(
|
let http_score = StaticFilesHttpScore {
|
||||||
|
folder_to_serve: Some(Url::LocalFolder(
|
||||||
"./data/watchguard/pxe-http-files".to_string(),
|
"./data/watchguard/pxe-http-files".to_string(),
|
||||||
));
|
)),
|
||||||
|
files: vec![],
|
||||||
|
};
|
||||||
let ipxe_score = IpxeScore::new();
|
let ipxe_score = IpxeScore::new();
|
||||||
let mut maestro = Maestro::initialize(inventory, topology).await.unwrap();
|
|
||||||
maestro.register_all(vec![
|
harmony_tui::run(
|
||||||
|
inventory,
|
||||||
|
topology,
|
||||||
|
vec![
|
||||||
Box::new(dns_score),
|
Box::new(dns_score),
|
||||||
Box::new(bootstrap_dhcp_score),
|
Box::new(bootstrap_dhcp_score),
|
||||||
Box::new(bootstrap_load_balancer_score),
|
Box::new(bootstrap_load_balancer_score),
|
||||||
@@ -140,6 +145,8 @@ async fn main() {
|
|||||||
Box::new(http_score),
|
Box::new(http_score),
|
||||||
Box::new(ipxe_score),
|
Box::new(ipxe_score),
|
||||||
Box::new(dhcp_score),
|
Box::new(dhcp_score),
|
||||||
]);
|
],
|
||||||
harmony_tui::init(maestro).await.unwrap();
|
)
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
}
|
}
|
||||||
|
|||||||
21
examples/okd_pxe/Cargo.toml
Normal file
21
examples/okd_pxe/Cargo.toml
Normal file
@@ -0,0 +1,21 @@
|
|||||||
|
[package]
|
||||||
|
name = "example-pxe"
|
||||||
|
edition = "2024"
|
||||||
|
version.workspace = true
|
||||||
|
readme.workspace = true
|
||||||
|
license.workspace = true
|
||||||
|
publish = false
|
||||||
|
|
||||||
|
[dependencies]
|
||||||
|
harmony = { path = "../../harmony" }
|
||||||
|
harmony_cli = { path = "../../harmony_cli" }
|
||||||
|
harmony_types = { path = "../../harmony_types" }
|
||||||
|
harmony_secret = { path = "../../harmony_secret" }
|
||||||
|
harmony_secret_derive = { path = "../../harmony_secret_derive" }
|
||||||
|
cidr = { workspace = true }
|
||||||
|
tokio = { workspace = true }
|
||||||
|
harmony_macros = { path = "../../harmony_macros" }
|
||||||
|
log = { workspace = true }
|
||||||
|
env_logger = { workspace = true }
|
||||||
|
url = { workspace = true }
|
||||||
|
serde.workspace = true
|
||||||
24
examples/okd_pxe/src/main.rs
Normal file
24
examples/okd_pxe/src/main.rs
Normal file
@@ -0,0 +1,24 @@
|
|||||||
|
mod topology;
|
||||||
|
|
||||||
|
use crate::topology::{get_inventory, get_topology};
|
||||||
|
use harmony::modules::okd::ipxe::OkdIpxeScore;
|
||||||
|
|
||||||
|
#[tokio::main]
|
||||||
|
async fn main() {
|
||||||
|
let inventory = get_inventory();
|
||||||
|
let topology = get_topology().await;
|
||||||
|
|
||||||
|
let kickstart_filename = "inventory.kickstart".to_string();
|
||||||
|
let cluster_pubkey_filename = "cluster_ssh_key.pub".to_string();
|
||||||
|
let harmony_inventory_agent = "harmony_inventory_agent".to_string();
|
||||||
|
|
||||||
|
let ipxe_score = OkdIpxeScore {
|
||||||
|
kickstart_filename,
|
||||||
|
harmony_inventory_agent,
|
||||||
|
cluster_pubkey_filename,
|
||||||
|
};
|
||||||
|
|
||||||
|
harmony_cli::run(inventory, topology, vec![Box::new(ipxe_score)], None)
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
}
|
||||||
78
examples/okd_pxe/src/topology.rs
Normal file
78
examples/okd_pxe/src/topology.rs
Normal file
@@ -0,0 +1,78 @@
|
|||||||
|
use cidr::Ipv4Cidr;
|
||||||
|
use harmony::{
|
||||||
|
hardware::{FirewallGroup, HostCategory, Location, PhysicalHost, SwitchGroup},
|
||||||
|
infra::opnsense::OPNSenseManagementInterface,
|
||||||
|
inventory::Inventory,
|
||||||
|
topology::{HAClusterTopology, LogicalHost, UnmanagedRouter},
|
||||||
|
};
|
||||||
|
use harmony_macros::{ip, ipv4};
|
||||||
|
use harmony_secret::{Secret, SecretManager};
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
use std::{net::IpAddr, sync::Arc};
|
||||||
|
|
||||||
|
#[derive(Secret, Serialize, Deserialize, Debug, PartialEq)]
|
||||||
|
struct OPNSenseFirewallConfig {
|
||||||
|
username: String,
|
||||||
|
password: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn get_topology() -> HAClusterTopology {
|
||||||
|
let firewall = harmony::topology::LogicalHost {
|
||||||
|
ip: ip!("192.168.1.1"),
|
||||||
|
name: String::from("opnsense-1"),
|
||||||
|
};
|
||||||
|
|
||||||
|
let config = SecretManager::get::<OPNSenseFirewallConfig>().await;
|
||||||
|
let config = config.unwrap();
|
||||||
|
|
||||||
|
let opnsense = Arc::new(
|
||||||
|
harmony::infra::opnsense::OPNSenseFirewall::new(
|
||||||
|
firewall,
|
||||||
|
None,
|
||||||
|
&config.username,
|
||||||
|
&config.password,
|
||||||
|
)
|
||||||
|
.await,
|
||||||
|
);
|
||||||
|
let lan_subnet = ipv4!("192.168.1.0");
|
||||||
|
let gateway_ipv4 = ipv4!("192.168.1.1");
|
||||||
|
let gateway_ip = IpAddr::V4(gateway_ipv4);
|
||||||
|
harmony::topology::HAClusterTopology {
|
||||||
|
domain_name: "demo.harmony.mcd".to_string(),
|
||||||
|
router: Arc::new(UnmanagedRouter::new(
|
||||||
|
gateway_ip,
|
||||||
|
Ipv4Cidr::new(lan_subnet, 24).unwrap(),
|
||||||
|
)),
|
||||||
|
load_balancer: opnsense.clone(),
|
||||||
|
firewall: opnsense.clone(),
|
||||||
|
tftp_server: opnsense.clone(),
|
||||||
|
http_server: opnsense.clone(),
|
||||||
|
dhcp_server: opnsense.clone(),
|
||||||
|
dns_server: opnsense.clone(),
|
||||||
|
control_plane: vec![LogicalHost {
|
||||||
|
ip: ip!("10.100.8.20"),
|
||||||
|
name: "cp0".to_string(),
|
||||||
|
}],
|
||||||
|
bootstrap_host: LogicalHost {
|
||||||
|
ip: ip!("10.100.8.20"),
|
||||||
|
name: "cp0".to_string(),
|
||||||
|
},
|
||||||
|
workers: vec![],
|
||||||
|
switch: vec![],
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn get_inventory() -> Inventory {
|
||||||
|
Inventory {
|
||||||
|
location: Location::new(
|
||||||
|
"Some virtual machine or maybe a physical machine if you're cool".to_string(),
|
||||||
|
"testopnsense".to_string(),
|
||||||
|
),
|
||||||
|
switch: SwitchGroup::from([]),
|
||||||
|
firewall: FirewallGroup::from([PhysicalHost::empty(HostCategory::Firewall)
|
||||||
|
.management(Arc::new(OPNSenseManagementInterface::new()))]),
|
||||||
|
storage_host: vec![],
|
||||||
|
worker_host: vec![],
|
||||||
|
control_plane_host: vec![],
|
||||||
|
}
|
||||||
|
}
|
||||||
7
examples/okd_pxe/ssh_example_key
Normal file
7
examples/okd_pxe/ssh_example_key
Normal file
@@ -0,0 +1,7 @@
|
|||||||
|
-----BEGIN OPENSSH PRIVATE KEY-----
|
||||||
|
b3BlbnNzaC1rZXktdjEAAAAABG5vbmUAAAAEbm9uZQAAAAAAAAABAAAAMwAAAAtzc2gtZW
|
||||||
|
QyNTUxOQAAACAcemw8pbwuvHFaYynxBbS0Cf3ThYuj1Utr7CDqjwySHAAAAJikacCNpGnA
|
||||||
|
jQAAAAtzc2gtZWQyNTUxOQAAACAcemw8pbwuvHFaYynxBbS0Cf3ThYuj1Utr7CDqjwySHA
|
||||||
|
AAAECiiKk4V6Q5cVs6axDM4sjAzZn/QCZLQekmYQXS9XbEYxx6bDylvC68cVpjKfEFtLQJ
|
||||||
|
/dOFi6PVS2vsIOqPDJIcAAAAEGplYW5nYWJAbGlsaWFuZTIBAgMEBQ==
|
||||||
|
-----END OPENSSH PRIVATE KEY-----
|
||||||
1
examples/okd_pxe/ssh_example_key.pub
Normal file
1
examples/okd_pxe/ssh_example_key.pub
Normal file
@@ -0,0 +1 @@
|
|||||||
|
ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBx6bDylvC68cVpjKfEFtLQJ/dOFi6PVS2vsIOqPDJIc jeangab@liliane2
|
||||||
@@ -8,7 +8,6 @@ use harmony::{
|
|||||||
hardware::{FirewallGroup, HostCategory, Location, PhysicalHost, SwitchGroup},
|
hardware::{FirewallGroup, HostCategory, Location, PhysicalHost, SwitchGroup},
|
||||||
infra::opnsense::OPNSenseManagementInterface,
|
infra::opnsense::OPNSenseManagementInterface,
|
||||||
inventory::Inventory,
|
inventory::Inventory,
|
||||||
maestro::Maestro,
|
|
||||||
modules::{
|
modules::{
|
||||||
dummy::{ErrorScore, PanicScore, SuccessScore},
|
dummy::{ErrorScore, PanicScore, SuccessScore},
|
||||||
http::StaticFilesHttpScore,
|
http::StaticFilesHttpScore,
|
||||||
@@ -81,11 +80,17 @@ async fn main() {
|
|||||||
let load_balancer_score = OKDLoadBalancerScore::new(&topology);
|
let load_balancer_score = OKDLoadBalancerScore::new(&topology);
|
||||||
|
|
||||||
let tftp_score = TftpScore::new(Url::LocalFolder("./data/watchguard/tftpboot".to_string()));
|
let tftp_score = TftpScore::new(Url::LocalFolder("./data/watchguard/tftpboot".to_string()));
|
||||||
let http_score = StaticFilesHttpScore::new(Url::LocalFolder(
|
let http_score = StaticFilesHttpScore {
|
||||||
|
folder_to_serve: Some(Url::LocalFolder(
|
||||||
"./data/watchguard/pxe-http-files".to_string(),
|
"./data/watchguard/pxe-http-files".to_string(),
|
||||||
));
|
)),
|
||||||
let mut maestro = Maestro::initialize(inventory, topology).await.unwrap();
|
files: vec![],
|
||||||
maestro.register_all(vec![
|
};
|
||||||
|
|
||||||
|
harmony_tui::run(
|
||||||
|
inventory,
|
||||||
|
topology,
|
||||||
|
vec![
|
||||||
Box::new(dns_score),
|
Box::new(dns_score),
|
||||||
Box::new(dhcp_score),
|
Box::new(dhcp_score),
|
||||||
Box::new(load_balancer_score),
|
Box::new(load_balancer_score),
|
||||||
@@ -98,6 +103,8 @@ async fn main() {
|
|||||||
Box::new(SuccessScore {}),
|
Box::new(SuccessScore {}),
|
||||||
Box::new(ErrorScore {}),
|
Box::new(ErrorScore {}),
|
||||||
Box::new(PanicScore {}),
|
Box::new(PanicScore {}),
|
||||||
]);
|
],
|
||||||
harmony_tui::init(maestro).await.unwrap();
|
)
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2,7 +2,6 @@ use std::net::{SocketAddr, SocketAddrV4};
|
|||||||
|
|
||||||
use harmony::{
|
use harmony::{
|
||||||
inventory::Inventory,
|
inventory::Inventory,
|
||||||
maestro::Maestro,
|
|
||||||
modules::{
|
modules::{
|
||||||
dns::DnsScore,
|
dns::DnsScore,
|
||||||
dummy::{ErrorScore, PanicScore, SuccessScore},
|
dummy::{ErrorScore, PanicScore, SuccessScore},
|
||||||
@@ -16,18 +15,19 @@ use harmony_macros::ipv4;
|
|||||||
|
|
||||||
#[tokio::main]
|
#[tokio::main]
|
||||||
async fn main() {
|
async fn main() {
|
||||||
let inventory = Inventory::autoload();
|
harmony_tui::run(
|
||||||
let topology = DummyInfra {};
|
Inventory::autoload(),
|
||||||
let mut maestro = Maestro::initialize(inventory, topology).await.unwrap();
|
DummyInfra {},
|
||||||
|
vec![
|
||||||
maestro.register_all(vec![
|
|
||||||
Box::new(SuccessScore {}),
|
Box::new(SuccessScore {}),
|
||||||
Box::new(ErrorScore {}),
|
Box::new(ErrorScore {}),
|
||||||
Box::new(PanicScore {}),
|
Box::new(PanicScore {}),
|
||||||
Box::new(DnsScore::new(vec![], None)),
|
Box::new(DnsScore::new(vec![], None)),
|
||||||
Box::new(build_large_score()),
|
Box::new(build_large_score()),
|
||||||
]);
|
],
|
||||||
harmony_tui::init(maestro).await.unwrap();
|
)
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
fn build_large_score() -> LoadBalancerScore {
|
fn build_large_score() -> LoadBalancerScore {
|
||||||
|
|||||||
@@ -5,6 +5,9 @@ version.workspace = true
|
|||||||
readme.workspace = true
|
readme.workspace = true
|
||||||
license.workspace = true
|
license.workspace = true
|
||||||
|
|
||||||
|
[features]
|
||||||
|
testing = []
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
rand = "0.9"
|
rand = "0.9"
|
||||||
hex = "0.4"
|
hex = "0.4"
|
||||||
@@ -13,8 +16,8 @@ reqwest = { version = "0.11", features = ["blocking", "json"] }
|
|||||||
russh = "0.45.0"
|
russh = "0.45.0"
|
||||||
rust-ipmi = "0.1.1"
|
rust-ipmi = "0.1.1"
|
||||||
semver = "1.0.23"
|
semver = "1.0.23"
|
||||||
serde = { version = "1.0.209", features = ["derive", "rc"] }
|
serde.workspace = true
|
||||||
serde_json = "1.0.127"
|
serde_json.workspace = true
|
||||||
tokio.workspace = true
|
tokio.workspace = true
|
||||||
derive-new.workspace = true
|
derive-new.workspace = true
|
||||||
log.workspace = true
|
log.workspace = true
|
||||||
@@ -35,8 +38,8 @@ serde-value.workspace = true
|
|||||||
helm-wrapper-rs = "0.4.0"
|
helm-wrapper-rs = "0.4.0"
|
||||||
non-blank-string-rs = "1.0.4"
|
non-blank-string-rs = "1.0.4"
|
||||||
k3d-rs = { path = "../k3d" }
|
k3d-rs = { path = "../k3d" }
|
||||||
directories = "6.0.0"
|
directories.workspace = true
|
||||||
lazy_static = "1.5.0"
|
lazy_static.workspace = true
|
||||||
dockerfile_builder = "0.1.5"
|
dockerfile_builder = "0.1.5"
|
||||||
temp-file = "0.1.9"
|
temp-file = "0.1.9"
|
||||||
convert_case.workspace = true
|
convert_case.workspace = true
|
||||||
@@ -56,7 +59,7 @@ similar.workspace = true
|
|||||||
futures-util = "0.3.31"
|
futures-util = "0.3.31"
|
||||||
tokio-util = "0.7.15"
|
tokio-util = "0.7.15"
|
||||||
strum = { version = "0.27.1", features = ["derive"] }
|
strum = { version = "0.27.1", features = ["derive"] }
|
||||||
tempfile = "3.20.0"
|
tempfile.workspace = true
|
||||||
serde_with = "3.14.0"
|
serde_with = "3.14.0"
|
||||||
schemars = "0.8.22"
|
schemars = "0.8.22"
|
||||||
kube-derive = "1.1.0"
|
kube-derive = "1.1.0"
|
||||||
@@ -64,6 +67,9 @@ bollard.workspace = true
|
|||||||
tar.workspace = true
|
tar.workspace = true
|
||||||
base64.workspace = true
|
base64.workspace = true
|
||||||
once_cell = "1.21.3"
|
once_cell = "1.21.3"
|
||||||
|
harmony_inventory_agent = { path = "../harmony_inventory_agent" }
|
||||||
|
harmony_secret_derive = { version = "0.1.0", path = "../harmony_secret_derive" }
|
||||||
|
askama = "0.14.0"
|
||||||
|
|
||||||
[dev-dependencies]
|
[dev-dependencies]
|
||||||
pretty_assertions.workspace = true
|
pretty_assertions.workspace = true
|
||||||
|
|||||||
BIN
harmony/harmony.rlib
Normal file
BIN
harmony/harmony.rlib
Normal file
Binary file not shown.
22
harmony/src/domain/data/file.rs
Normal file
22
harmony/src/domain/data/file.rs
Normal file
@@ -0,0 +1,22 @@
|
|||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||||
|
pub struct FileContent {
|
||||||
|
pub path: FilePath,
|
||||||
|
pub content: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||||
|
pub enum FilePath {
|
||||||
|
Relative(String),
|
||||||
|
Absolute(String),
|
||||||
|
}
|
||||||
|
|
||||||
|
impl std::fmt::Display for FilePath {
|
||||||
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||||
|
match self {
|
||||||
|
FilePath::Relative(path) => f.write_fmt(format_args!("./{path}")),
|
||||||
|
FilePath::Absolute(path) => f.write_fmt(format_args!("/{path}")),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,4 +1,6 @@
|
|||||||
|
mod file;
|
||||||
mod id;
|
mod id;
|
||||||
mod version;
|
mod version;
|
||||||
|
pub use file::*;
|
||||||
pub use id::*;
|
pub use id::*;
|
||||||
pub use version::*;
|
pub use version::*;
|
||||||
|
|||||||
@@ -1,6 +1,7 @@
|
|||||||
use log::debug;
|
|
||||||
use once_cell::sync::Lazy;
|
use once_cell::sync::Lazy;
|
||||||
use tokio::sync::broadcast;
|
use std::{collections::HashMap, sync::Mutex};
|
||||||
|
|
||||||
|
use crate::modules::application::ApplicationFeatureStatus;
|
||||||
|
|
||||||
use super::{
|
use super::{
|
||||||
interpret::{InterpretError, Outcome},
|
interpret::{InterpretError, Outcome},
|
||||||
@@ -10,13 +11,16 @@ use super::{
|
|||||||
#[derive(Debug, Clone)]
|
#[derive(Debug, Clone)]
|
||||||
pub enum HarmonyEvent {
|
pub enum HarmonyEvent {
|
||||||
HarmonyStarted,
|
HarmonyStarted,
|
||||||
|
HarmonyFinished,
|
||||||
InterpretExecutionStarted {
|
InterpretExecutionStarted {
|
||||||
|
execution_id: String,
|
||||||
topology: String,
|
topology: String,
|
||||||
interpret: String,
|
interpret: String,
|
||||||
score: String,
|
score: String,
|
||||||
message: String,
|
message: String,
|
||||||
},
|
},
|
||||||
InterpretExecutionFinished {
|
InterpretExecutionFinished {
|
||||||
|
execution_id: String,
|
||||||
topology: String,
|
topology: String,
|
||||||
interpret: String,
|
interpret: String,
|
||||||
score: String,
|
score: String,
|
||||||
@@ -27,40 +31,54 @@ pub enum HarmonyEvent {
|
|||||||
status: TopologyStatus,
|
status: TopologyStatus,
|
||||||
message: Option<String>,
|
message: Option<String>,
|
||||||
},
|
},
|
||||||
|
ApplicationFeatureStateChanged {
|
||||||
|
topology: String,
|
||||||
|
application: String,
|
||||||
|
feature: String,
|
||||||
|
status: ApplicationFeatureStatus,
|
||||||
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
static HARMONY_EVENT_BUS: Lazy<broadcast::Sender<HarmonyEvent>> = Lazy::new(|| {
|
type Subscriber = Box<dyn Fn(&HarmonyEvent) + Send + Sync>;
|
||||||
// TODO: Adjust channel capacity
|
|
||||||
let (tx, _rx) = broadcast::channel(100);
|
|
||||||
tx
|
|
||||||
});
|
|
||||||
|
|
||||||
pub fn instrument(event: HarmonyEvent) -> Result<(), &'static str> {
|
static SUBSCRIBERS: Lazy<Mutex<HashMap<String, Subscriber>>> =
|
||||||
match HARMONY_EVENT_BUS.send(event) {
|
Lazy::new(|| Mutex::new(HashMap::new()));
|
||||||
Ok(_) => Ok(()),
|
|
||||||
Err(_) => Err("send error: no subscribers"),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
pub async fn subscribe<F, Fut>(name: &str, mut handler: F)
|
/// Subscribes a listener to all instrumentation events.
|
||||||
|
///
|
||||||
|
/// Simply provide a unique name and a closure to run when an event happens.
|
||||||
|
///
|
||||||
|
/// # Example
|
||||||
|
/// ```
|
||||||
|
/// use harmony::instrumentation;
|
||||||
|
/// instrumentation::subscribe("my_logger", |event| {
|
||||||
|
/// println!("Event occurred: {:?}", event);
|
||||||
|
/// });
|
||||||
|
/// ```
|
||||||
|
pub fn subscribe<F>(name: &str, callback: F)
|
||||||
where
|
where
|
||||||
F: FnMut(HarmonyEvent) -> Fut + Send + 'static,
|
F: Fn(&HarmonyEvent) + Send + Sync + 'static,
|
||||||
Fut: Future<Output = bool> + Send,
|
|
||||||
{
|
{
|
||||||
let mut rx = HARMONY_EVENT_BUS.subscribe();
|
let mut subs = SUBSCRIBERS.lock().unwrap();
|
||||||
debug!("[{name}] Service started. Listening for events...");
|
subs.insert(name.to_string(), Box::new(callback));
|
||||||
loop {
|
|
||||||
match rx.recv().await {
|
|
||||||
Ok(event) => {
|
|
||||||
if !handler(event).await {
|
|
||||||
debug!("[{name}] Handler requested exit.");
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Err(broadcast::error::RecvError::Lagged(n)) => {
|
|
||||||
debug!("[{name}] Lagged behind by {n} messages.");
|
|
||||||
}
|
|
||||||
Err(_) => break,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Instruments an event, notifying all subscribers.
|
||||||
|
///
|
||||||
|
/// This will call every closure that was registered with `subscribe`.
|
||||||
|
///
|
||||||
|
/// # Example
|
||||||
|
/// ```
|
||||||
|
/// use harmony::instrumentation;
|
||||||
|
/// use harmony::instrumentation::HarmonyEvent;
|
||||||
|
/// instrumentation::instrument(HarmonyEvent::HarmonyStarted);
|
||||||
|
/// ```
|
||||||
|
pub fn instrument(event: HarmonyEvent) -> Result<(), &'static str> {
|
||||||
|
let subs = SUBSCRIBERS.lock().unwrap();
|
||||||
|
|
||||||
|
for callback in subs.values() {
|
||||||
|
callback(&event);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -32,6 +32,7 @@ pub enum InterpretName {
|
|||||||
Lamp,
|
Lamp,
|
||||||
ApplicationMonitoring,
|
ApplicationMonitoring,
|
||||||
K8sPrometheusCrdAlerting,
|
K8sPrometheusCrdAlerting,
|
||||||
|
DiscoverInventoryAgent,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl std::fmt::Display for InterpretName {
|
impl std::fmt::Display for InterpretName {
|
||||||
@@ -58,6 +59,7 @@ impl std::fmt::Display for InterpretName {
|
|||||||
InterpretName::Lamp => f.write_str("LAMP"),
|
InterpretName::Lamp => f.write_str("LAMP"),
|
||||||
InterpretName::ApplicationMonitoring => f.write_str("ApplicationMonitoring"),
|
InterpretName::ApplicationMonitoring => f.write_str("ApplicationMonitoring"),
|
||||||
InterpretName::K8sPrometheusCrdAlerting => f.write_str("K8sPrometheusCrdAlerting"),
|
InterpretName::K8sPrometheusCrdAlerting => f.write_str("K8sPrometheusCrdAlerting"),
|
||||||
|
InterpretName::DiscoverInventoryAgent => f.write_str("DiscoverInventoryAgent"),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -74,6 +74,7 @@ impl<T: Topology> Maestro<T> {
|
|||||||
|
|
||||||
fn is_topology_initialized(&self) -> bool {
|
fn is_topology_initialized(&self) -> bool {
|
||||||
self.topology_state.status == TopologyStatus::Success
|
self.topology_state.status == TopologyStatus::Success
|
||||||
|
|| self.topology_state.status == TopologyStatus::Noop
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn interpret(&self, score: Box<dyn Score<T>>) -> Result<Outcome, InterpretError> {
|
pub async fn interpret(&self, score: Box<dyn Score<T>>) -> Result<Outcome, InterpretError> {
|
||||||
|
|||||||
@@ -5,6 +5,7 @@ use serde::Serialize;
|
|||||||
use serde_value::Value;
|
use serde_value::Value;
|
||||||
|
|
||||||
use super::{
|
use super::{
|
||||||
|
data::Id,
|
||||||
instrumentation::{self, HarmonyEvent},
|
instrumentation::{self, HarmonyEvent},
|
||||||
interpret::{Interpret, InterpretError, Outcome},
|
interpret::{Interpret, InterpretError, Outcome},
|
||||||
inventory::Inventory,
|
inventory::Inventory,
|
||||||
@@ -20,9 +21,11 @@ pub trait Score<T: Topology>:
|
|||||||
inventory: &Inventory,
|
inventory: &Inventory,
|
||||||
topology: &T,
|
topology: &T,
|
||||||
) -> Result<Outcome, InterpretError> {
|
) -> Result<Outcome, InterpretError> {
|
||||||
|
let id = Id::default();
|
||||||
let interpret = self.create_interpret();
|
let interpret = self.create_interpret();
|
||||||
|
|
||||||
instrumentation::instrument(HarmonyEvent::InterpretExecutionStarted {
|
instrumentation::instrument(HarmonyEvent::InterpretExecutionStarted {
|
||||||
|
execution_id: id.clone().to_string(),
|
||||||
topology: topology.name().into(),
|
topology: topology.name().into(),
|
||||||
interpret: interpret.get_name().to_string(),
|
interpret: interpret.get_name().to_string(),
|
||||||
score: self.name(),
|
score: self.name(),
|
||||||
@@ -32,6 +35,7 @@ pub trait Score<T: Topology>:
|
|||||||
let result = interpret.execute(inventory, topology).await;
|
let result = interpret.execute(inventory, topology).await;
|
||||||
|
|
||||||
instrumentation::instrument(HarmonyEvent::InterpretExecutionFinished {
|
instrumentation::instrument(HarmonyEvent::InterpretExecutionFinished {
|
||||||
|
execution_id: id.clone().to_string(),
|
||||||
topology: topology.name().into(),
|
topology: topology.name().into(),
|
||||||
interpret: interpret.get_name().to_string(),
|
interpret: interpret.get_name().to_string(),
|
||||||
score: self.name(),
|
score: self.name(),
|
||||||
|
|||||||
@@ -1,9 +1,12 @@
|
|||||||
use async_trait::async_trait;
|
use async_trait::async_trait;
|
||||||
use harmony_macros::ip;
|
use harmony_macros::ip;
|
||||||
use harmony_types::net::MacAddress;
|
use harmony_types::net::MacAddress;
|
||||||
|
use log::debug;
|
||||||
use log::info;
|
use log::info;
|
||||||
|
|
||||||
|
use crate::data::FileContent;
|
||||||
use crate::executors::ExecutorError;
|
use crate::executors::ExecutorError;
|
||||||
|
use crate::topology::PxeOptions;
|
||||||
|
|
||||||
use super::DHCPStaticEntry;
|
use super::DHCPStaticEntry;
|
||||||
use super::DhcpServer;
|
use super::DhcpServer;
|
||||||
@@ -49,9 +52,10 @@ impl Topology for HAClusterTopology {
|
|||||||
"HAClusterTopology"
|
"HAClusterTopology"
|
||||||
}
|
}
|
||||||
async fn ensure_ready(&self) -> Result<PreparationOutcome, PreparationError> {
|
async fn ensure_ready(&self) -> Result<PreparationOutcome, PreparationError> {
|
||||||
todo!(
|
debug!(
|
||||||
"ensure_ready, not entirely sure what it should do here, probably something like verify that the hosts are reachable and all services are up and ready."
|
"ensure_ready, not entirely sure what it should do here, probably something like verify that the hosts are reachable and all services are up and ready."
|
||||||
)
|
);
|
||||||
|
Ok(PreparationOutcome::Noop)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -153,12 +157,10 @@ impl DhcpServer for HAClusterTopology {
|
|||||||
async fn list_static_mappings(&self) -> Vec<(MacAddress, IpAddress)> {
|
async fn list_static_mappings(&self) -> Vec<(MacAddress, IpAddress)> {
|
||||||
self.dhcp_server.list_static_mappings().await
|
self.dhcp_server.list_static_mappings().await
|
||||||
}
|
}
|
||||||
async fn set_next_server(&self, ip: IpAddress) -> Result<(), ExecutorError> {
|
async fn set_pxe_options(&self, options: PxeOptions) -> Result<(), ExecutorError> {
|
||||||
self.dhcp_server.set_next_server(ip).await
|
self.dhcp_server.set_pxe_options(options).await
|
||||||
}
|
|
||||||
async fn set_boot_filename(&self, boot_filename: &str) -> Result<(), ExecutorError> {
|
|
||||||
self.dhcp_server.set_boot_filename(boot_filename).await
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn get_ip(&self) -> IpAddress {
|
fn get_ip(&self) -> IpAddress {
|
||||||
self.dhcp_server.get_ip()
|
self.dhcp_server.get_ip()
|
||||||
}
|
}
|
||||||
@@ -168,16 +170,6 @@ impl DhcpServer for HAClusterTopology {
|
|||||||
async fn commit_config(&self) -> Result<(), ExecutorError> {
|
async fn commit_config(&self) -> Result<(), ExecutorError> {
|
||||||
self.dhcp_server.commit_config().await
|
self.dhcp_server.commit_config().await
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn set_filename(&self, filename: &str) -> Result<(), ExecutorError> {
|
|
||||||
self.dhcp_server.set_filename(filename).await
|
|
||||||
}
|
|
||||||
async fn set_filename64(&self, filename64: &str) -> Result<(), ExecutorError> {
|
|
||||||
self.dhcp_server.set_filename64(filename64).await
|
|
||||||
}
|
|
||||||
async fn set_filenameipxe(&self, filenameipxe: &str) -> Result<(), ExecutorError> {
|
|
||||||
self.dhcp_server.set_filenameipxe(filenameipxe).await
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[async_trait]
|
#[async_trait]
|
||||||
@@ -221,17 +213,21 @@ impl HttpServer for HAClusterTopology {
|
|||||||
self.http_server.serve_files(url).await
|
self.http_server.serve_files(url).await
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async fn serve_file_content(&self, file: &FileContent) -> Result<(), ExecutorError> {
|
||||||
|
self.http_server.serve_file_content(file).await
|
||||||
|
}
|
||||||
|
|
||||||
fn get_ip(&self) -> IpAddress {
|
fn get_ip(&self) -> IpAddress {
|
||||||
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
self.http_server.get_ip()
|
||||||
}
|
}
|
||||||
async fn ensure_initialized(&self) -> Result<(), ExecutorError> {
|
async fn ensure_initialized(&self) -> Result<(), ExecutorError> {
|
||||||
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
self.http_server.ensure_initialized().await
|
||||||
}
|
}
|
||||||
async fn commit_config(&self) -> Result<(), ExecutorError> {
|
async fn commit_config(&self) -> Result<(), ExecutorError> {
|
||||||
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
self.http_server.commit_config().await
|
||||||
}
|
}
|
||||||
async fn reload_restart(&self) -> Result<(), ExecutorError> {
|
async fn reload_restart(&self) -> Result<(), ExecutorError> {
|
||||||
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
self.http_server.reload_restart().await
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -241,7 +237,7 @@ pub struct DummyInfra;
|
|||||||
#[async_trait]
|
#[async_trait]
|
||||||
impl Topology for DummyInfra {
|
impl Topology for DummyInfra {
|
||||||
fn name(&self) -> &str {
|
fn name(&self) -> &str {
|
||||||
todo!()
|
"DummyInfra"
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn ensure_ready(&self) -> Result<PreparationOutcome, PreparationError> {
|
async fn ensure_ready(&self) -> Result<PreparationOutcome, PreparationError> {
|
||||||
@@ -299,19 +295,7 @@ impl DhcpServer for DummyInfra {
|
|||||||
async fn list_static_mappings(&self) -> Vec<(MacAddress, IpAddress)> {
|
async fn list_static_mappings(&self) -> Vec<(MacAddress, IpAddress)> {
|
||||||
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
||||||
}
|
}
|
||||||
async fn set_next_server(&self, _ip: IpAddress) -> Result<(), ExecutorError> {
|
async fn set_pxe_options(&self, _options: PxeOptions) -> Result<(), ExecutorError> {
|
||||||
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
|
||||||
}
|
|
||||||
async fn set_boot_filename(&self, _boot_filename: &str) -> Result<(), ExecutorError> {
|
|
||||||
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
|
||||||
}
|
|
||||||
async fn set_filename(&self, _filename: &str) -> Result<(), ExecutorError> {
|
|
||||||
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
|
||||||
}
|
|
||||||
async fn set_filename64(&self, _filename: &str) -> Result<(), ExecutorError> {
|
|
||||||
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
|
||||||
}
|
|
||||||
async fn set_filenameipxe(&self, _filenameipxe: &str) -> Result<(), ExecutorError> {
|
|
||||||
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
||||||
}
|
}
|
||||||
fn get_ip(&self) -> IpAddress {
|
fn get_ip(&self) -> IpAddress {
|
||||||
@@ -381,6 +365,9 @@ impl HttpServer for DummyInfra {
|
|||||||
async fn serve_files(&self, _url: &Url) -> Result<(), ExecutorError> {
|
async fn serve_files(&self, _url: &Url) -> Result<(), ExecutorError> {
|
||||||
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
||||||
}
|
}
|
||||||
|
async fn serve_file_content(&self, _file: &FileContent) -> Result<(), ExecutorError> {
|
||||||
|
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
||||||
|
}
|
||||||
fn get_ip(&self) -> IpAddress {
|
fn get_ip(&self) -> IpAddress {
|
||||||
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
unimplemented!("{}", UNIMPLEMENTED_DUMMY_INFRA)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
use crate::executors::ExecutorError;
|
use crate::{data::FileContent, executors::ExecutorError};
|
||||||
use async_trait::async_trait;
|
use async_trait::async_trait;
|
||||||
|
|
||||||
use super::{IpAddress, Url};
|
use super::{IpAddress, Url};
|
||||||
@@ -6,6 +6,7 @@ use super::{IpAddress, Url};
|
|||||||
#[async_trait]
|
#[async_trait]
|
||||||
pub trait HttpServer: Send + Sync {
|
pub trait HttpServer: Send + Sync {
|
||||||
async fn serve_files(&self, url: &Url) -> Result<(), ExecutorError>;
|
async fn serve_files(&self, url: &Url) -> Result<(), ExecutorError>;
|
||||||
|
async fn serve_file_content(&self, file: &FileContent) -> Result<(), ExecutorError>;
|
||||||
fn get_ip(&self) -> IpAddress;
|
fn get_ip(&self) -> IpAddress;
|
||||||
|
|
||||||
// async fn set_ip(&self, ip: IpAddress) -> Result<(), ExecutorError>;
|
// async fn set_ip(&self, ip: IpAddress) -> Result<(), ExecutorError>;
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ use k8s_openapi::{
|
|||||||
};
|
};
|
||||||
use kube::{
|
use kube::{
|
||||||
Client, Config, Error, Resource,
|
Client, Config, Error, Resource,
|
||||||
api::{Api, AttachParams, ListParams, Patch, PatchParams, ResourceExt},
|
api::{Api, AttachParams, DeleteParams, ListParams, Patch, PatchParams, ResourceExt},
|
||||||
config::{KubeConfigOptions, Kubeconfig},
|
config::{KubeConfigOptions, Kubeconfig},
|
||||||
core::ErrorResponse,
|
core::ErrorResponse,
|
||||||
runtime::reflector::Lookup,
|
runtime::reflector::Lookup,
|
||||||
@@ -17,7 +17,9 @@ use kube::{
|
|||||||
};
|
};
|
||||||
use log::{debug, error, trace};
|
use log::{debug, error, trace};
|
||||||
use serde::{Serialize, de::DeserializeOwned};
|
use serde::{Serialize, de::DeserializeOwned};
|
||||||
|
use serde_json::json;
|
||||||
use similar::TextDiff;
|
use similar::TextDiff;
|
||||||
|
use tokio::io::AsyncReadExt;
|
||||||
|
|
||||||
#[derive(new, Clone)]
|
#[derive(new, Clone)]
|
||||||
pub struct K8sClient {
|
pub struct K8sClient {
|
||||||
@@ -51,6 +53,66 @@ impl K8sClient {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub async fn get_deployment(
|
||||||
|
&self,
|
||||||
|
name: &str,
|
||||||
|
namespace: Option<&str>,
|
||||||
|
) -> Result<Option<Deployment>, Error> {
|
||||||
|
let deps: Api<Deployment> = if let Some(ns) = namespace {
|
||||||
|
Api::namespaced(self.client.clone(), ns)
|
||||||
|
} else {
|
||||||
|
Api::default_namespaced(self.client.clone())
|
||||||
|
};
|
||||||
|
Ok(deps.get_opt(name).await?)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn get_pod(&self, name: &str, namespace: Option<&str>) -> Result<Option<Pod>, Error> {
|
||||||
|
let pods: Api<Pod> = if let Some(ns) = namespace {
|
||||||
|
Api::namespaced(self.client.clone(), ns)
|
||||||
|
} else {
|
||||||
|
Api::default_namespaced(self.client.clone())
|
||||||
|
};
|
||||||
|
Ok(pods.get_opt(name).await?)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn scale_deployment(
|
||||||
|
&self,
|
||||||
|
name: &str,
|
||||||
|
namespace: Option<&str>,
|
||||||
|
replicas: u32,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
let deployments: Api<Deployment> = if let Some(ns) = namespace {
|
||||||
|
Api::namespaced(self.client.clone(), ns)
|
||||||
|
} else {
|
||||||
|
Api::default_namespaced(self.client.clone())
|
||||||
|
};
|
||||||
|
|
||||||
|
let patch = json!({
|
||||||
|
"spec": {
|
||||||
|
"replicas": replicas
|
||||||
|
}
|
||||||
|
});
|
||||||
|
let pp = PatchParams::default();
|
||||||
|
let scale = Patch::Apply(&patch);
|
||||||
|
deployments.patch_scale(name, &pp, &scale).await?;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn delete_deployment(
|
||||||
|
&self,
|
||||||
|
name: &str,
|
||||||
|
namespace: Option<&str>,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
let deployments: Api<Deployment> = if let Some(ns) = namespace {
|
||||||
|
Api::namespaced(self.client.clone(), ns)
|
||||||
|
} else {
|
||||||
|
Api::default_namespaced(self.client.clone())
|
||||||
|
};
|
||||||
|
let delete_params = DeleteParams::default();
|
||||||
|
deployments.delete(name, &delete_params).await?;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
pub async fn wait_until_deployment_ready(
|
pub async fn wait_until_deployment_ready(
|
||||||
&self,
|
&self,
|
||||||
name: String,
|
name: String,
|
||||||
@@ -76,6 +138,71 @@ impl K8sClient {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Will execute a commond in the first pod found that matches the specified label
|
||||||
|
/// '{label}={name}'
|
||||||
|
pub async fn exec_app_capture_output(
|
||||||
|
&self,
|
||||||
|
name: String,
|
||||||
|
label: String,
|
||||||
|
namespace: Option<&str>,
|
||||||
|
command: Vec<&str>,
|
||||||
|
) -> Result<String, String> {
|
||||||
|
let api: Api<Pod>;
|
||||||
|
|
||||||
|
if let Some(ns) = namespace {
|
||||||
|
api = Api::namespaced(self.client.clone(), ns);
|
||||||
|
} else {
|
||||||
|
api = Api::default_namespaced(self.client.clone());
|
||||||
|
}
|
||||||
|
let pod_list = api
|
||||||
|
.list(&ListParams::default().labels(format!("{label}={name}").as_str()))
|
||||||
|
.await
|
||||||
|
.expect("couldn't get list of pods");
|
||||||
|
|
||||||
|
let res = api
|
||||||
|
.exec(
|
||||||
|
pod_list
|
||||||
|
.items
|
||||||
|
.first()
|
||||||
|
.expect("couldn't get pod")
|
||||||
|
.name()
|
||||||
|
.expect("couldn't get pod name")
|
||||||
|
.into_owned()
|
||||||
|
.as_str(),
|
||||||
|
command,
|
||||||
|
&AttachParams::default().stdout(true).stderr(true),
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
match res {
|
||||||
|
Err(e) => Err(e.to_string()),
|
||||||
|
Ok(mut process) => {
|
||||||
|
let status = process
|
||||||
|
.take_status()
|
||||||
|
.expect("Couldn't get status")
|
||||||
|
.await
|
||||||
|
.expect("Couldn't unwrap status");
|
||||||
|
|
||||||
|
if let Some(s) = status.status {
|
||||||
|
let mut stdout_buf = String::new();
|
||||||
|
if let Some(mut stdout) = process.stdout().take() {
|
||||||
|
stdout
|
||||||
|
.read_to_string(&mut stdout_buf)
|
||||||
|
.await
|
||||||
|
.map_err(|e| format!("Failed to get status stdout {e}"))?;
|
||||||
|
}
|
||||||
|
debug!("Status: {} - {:?}", s, status.details);
|
||||||
|
if s == "Success" {
|
||||||
|
Ok(stdout_buf)
|
||||||
|
} else {
|
||||||
|
Err(s)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
Err("Couldn't get inner status of pod exec".to_string())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// Will execute a command in the first pod found that matches the label `app.kubernetes.io/name={name}`
|
/// Will execute a command in the first pod found that matches the label `app.kubernetes.io/name={name}`
|
||||||
pub async fn exec_app(
|
pub async fn exec_app(
|
||||||
&self,
|
&self,
|
||||||
@@ -120,7 +247,7 @@ impl K8sClient {
|
|||||||
.expect("Couldn't unwrap status");
|
.expect("Couldn't unwrap status");
|
||||||
|
|
||||||
if let Some(s) = status.status {
|
if let Some(s) = status.status {
|
||||||
debug!("Status: {}", s);
|
debug!("Status: {} - {:?}", s, status.details);
|
||||||
if s == "Success" { Ok(()) } else { Err(s) }
|
if s == "Success" { Ok(()) } else { Err(s) }
|
||||||
} else {
|
} else {
|
||||||
Err("Couldn't get inner status of pod exec".to_string())
|
Err("Couldn't get inner status of pod exec".to_string())
|
||||||
|
|||||||
@@ -28,7 +28,13 @@ use super::{
|
|||||||
PreparationOutcome, Topology,
|
PreparationOutcome, Topology,
|
||||||
k8s::K8sClient,
|
k8s::K8sClient,
|
||||||
oberservability::monitoring::AlertReceiver,
|
oberservability::monitoring::AlertReceiver,
|
||||||
tenant::{TenantConfig, TenantManager, k8s::K8sTenantManager},
|
tenant::{
|
||||||
|
TenantConfig, TenantManager,
|
||||||
|
k8s::K8sTenantManager,
|
||||||
|
network_policy::{
|
||||||
|
K3dNetworkPolicyStrategy, NetworkPolicyStrategy, NoopNetworkPolicyStrategy,
|
||||||
|
},
|
||||||
|
},
|
||||||
};
|
};
|
||||||
|
|
||||||
#[derive(Clone, Debug)]
|
#[derive(Clone, Debug)]
|
||||||
@@ -250,16 +256,21 @@ impl K8sAnywhereTopology {
|
|||||||
Ok(Some(state))
|
Ok(Some(state))
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn ensure_k8s_tenant_manager(&self) -> Result<(), String> {
|
async fn ensure_k8s_tenant_manager(&self, k8s_state: &K8sState) -> Result<(), String> {
|
||||||
if self.tenant_manager.get().is_some() {
|
if self.tenant_manager.get().is_some() {
|
||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
|
|
||||||
self.tenant_manager
|
self.tenant_manager
|
||||||
.get_or_try_init(async || -> Result<K8sTenantManager, String> {
|
.get_or_try_init(async || -> Result<K8sTenantManager, String> {
|
||||||
// TOOD: checker si K8s ou K3d/s tenant manager (ref. issue https://git.nationtech.io/NationTech/harmony/issues/94)
|
|
||||||
let k8s_client = self.k8s_client().await?;
|
let k8s_client = self.k8s_client().await?;
|
||||||
Ok(K8sTenantManager::new(k8s_client))
|
let network_policy_strategy: Box<dyn NetworkPolicyStrategy> = match k8s_state.source
|
||||||
|
{
|
||||||
|
K8sSource::LocalK3d => Box::new(K3dNetworkPolicyStrategy::new()),
|
||||||
|
K8sSource::Kubeconfig => Box::new(NoopNetworkPolicyStrategy::new()),
|
||||||
|
};
|
||||||
|
|
||||||
|
Ok(K8sTenantManager::new(k8s_client, network_policy_strategy))
|
||||||
})
|
})
|
||||||
.await?;
|
.await?;
|
||||||
|
|
||||||
@@ -390,7 +401,7 @@ impl Topology for K8sAnywhereTopology {
|
|||||||
"no K8s client could be found or installed".to_string(),
|
"no K8s client could be found or installed".to_string(),
|
||||||
))?;
|
))?;
|
||||||
|
|
||||||
self.ensure_k8s_tenant_manager()
|
self.ensure_k8s_tenant_manager(k8s_state)
|
||||||
.await
|
.await
|
||||||
.map_err(PreparationError::new)?;
|
.map_err(PreparationError::new)?;
|
||||||
|
|
||||||
|
|||||||
@@ -46,16 +46,19 @@ pub trait K8sclient: Send + Sync {
|
|||||||
async fn k8s_client(&self) -> Result<Arc<K8sClient>, String>;
|
async fn k8s_client(&self) -> Result<Arc<K8sClient>, String>;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub struct PxeOptions {
|
||||||
|
pub ipxe_filename: String,
|
||||||
|
pub bios_filename: String,
|
||||||
|
pub efi_filename: String,
|
||||||
|
pub tftp_ip: Option<IpAddress>,
|
||||||
|
}
|
||||||
|
|
||||||
#[async_trait]
|
#[async_trait]
|
||||||
pub trait DhcpServer: Send + Sync + std::fmt::Debug {
|
pub trait DhcpServer: Send + Sync + std::fmt::Debug {
|
||||||
async fn add_static_mapping(&self, entry: &DHCPStaticEntry) -> Result<(), ExecutorError>;
|
async fn add_static_mapping(&self, entry: &DHCPStaticEntry) -> Result<(), ExecutorError>;
|
||||||
async fn remove_static_mapping(&self, mac: &MacAddress) -> Result<(), ExecutorError>;
|
async fn remove_static_mapping(&self, mac: &MacAddress) -> Result<(), ExecutorError>;
|
||||||
async fn list_static_mappings(&self) -> Vec<(MacAddress, IpAddress)>;
|
async fn list_static_mappings(&self) -> Vec<(MacAddress, IpAddress)>;
|
||||||
async fn set_next_server(&self, ip: IpAddress) -> Result<(), ExecutorError>;
|
async fn set_pxe_options(&self, pxe_options: PxeOptions) -> Result<(), ExecutorError>;
|
||||||
async fn set_boot_filename(&self, boot_filename: &str) -> Result<(), ExecutorError>;
|
|
||||||
async fn set_filename(&self, filename: &str) -> Result<(), ExecutorError>;
|
|
||||||
async fn set_filename64(&self, filename64: &str) -> Result<(), ExecutorError>;
|
|
||||||
async fn set_filenameipxe(&self, filenameipxe: &str) -> Result<(), ExecutorError>;
|
|
||||||
fn get_ip(&self) -> IpAddress;
|
fn get_ip(&self) -> IpAddress;
|
||||||
fn get_host(&self) -> LogicalHost;
|
fn get_host(&self) -> LogicalHost;
|
||||||
async fn commit_config(&self) -> Result<(), ExecutorError>;
|
async fn commit_config(&self) -> Result<(), ExecutorError>;
|
||||||
|
|||||||
@@ -20,24 +20,27 @@ use serde::de::DeserializeOwned;
|
|||||||
use serde_json::json;
|
use serde_json::json;
|
||||||
use tokio::sync::OnceCell;
|
use tokio::sync::OnceCell;
|
||||||
|
|
||||||
use super::{TenantConfig, TenantManager};
|
use super::{TenantConfig, TenantManager, network_policy::NetworkPolicyStrategy};
|
||||||
|
|
||||||
#[derive(Clone, Debug)]
|
#[derive(Debug)]
|
||||||
pub struct K8sTenantManager {
|
pub struct K8sTenantManager {
|
||||||
k8s_client: Arc<K8sClient>,
|
k8s_client: Arc<K8sClient>,
|
||||||
k8s_tenant_config: Arc<OnceCell<TenantConfig>>,
|
k8s_tenant_config: Arc<OnceCell<TenantConfig>>,
|
||||||
|
network_policy_strategy: Box<dyn NetworkPolicyStrategy>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl K8sTenantManager {
|
impl K8sTenantManager {
|
||||||
pub fn new(client: Arc<K8sClient>) -> Self {
|
pub fn new(
|
||||||
|
client: Arc<K8sClient>,
|
||||||
|
network_policy_strategy: Box<dyn NetworkPolicyStrategy>,
|
||||||
|
) -> Self {
|
||||||
Self {
|
Self {
|
||||||
k8s_client: client,
|
k8s_client: client,
|
||||||
k8s_tenant_config: Arc::new(OnceCell::new()),
|
k8s_tenant_config: Arc::new(OnceCell::new()),
|
||||||
}
|
network_policy_strategy,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl K8sTenantManager {
|
|
||||||
fn get_namespace_name(&self, config: &TenantConfig) -> String {
|
fn get_namespace_name(&self, config: &TenantConfig) -> String {
|
||||||
config.name.clone()
|
config.name.clone()
|
||||||
}
|
}
|
||||||
@@ -218,29 +221,6 @@ impl K8sTenantManager {
|
|||||||
}
|
}
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"to": [
|
|
||||||
{
|
|
||||||
"ipBlock": {
|
|
||||||
"cidr": "10.43.0.1/32",
|
|
||||||
}
|
|
||||||
}
|
|
||||||
]
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"to": [
|
|
||||||
{
|
|
||||||
//TODO this ip is from the docker network that k3d is running on
|
|
||||||
//since k3d does not deploy kube-api-server as a pod it needs to ahve the ip
|
|
||||||
//address opened up
|
|
||||||
//need to find a way to automatically detect the ip address from the docker
|
|
||||||
//network
|
|
||||||
"ipBlock": {
|
|
||||||
"cidr": "172.18.0.0/16",
|
|
||||||
}
|
|
||||||
}
|
|
||||||
]
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"to": [
|
"to": [
|
||||||
{
|
{
|
||||||
@@ -410,12 +390,27 @@ impl K8sTenantManager {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl Clone for K8sTenantManager {
|
||||||
|
fn clone(&self) -> Self {
|
||||||
|
Self {
|
||||||
|
k8s_client: self.k8s_client.clone(),
|
||||||
|
k8s_tenant_config: self.k8s_tenant_config.clone(),
|
||||||
|
network_policy_strategy: self.network_policy_strategy.clone_box(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#[async_trait]
|
#[async_trait]
|
||||||
impl TenantManager for K8sTenantManager {
|
impl TenantManager for K8sTenantManager {
|
||||||
async fn provision_tenant(&self, config: &TenantConfig) -> Result<(), ExecutorError> {
|
async fn provision_tenant(&self, config: &TenantConfig) -> Result<(), ExecutorError> {
|
||||||
let namespace = self.build_namespace(config)?;
|
let namespace = self.build_namespace(config)?;
|
||||||
let resource_quota = self.build_resource_quota(config)?;
|
let resource_quota = self.build_resource_quota(config)?;
|
||||||
|
|
||||||
let network_policy = self.build_network_policy(config)?;
|
let network_policy = self.build_network_policy(config)?;
|
||||||
|
let network_policy = self
|
||||||
|
.network_policy_strategy
|
||||||
|
.adjust_policy(network_policy, config);
|
||||||
|
|
||||||
let resource_limit_range = self.build_limit_range(config)?;
|
let resource_limit_range = self.build_limit_range(config)?;
|
||||||
|
|
||||||
self.ensure_constraints(&namespace)?;
|
self.ensure_constraints(&namespace)?;
|
||||||
|
|||||||
@@ -1,11 +1,11 @@
|
|||||||
pub mod k8s;
|
pub mod k8s;
|
||||||
mod manager;
|
mod manager;
|
||||||
use std::str::FromStr;
|
pub mod network_policy;
|
||||||
|
|
||||||
pub use manager::*;
|
|
||||||
use serde::{Deserialize, Serialize};
|
|
||||||
|
|
||||||
use crate::data::Id;
|
use crate::data::Id;
|
||||||
|
pub use manager::*;
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
use std::str::FromStr;
|
||||||
|
|
||||||
#[derive(Debug, Clone, PartialEq, Serialize, Deserialize)] // Assuming serde for Scores
|
#[derive(Debug, Clone, PartialEq, Serialize, Deserialize)] // Assuming serde for Scores
|
||||||
pub struct TenantConfig {
|
pub struct TenantConfig {
|
||||||
|
|||||||
120
harmony/src/domain/topology/tenant/network_policy.rs
Normal file
120
harmony/src/domain/topology/tenant/network_policy.rs
Normal file
@@ -0,0 +1,120 @@
|
|||||||
|
use k8s_openapi::api::networking::v1::{
|
||||||
|
IPBlock, NetworkPolicy, NetworkPolicyEgressRule, NetworkPolicyPeer, NetworkPolicySpec,
|
||||||
|
};
|
||||||
|
|
||||||
|
use super::TenantConfig;
|
||||||
|
|
||||||
|
pub trait NetworkPolicyStrategy: Send + Sync + std::fmt::Debug {
|
||||||
|
fn clone_box(&self) -> Box<dyn NetworkPolicyStrategy>;
|
||||||
|
|
||||||
|
fn adjust_policy(&self, policy: NetworkPolicy, config: &TenantConfig) -> NetworkPolicy;
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Clone, Debug)]
|
||||||
|
pub struct NoopNetworkPolicyStrategy {}
|
||||||
|
|
||||||
|
impl NoopNetworkPolicyStrategy {
|
||||||
|
pub fn new() -> Self {
|
||||||
|
Self {}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Default for NoopNetworkPolicyStrategy {
|
||||||
|
fn default() -> Self {
|
||||||
|
Self::new()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl NetworkPolicyStrategy for NoopNetworkPolicyStrategy {
|
||||||
|
fn clone_box(&self) -> Box<dyn NetworkPolicyStrategy> {
|
||||||
|
Box::new(self.clone())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn adjust_policy(&self, policy: NetworkPolicy, _config: &TenantConfig) -> NetworkPolicy {
|
||||||
|
policy
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Clone, Debug)]
|
||||||
|
pub struct K3dNetworkPolicyStrategy {}
|
||||||
|
|
||||||
|
impl K3dNetworkPolicyStrategy {
|
||||||
|
pub fn new() -> Self {
|
||||||
|
Self {}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Default for K3dNetworkPolicyStrategy {
|
||||||
|
fn default() -> Self {
|
||||||
|
Self::new()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl NetworkPolicyStrategy for K3dNetworkPolicyStrategy {
|
||||||
|
fn clone_box(&self) -> Box<dyn NetworkPolicyStrategy> {
|
||||||
|
Box::new(self.clone())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn adjust_policy(&self, policy: NetworkPolicy, _config: &TenantConfig) -> NetworkPolicy {
|
||||||
|
let mut egress = policy
|
||||||
|
.spec
|
||||||
|
.clone()
|
||||||
|
.unwrap_or_default()
|
||||||
|
.egress
|
||||||
|
.clone()
|
||||||
|
.unwrap_or_default();
|
||||||
|
egress.push(NetworkPolicyEgressRule {
|
||||||
|
to: Some(vec![NetworkPolicyPeer {
|
||||||
|
ip_block: Some(IPBlock {
|
||||||
|
cidr: "172.18.0.0/16".into(), // TODO: query the IP range https://git.nationtech.io/NationTech/harmony/issues/108
|
||||||
|
..Default::default()
|
||||||
|
}),
|
||||||
|
..Default::default()
|
||||||
|
}]),
|
||||||
|
..Default::default()
|
||||||
|
});
|
||||||
|
|
||||||
|
NetworkPolicy {
|
||||||
|
spec: Some(NetworkPolicySpec {
|
||||||
|
egress: Some(egress),
|
||||||
|
..policy.spec.unwrap_or_default()
|
||||||
|
}),
|
||||||
|
..policy
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use k8s_openapi::api::networking::v1::{
|
||||||
|
IPBlock, NetworkPolicy, NetworkPolicyEgressRule, NetworkPolicyPeer, NetworkPolicySpec,
|
||||||
|
};
|
||||||
|
|
||||||
|
use super::{K3dNetworkPolicyStrategy, NetworkPolicyStrategy};
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
pub fn should_add_ip_block_for_k3d_harmony_server() {
|
||||||
|
let strategy = K3dNetworkPolicyStrategy::new();
|
||||||
|
|
||||||
|
let policy =
|
||||||
|
strategy.adjust_policy(NetworkPolicy::default(), &super::TenantConfig::default());
|
||||||
|
|
||||||
|
let expected_policy = NetworkPolicy {
|
||||||
|
spec: Some(NetworkPolicySpec {
|
||||||
|
egress: Some(vec![NetworkPolicyEgressRule {
|
||||||
|
to: Some(vec![NetworkPolicyPeer {
|
||||||
|
ip_block: Some(IPBlock {
|
||||||
|
cidr: "172.18.0.0/16".into(),
|
||||||
|
..Default::default()
|
||||||
|
}),
|
||||||
|
..Default::default()
|
||||||
|
}]),
|
||||||
|
..Default::default()
|
||||||
|
}]),
|
||||||
|
..Default::default()
|
||||||
|
}),
|
||||||
|
..Default::default()
|
||||||
|
};
|
||||||
|
assert_eq!(expected_policy, policy);
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,10 +1,10 @@
|
|||||||
use async_trait::async_trait;
|
use async_trait::async_trait;
|
||||||
use harmony_types::net::MacAddress;
|
use harmony_types::net::MacAddress;
|
||||||
use log::debug;
|
use log::info;
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
executors::ExecutorError,
|
executors::ExecutorError,
|
||||||
topology::{DHCPStaticEntry, DhcpServer, IpAddress, LogicalHost},
|
topology::{DHCPStaticEntry, DhcpServer, IpAddress, LogicalHost, PxeOptions},
|
||||||
};
|
};
|
||||||
|
|
||||||
use super::OPNSenseFirewall;
|
use super::OPNSenseFirewall;
|
||||||
@@ -26,7 +26,7 @@ impl DhcpServer for OPNSenseFirewall {
|
|||||||
.unwrap();
|
.unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
debug!("Registered {:?}", entry);
|
info!("Registered {:?}", entry);
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -46,57 +46,25 @@ impl DhcpServer for OPNSenseFirewall {
|
|||||||
self.host.clone()
|
self.host.clone()
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn set_next_server(&self, ip: IpAddress) -> Result<(), ExecutorError> {
|
async fn set_pxe_options(&self, options: PxeOptions) -> Result<(), ExecutorError> {
|
||||||
let ipv4 = match ip {
|
|
||||||
std::net::IpAddr::V4(ipv4_addr) => ipv4_addr,
|
|
||||||
std::net::IpAddr::V6(_) => todo!("ipv6 not supported yet"),
|
|
||||||
};
|
|
||||||
{
|
|
||||||
let mut writable_opnsense = self.opnsense_config.write().await;
|
let mut writable_opnsense = self.opnsense_config.write().await;
|
||||||
writable_opnsense.dhcp().set_next_server(ipv4);
|
let PxeOptions {
|
||||||
debug!("OPNsense dhcp server set next server {ipv4}");
|
ipxe_filename,
|
||||||
}
|
bios_filename,
|
||||||
|
efi_filename,
|
||||||
Ok(())
|
tftp_ip,
|
||||||
}
|
} = options;
|
||||||
|
writable_opnsense
|
||||||
async fn set_boot_filename(&self, boot_filename: &str) -> Result<(), ExecutorError> {
|
.dhcp()
|
||||||
{
|
.set_pxe_options(
|
||||||
let mut writable_opnsense = self.opnsense_config.write().await;
|
tftp_ip.map(|i| i.to_string()),
|
||||||
writable_opnsense.dhcp().set_boot_filename(boot_filename);
|
bios_filename,
|
||||||
debug!("OPNsense dhcp server set boot filename {boot_filename}");
|
efi_filename,
|
||||||
}
|
ipxe_filename,
|
||||||
|
)
|
||||||
Ok(())
|
.await
|
||||||
}
|
.map_err(|dhcp_error| {
|
||||||
|
ExecutorError::UnexpectedError(format!("Failed to set_pxe_options : {dhcp_error}"))
|
||||||
async fn set_filename(&self, filename: &str) -> Result<(), ExecutorError> {
|
})
|
||||||
{
|
|
||||||
let mut writable_opnsense = self.opnsense_config.write().await;
|
|
||||||
writable_opnsense.dhcp().set_filename(filename);
|
|
||||||
debug!("OPNsense dhcp server set filename {filename}");
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn set_filename64(&self, filename: &str) -> Result<(), ExecutorError> {
|
|
||||||
{
|
|
||||||
let mut writable_opnsense = self.opnsense_config.write().await;
|
|
||||||
writable_opnsense.dhcp().set_filename64(filename);
|
|
||||||
debug!("OPNsense dhcp server set filename {filename}");
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn set_filenameipxe(&self, filenameipxe: &str) -> Result<(), ExecutorError> {
|
|
||||||
{
|
|
||||||
let mut writable_opnsense = self.opnsense_config.write().await;
|
|
||||||
writable_opnsense.dhcp().set_filenameipxe(filenameipxe);
|
|
||||||
debug!("OPNsense dhcp server set filenameipxe {filenameipxe}");
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2,23 +2,23 @@ use async_trait::async_trait;
|
|||||||
use log::info;
|
use log::info;
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
|
data::FileContent,
|
||||||
executors::ExecutorError,
|
executors::ExecutorError,
|
||||||
topology::{HttpServer, IpAddress, Url},
|
topology::{HttpServer, IpAddress, Url},
|
||||||
};
|
};
|
||||||
|
|
||||||
use super::OPNSenseFirewall;
|
use super::OPNSenseFirewall;
|
||||||
|
const OPNSENSE_HTTP_ROOT_PATH: &str = "/usr/local/http";
|
||||||
|
|
||||||
#[async_trait]
|
#[async_trait]
|
||||||
impl HttpServer for OPNSenseFirewall {
|
impl HttpServer for OPNSenseFirewall {
|
||||||
async fn serve_files(&self, url: &Url) -> Result<(), ExecutorError> {
|
async fn serve_files(&self, url: &Url) -> Result<(), ExecutorError> {
|
||||||
let http_root_path = "/usr/local/http";
|
|
||||||
|
|
||||||
let config = self.opnsense_config.read().await;
|
let config = self.opnsense_config.read().await;
|
||||||
info!("Uploading files from url {url} to {http_root_path}");
|
info!("Uploading files from url {url} to {OPNSENSE_HTTP_ROOT_PATH}");
|
||||||
match url {
|
match url {
|
||||||
Url::LocalFolder(path) => {
|
Url::LocalFolder(path) => {
|
||||||
config
|
config
|
||||||
.upload_files(path, http_root_path)
|
.upload_files(path, OPNSENSE_HTTP_ROOT_PATH)
|
||||||
.await
|
.await
|
||||||
.map_err(|e| ExecutorError::UnexpectedError(e.to_string()))?;
|
.map_err(|e| ExecutorError::UnexpectedError(e.to_string()))?;
|
||||||
}
|
}
|
||||||
@@ -27,8 +27,29 @@ impl HttpServer for OPNSenseFirewall {
|
|||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async fn serve_file_content(&self, file: &FileContent) -> Result<(), ExecutorError> {
|
||||||
|
let path = match &file.path {
|
||||||
|
crate::data::FilePath::Relative(path) => {
|
||||||
|
format!("{OPNSENSE_HTTP_ROOT_PATH}/{}", path.to_string())
|
||||||
|
}
|
||||||
|
crate::data::FilePath::Absolute(path) => {
|
||||||
|
return Err(ExecutorError::ConfigurationError(format!(
|
||||||
|
"Cannot serve file from http server with absolute path : {path}"
|
||||||
|
)));
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
let config = self.opnsense_config.read().await;
|
||||||
|
info!("Uploading file content to {}", path);
|
||||||
|
config
|
||||||
|
.upload_file_content(&path, &file.content)
|
||||||
|
.await
|
||||||
|
.map_err(|e| ExecutorError::UnexpectedError(e.to_string()))?;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
fn get_ip(&self) -> IpAddress {
|
fn get_ip(&self) -> IpAddress {
|
||||||
todo!();
|
OPNSenseFirewall::get_ip(self)
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn commit_config(&self) -> Result<(), ExecutorError> {
|
async fn commit_config(&self) -> Result<(), ExecutorError> {
|
||||||
|
|||||||
@@ -28,7 +28,7 @@ impl TftpServer for OPNSenseFirewall {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn get_ip(&self) -> IpAddress {
|
fn get_ip(&self) -> IpAddress {
|
||||||
todo!()
|
OPNSenseFirewall::get_ip(self)
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn set_ip(&self, ip: IpAddress) -> Result<(), ExecutorError> {
|
async fn set_ip(&self, ip: IpAddress) -> Result<(), ExecutorError> {
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
use std::{io::Write, process::Command, sync::Arc};
|
use std::{io::Write, process::Command, sync::Arc};
|
||||||
|
|
||||||
use async_trait::async_trait;
|
use async_trait::async_trait;
|
||||||
use log::{debug, error};
|
use log::info;
|
||||||
use serde_yaml::Value;
|
use serde_yaml::Value;
|
||||||
use tempfile::NamedTempFile;
|
use tempfile::NamedTempFile;
|
||||||
|
|
||||||
@@ -56,14 +56,11 @@ impl<A: OCICompliant + HelmPackage> ContinuousDelivery<A> {
|
|||||||
chart_url: String,
|
chart_url: String,
|
||||||
image_name: String,
|
image_name: String,
|
||||||
) -> Result<(), String> {
|
) -> Result<(), String> {
|
||||||
error!(
|
// TODO: This works only with local k3d installations, which is fine only for current demo purposes. We assume usage of K8sAnywhereTopology"
|
||||||
"FIXME This works only with local k3d installations, which is fine only for current demo purposes. We assume usage of K8sAnywhereTopology"
|
// https://git.nationtech.io/NationTech/harmony/issues/106
|
||||||
);
|
|
||||||
|
|
||||||
error!("TODO hardcoded k3d bin path is wrong");
|
|
||||||
let k3d_bin_path = (*HARMONY_DATA_DIR).join("k3d").join("k3d");
|
let k3d_bin_path = (*HARMONY_DATA_DIR).join("k3d").join("k3d");
|
||||||
// --- 1. Import the container image into the k3d cluster ---
|
// --- 1. Import the container image into the k3d cluster ---
|
||||||
debug!(
|
info!(
|
||||||
"Importing image '{}' into k3d cluster 'harmony'",
|
"Importing image '{}' into k3d cluster 'harmony'",
|
||||||
image_name
|
image_name
|
||||||
);
|
);
|
||||||
@@ -80,7 +77,7 @@ impl<A: OCICompliant + HelmPackage> ContinuousDelivery<A> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// --- 2. Get the kubeconfig for the k3d cluster and write it to a temp file ---
|
// --- 2. Get the kubeconfig for the k3d cluster and write it to a temp file ---
|
||||||
debug!("Retrieving kubeconfig for k3d cluster 'harmony'");
|
info!("Retrieving kubeconfig for k3d cluster 'harmony'");
|
||||||
let kubeconfig_output = Command::new(&k3d_bin_path)
|
let kubeconfig_output = Command::new(&k3d_bin_path)
|
||||||
.args(["kubeconfig", "get", "harmony"])
|
.args(["kubeconfig", "get", "harmony"])
|
||||||
.output()
|
.output()
|
||||||
@@ -101,7 +98,7 @@ impl<A: OCICompliant + HelmPackage> ContinuousDelivery<A> {
|
|||||||
let kubeconfig_path = temp_kubeconfig.path().to_str().unwrap();
|
let kubeconfig_path = temp_kubeconfig.path().to_str().unwrap();
|
||||||
|
|
||||||
// --- 3. Install or upgrade the Helm chart in the cluster ---
|
// --- 3. Install or upgrade the Helm chart in the cluster ---
|
||||||
debug!(
|
info!(
|
||||||
"Deploying Helm chart '{}' to namespace '{}'",
|
"Deploying Helm chart '{}' to namespace '{}'",
|
||||||
chart_url, app_name
|
chart_url, app_name
|
||||||
);
|
);
|
||||||
@@ -131,7 +128,7 @@ impl<A: OCICompliant + HelmPackage> ContinuousDelivery<A> {
|
|||||||
));
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
debug!("Successfully deployed '{}' to local k3d cluster.", app_name);
|
info!("Successfully deployed '{}' to local k3d cluster.", app_name);
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -151,14 +148,12 @@ impl<
|
|||||||
// Or ask for it when unknown
|
// Or ask for it when unknown
|
||||||
|
|
||||||
let helm_chart = self.application.build_push_helm_package(&image).await?;
|
let helm_chart = self.application.build_push_helm_package(&image).await?;
|
||||||
debug!("Pushed new helm chart {helm_chart}");
|
|
||||||
|
|
||||||
error!("TODO Make building image configurable/skippable if image already exists (prompt)");
|
// TODO: Make building image configurable/skippable if image already exists (prompt)")
|
||||||
|
// https://git.nationtech.io/NationTech/harmony/issues/104
|
||||||
let image = self.application.build_push_oci_image().await?;
|
let image = self.application.build_push_oci_image().await?;
|
||||||
debug!("Pushed new docker image {image}");
|
|
||||||
|
|
||||||
debug!("Installing ContinuousDelivery feature");
|
// TODO: this is a temporary hack for demo purposes, the deployment target should be driven
|
||||||
// TODO this is a temporary hack for demo purposes, the deployment target should be driven
|
|
||||||
// by the topology only and we should not have to know how to perform tasks like this for
|
// by the topology only and we should not have to know how to perform tasks like this for
|
||||||
// which the topology should be responsible.
|
// which the topology should be responsible.
|
||||||
//
|
//
|
||||||
@@ -171,17 +166,20 @@ impl<
|
|||||||
// access it. This forces every Topology to understand the concept of targets though... So
|
// access it. This forces every Topology to understand the concept of targets though... So
|
||||||
// instead I'll create a new Capability which is MultiTargetTopology and we'll see how it
|
// instead I'll create a new Capability which is MultiTargetTopology and we'll see how it
|
||||||
// goes. It still does not feel right though.
|
// goes. It still does not feel right though.
|
||||||
|
//
|
||||||
|
// https://git.nationtech.io/NationTech/harmony/issues/106
|
||||||
match topology.current_target() {
|
match topology.current_target() {
|
||||||
DeploymentTarget::LocalDev => {
|
DeploymentTarget::LocalDev => {
|
||||||
|
info!("Deploying {} locally...", self.application.name());
|
||||||
self.deploy_to_local_k3d(self.application.name(), helm_chart, image)
|
self.deploy_to_local_k3d(self.application.name(), helm_chart, image)
|
||||||
.await?;
|
.await?;
|
||||||
}
|
}
|
||||||
target => {
|
target => {
|
||||||
debug!("Deploying to target {target:?}");
|
info!("Deploying {} to target {target:?}", self.application.name());
|
||||||
let score = ArgoHelmScore {
|
let score = ArgoHelmScore {
|
||||||
namespace: "harmonydemo-staging".to_string(),
|
namespace: "harmony-example-rust-webapp".to_string(),
|
||||||
openshift: false,
|
openshift: true,
|
||||||
domain: "argo.harmonydemo.apps.st.mcd".to_string(),
|
domain: "argo.harmonydemo.apps.ncd0.harmony.mcd".to_string(),
|
||||||
argo_apps: vec![ArgoApplication::from(CDApplicationConfig {
|
argo_apps: vec![ArgoApplication::from(CDApplicationConfig {
|
||||||
// helm pull oci://hub.nationtech.io/harmony/harmony-example-rust-webapp-chart --version 0.1.0
|
// helm pull oci://hub.nationtech.io/harmony/harmony-example-rust-webapp-chart --version 0.1.0
|
||||||
version: Version::from("0.1.0").unwrap(),
|
version: Version::from("0.1.0").unwrap(),
|
||||||
@@ -189,7 +187,7 @@ impl<
|
|||||||
helm_chart_name: "harmony-example-rust-webapp-chart".to_string(),
|
helm_chart_name: "harmony-example-rust-webapp-chart".to_string(),
|
||||||
values_overrides: None,
|
values_overrides: None,
|
||||||
name: "harmony-demo-rust-webapp".to_string(),
|
name: "harmony-demo-rust-webapp".to_string(),
|
||||||
namespace: "harmonydemo-staging".to_string(),
|
namespace: "harmony-example-rust-webapp".to_string(),
|
||||||
})],
|
})],
|
||||||
};
|
};
|
||||||
score
|
score
|
||||||
|
|||||||
@@ -1,5 +1,4 @@
|
|||||||
use async_trait::async_trait;
|
use async_trait::async_trait;
|
||||||
use log::error;
|
|
||||||
use non_blank_string_rs::NonBlankString;
|
use non_blank_string_rs::NonBlankString;
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
use std::str::FromStr;
|
use std::str::FromStr;
|
||||||
@@ -50,7 +49,6 @@ impl<T: Topology + K8sclient + HelmCommand> Interpret<T> for ArgoInterpret {
|
|||||||
inventory: &Inventory,
|
inventory: &Inventory,
|
||||||
topology: &T,
|
topology: &T,
|
||||||
) -> Result<Outcome, InterpretError> {
|
) -> Result<Outcome, InterpretError> {
|
||||||
error!("Uncomment below, only disabled for debugging");
|
|
||||||
self.score.interpret(inventory, topology).await?;
|
self.score.interpret(inventory, topology).await?;
|
||||||
|
|
||||||
let k8s_client = topology.k8s_client().await?;
|
let k8s_client = topology.k8s_client().await?;
|
||||||
@@ -58,9 +56,14 @@ impl<T: Topology + K8sclient + HelmCommand> Interpret<T> for ArgoInterpret {
|
|||||||
.apply_yaml_many(&self.argo_apps.iter().map(|a| a.to_yaml()).collect(), None)
|
.apply_yaml_many(&self.argo_apps.iter().map(|a| a.to_yaml()).collect(), None)
|
||||||
.await
|
.await
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
Ok(Outcome::success(format!(
|
Ok(Outcome::success(format!(
|
||||||
"Successfully installed ArgoCD and {} Applications",
|
"ArgoCD installed with {} {}",
|
||||||
self.argo_apps.len()
|
self.argo_apps.len(),
|
||||||
|
match self.argo_apps.len() {
|
||||||
|
1 => "application",
|
||||||
|
_ => "applications",
|
||||||
|
}
|
||||||
)))
|
)))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -4,6 +4,7 @@ use crate::modules::application::{Application, ApplicationFeature};
|
|||||||
use crate::modules::monitoring::application_monitoring::application_monitoring_score::ApplicationMonitoringScore;
|
use crate::modules::monitoring::application_monitoring::application_monitoring_score::ApplicationMonitoringScore;
|
||||||
use crate::modules::monitoring::kube_prometheus::crd::crd_alertmanager_config::CRDPrometheus;
|
use crate::modules::monitoring::kube_prometheus::crd::crd_alertmanager_config::CRDPrometheus;
|
||||||
|
|
||||||
|
use crate::topology::MultiTargetTopology;
|
||||||
use crate::{
|
use crate::{
|
||||||
inventory::Inventory,
|
inventory::Inventory,
|
||||||
modules::monitoring::{
|
modules::monitoring::{
|
||||||
@@ -33,6 +34,7 @@ impl<
|
|||||||
+ 'static
|
+ 'static
|
||||||
+ TenantManager
|
+ TenantManager
|
||||||
+ K8sclient
|
+ K8sclient
|
||||||
|
+ MultiTargetTopology
|
||||||
+ std::fmt::Debug
|
+ std::fmt::Debug
|
||||||
+ PrometheusApplicationMonitoring<CRDPrometheus>,
|
+ PrometheusApplicationMonitoring<CRDPrometheus>,
|
||||||
> ApplicationFeature<T> for Monitoring
|
> ApplicationFeature<T> for Monitoring
|
||||||
@@ -55,11 +57,11 @@ impl<
|
|||||||
};
|
};
|
||||||
let ntfy = NtfyScore {
|
let ntfy = NtfyScore {
|
||||||
namespace: namespace.clone(),
|
namespace: namespace.clone(),
|
||||||
host: "localhost".to_string(),
|
host: "ntfy.harmonydemo.apps.ncd0.harmony.mcd".to_string(),
|
||||||
};
|
};
|
||||||
ntfy.interpret(&Inventory::empty(), topology)
|
ntfy.interpret(&Inventory::empty(), topology)
|
||||||
.await
|
.await
|
||||||
.expect("couldn't create interpret for ntfy");
|
.map_err(|e| e.to_string())?;
|
||||||
|
|
||||||
let ntfy_default_auth_username = "harmony";
|
let ntfy_default_auth_username = "harmony";
|
||||||
let ntfy_default_auth_password = "harmony";
|
let ntfy_default_auth_password = "harmony";
|
||||||
@@ -96,7 +98,7 @@ impl<
|
|||||||
alerting_score
|
alerting_score
|
||||||
.interpret(&Inventory::empty(), topology)
|
.interpret(&Inventory::empty(), topology)
|
||||||
.await
|
.await
|
||||||
.unwrap();
|
.map_err(|e| e.to_string())?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
fn name(&self) -> String {
|
fn name(&self) -> String {
|
||||||
|
|||||||
@@ -14,11 +14,19 @@ use serde::Serialize;
|
|||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
data::{Id, Version},
|
data::{Id, Version},
|
||||||
|
instrumentation::{self, HarmonyEvent},
|
||||||
interpret::{Interpret, InterpretError, InterpretName, InterpretStatus, Outcome},
|
interpret::{Interpret, InterpretError, InterpretName, InterpretStatus, Outcome},
|
||||||
inventory::Inventory,
|
inventory::Inventory,
|
||||||
topology::Topology,
|
topology::Topology,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
#[derive(Clone, Debug)]
|
||||||
|
pub enum ApplicationFeatureStatus {
|
||||||
|
Installing,
|
||||||
|
Installed,
|
||||||
|
Failed { details: String },
|
||||||
|
}
|
||||||
|
|
||||||
pub trait Application: std::fmt::Debug + Send + Sync {
|
pub trait Application: std::fmt::Debug + Send + Sync {
|
||||||
fn name(&self) -> String;
|
fn name(&self) -> String;
|
||||||
}
|
}
|
||||||
@@ -47,20 +55,41 @@ impl<A: Application, T: Topology + std::fmt::Debug> Interpret<T> for Application
|
|||||||
.join(", ")
|
.join(", ")
|
||||||
);
|
);
|
||||||
for feature in self.features.iter() {
|
for feature in self.features.iter() {
|
||||||
debug!(
|
instrumentation::instrument(HarmonyEvent::ApplicationFeatureStateChanged {
|
||||||
"Installing feature {} for application {app_name}",
|
topology: topology.name().into(),
|
||||||
feature.name()
|
application: self.application.name(),
|
||||||
);
|
feature: feature.name(),
|
||||||
|
status: ApplicationFeatureStatus::Installing,
|
||||||
|
})
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
let _ = match feature.ensure_installed(topology).await {
|
let _ = match feature.ensure_installed(topology).await {
|
||||||
Ok(()) => (),
|
Ok(()) => {
|
||||||
|
instrumentation::instrument(HarmonyEvent::ApplicationFeatureStateChanged {
|
||||||
|
topology: topology.name().into(),
|
||||||
|
application: self.application.name(),
|
||||||
|
feature: feature.name(),
|
||||||
|
status: ApplicationFeatureStatus::Installed,
|
||||||
|
})
|
||||||
|
.unwrap();
|
||||||
|
}
|
||||||
Err(msg) => {
|
Err(msg) => {
|
||||||
|
instrumentation::instrument(HarmonyEvent::ApplicationFeatureStateChanged {
|
||||||
|
topology: topology.name().into(),
|
||||||
|
application: self.application.name(),
|
||||||
|
feature: feature.name(),
|
||||||
|
status: ApplicationFeatureStatus::Failed {
|
||||||
|
details: msg.clone(),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
.unwrap();
|
||||||
return Err(InterpretError::new(format!(
|
return Err(InterpretError::new(format!(
|
||||||
"Application Interpret failed to install feature : {msg}"
|
"Application Interpret failed to install feature : {msg}"
|
||||||
)));
|
)));
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
Ok(Outcome::success("successfully created app".to_string()))
|
Ok(Outcome::success("Application created".to_string()))
|
||||||
}
|
}
|
||||||
|
|
||||||
fn get_name(&self) -> InterpretName {
|
fn get_name(&self) -> InterpretName {
|
||||||
|
|||||||
@@ -10,7 +10,7 @@ use dockerfile_builder::Dockerfile;
|
|||||||
use dockerfile_builder::instruction::{CMD, COPY, ENV, EXPOSE, FROM, RUN, USER, WORKDIR};
|
use dockerfile_builder::instruction::{CMD, COPY, ENV, EXPOSE, FROM, RUN, USER, WORKDIR};
|
||||||
use dockerfile_builder::instruction_builder::CopyBuilder;
|
use dockerfile_builder::instruction_builder::CopyBuilder;
|
||||||
use futures_util::StreamExt;
|
use futures_util::StreamExt;
|
||||||
use log::{debug, error, log_enabled};
|
use log::{debug, info, log_enabled};
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
use tar::Archive;
|
use tar::Archive;
|
||||||
|
|
||||||
@@ -46,7 +46,7 @@ where
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn name(&self) -> String {
|
fn name(&self) -> String {
|
||||||
format!("Application: {}", self.application.name())
|
format!("{} [ApplicationScore]", self.application.name())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -73,19 +73,19 @@ impl Application for RustWebapp {
|
|||||||
#[async_trait]
|
#[async_trait]
|
||||||
impl HelmPackage for RustWebapp {
|
impl HelmPackage for RustWebapp {
|
||||||
async fn build_push_helm_package(&self, image_url: &str) -> Result<String, String> {
|
async fn build_push_helm_package(&self, image_url: &str) -> Result<String, String> {
|
||||||
debug!("Starting Helm chart build and push for '{}'", self.name);
|
info!("Starting Helm chart build and push for '{}'", self.name);
|
||||||
|
|
||||||
// 1. Create the Helm chart files on disk.
|
// 1. Create the Helm chart files on disk.
|
||||||
let chart_dir = self
|
let chart_dir = self
|
||||||
.create_helm_chart_files(image_url)
|
.create_helm_chart_files(image_url)
|
||||||
.map_err(|e| format!("Failed to create Helm chart files: {}", e))?;
|
.map_err(|e| format!("Failed to create Helm chart files: {}", e))?;
|
||||||
debug!("Successfully created Helm chart files in {:?}", chart_dir);
|
info!("Successfully created Helm chart files in {:?}", chart_dir);
|
||||||
|
|
||||||
// 2. Package the chart into a .tgz archive.
|
// 2. Package the chart into a .tgz archive.
|
||||||
let packaged_chart_path = self
|
let packaged_chart_path = self
|
||||||
.package_helm_chart(&chart_dir)
|
.package_helm_chart(&chart_dir)
|
||||||
.map_err(|e| format!("Failed to package Helm chart: {}", e))?;
|
.map_err(|e| format!("Failed to package Helm chart: {}", e))?;
|
||||||
debug!(
|
info!(
|
||||||
"Successfully packaged Helm chart: {}",
|
"Successfully packaged Helm chart: {}",
|
||||||
packaged_chart_path.to_string_lossy()
|
packaged_chart_path.to_string_lossy()
|
||||||
);
|
);
|
||||||
@@ -94,7 +94,7 @@ impl HelmPackage for RustWebapp {
|
|||||||
let oci_chart_url = self
|
let oci_chart_url = self
|
||||||
.push_helm_chart(&packaged_chart_path)
|
.push_helm_chart(&packaged_chart_path)
|
||||||
.map_err(|e| format!("Failed to push Helm chart: {}", e))?;
|
.map_err(|e| format!("Failed to push Helm chart: {}", e))?;
|
||||||
debug!("Successfully pushed Helm chart to: {}", oci_chart_url);
|
info!("Successfully pushed Helm chart to: {}", oci_chart_url);
|
||||||
|
|
||||||
Ok(oci_chart_url)
|
Ok(oci_chart_url)
|
||||||
}
|
}
|
||||||
@@ -107,20 +107,20 @@ impl OCICompliant for RustWebapp {
|
|||||||
async fn build_push_oci_image(&self) -> Result<String, String> {
|
async fn build_push_oci_image(&self) -> Result<String, String> {
|
||||||
// This function orchestrates the build and push process.
|
// This function orchestrates the build and push process.
|
||||||
// It's async to match the trait definition, though the underlying docker commands are blocking.
|
// It's async to match the trait definition, though the underlying docker commands are blocking.
|
||||||
debug!("Starting OCI image build and push for '{}'", self.name);
|
info!("Starting OCI image build and push for '{}'", self.name);
|
||||||
|
|
||||||
// 1. Build the image by calling the synchronous helper function.
|
// 1. Build the image by calling the synchronous helper function.
|
||||||
let image_tag = self.image_name();
|
let image_tag = self.image_name();
|
||||||
self.build_docker_image(&image_tag)
|
self.build_docker_image(&image_tag)
|
||||||
.await
|
.await
|
||||||
.map_err(|e| format!("Failed to build Docker image: {}", e))?;
|
.map_err(|e| format!("Failed to build Docker image: {}", e))?;
|
||||||
debug!("Successfully built Docker image: {}", image_tag);
|
info!("Successfully built Docker image: {}", image_tag);
|
||||||
|
|
||||||
// 2. Push the image to the registry.
|
// 2. Push the image to the registry.
|
||||||
self.push_docker_image(&image_tag)
|
self.push_docker_image(&image_tag)
|
||||||
.await
|
.await
|
||||||
.map_err(|e| format!("Failed to push Docker image: {}", e))?;
|
.map_err(|e| format!("Failed to push Docker image: {}", e))?;
|
||||||
debug!("Successfully pushed Docker image to: {}", image_tag);
|
info!("Successfully pushed Docker image to: {}", image_tag);
|
||||||
|
|
||||||
Ok(image_tag)
|
Ok(image_tag)
|
||||||
}
|
}
|
||||||
@@ -195,7 +195,7 @@ impl RustWebapp {
|
|||||||
);
|
);
|
||||||
|
|
||||||
while let Some(msg) = image_build_stream.next().await {
|
while let Some(msg) = image_build_stream.next().await {
|
||||||
println!("Message: {msg:?}");
|
debug!("Message: {msg:?}");
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(image_name.to_string())
|
Ok(image_name.to_string())
|
||||||
@@ -219,7 +219,7 @@ impl RustWebapp {
|
|||||||
);
|
);
|
||||||
|
|
||||||
while let Some(msg) = push_image_stream.next().await {
|
while let Some(msg) = push_image_stream.next().await {
|
||||||
println!("Message: {msg:?}");
|
debug!("Message: {msg:?}");
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(image_tag.to_string())
|
Ok(image_tag.to_string())
|
||||||
@@ -288,9 +288,8 @@ impl RustWebapp {
|
|||||||
.unwrap(),
|
.unwrap(),
|
||||||
);
|
);
|
||||||
// Copy the compiled binary from the builder stage.
|
// Copy the compiled binary from the builder stage.
|
||||||
error!(
|
// TODO: Should not be using score name here, instead should use name from Cargo.toml
|
||||||
"FIXME Should not be using score name here, instead should use name from Cargo.toml"
|
// https://git.nationtech.io/NationTech/harmony/issues/105
|
||||||
);
|
|
||||||
let binary_path_in_builder = format!("/app/target/release/{}", self.name);
|
let binary_path_in_builder = format!("/app/target/release/{}", self.name);
|
||||||
let binary_path_in_final = format!("/home/appuser/{}", self.name);
|
let binary_path_in_final = format!("/home/appuser/{}", self.name);
|
||||||
dockerfile.push(
|
dockerfile.push(
|
||||||
@@ -328,9 +327,8 @@ impl RustWebapp {
|
|||||||
));
|
));
|
||||||
|
|
||||||
// Copy only the compiled binary from the builder stage.
|
// Copy only the compiled binary from the builder stage.
|
||||||
error!(
|
// TODO: Should not be using score name here, instead should use name from Cargo.toml
|
||||||
"FIXME Should not be using score name here, instead should use name from Cargo.toml"
|
// https://git.nationtech.io/NationTech/harmony/issues/105
|
||||||
);
|
|
||||||
let binary_path_in_builder = format!("/app/target/release/{}", self.name);
|
let binary_path_in_builder = format!("/app/target/release/{}", self.name);
|
||||||
let binary_path_in_final = format!("/usr/local/bin/{}", self.name);
|
let binary_path_in_final = format!("/usr/local/bin/{}", self.name);
|
||||||
dockerfile.push(
|
dockerfile.push(
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ use crate::{
|
|||||||
domain::{data::Version, interpret::InterpretStatus},
|
domain::{data::Version, interpret::InterpretStatus},
|
||||||
interpret::{Interpret, InterpretError, InterpretName, Outcome},
|
interpret::{Interpret, InterpretError, InterpretName, Outcome},
|
||||||
inventory::Inventory,
|
inventory::Inventory,
|
||||||
topology::{DHCPStaticEntry, DhcpServer, HostBinding, IpAddress, Topology},
|
topology::{DHCPStaticEntry, DhcpServer, HostBinding, IpAddress, PxeOptions, Topology},
|
||||||
};
|
};
|
||||||
|
|
||||||
use crate::domain::score::Score;
|
use crate::domain::score::Score;
|
||||||
@@ -98,69 +98,14 @@ impl DhcpInterpret {
|
|||||||
_inventory: &Inventory,
|
_inventory: &Inventory,
|
||||||
dhcp_server: &D,
|
dhcp_server: &D,
|
||||||
) -> Result<Outcome, InterpretError> {
|
) -> Result<Outcome, InterpretError> {
|
||||||
let next_server_outcome = match self.score.next_server {
|
let pxe_options = PxeOptions {
|
||||||
Some(next_server) => {
|
ipxe_filename: self.score.filenameipxe.clone().unwrap_or_default(),
|
||||||
dhcp_server.set_next_server(next_server).await?;
|
bios_filename: self.score.filename.clone().unwrap_or_default(),
|
||||||
Outcome::new(
|
efi_filename: self.score.filename64.clone().unwrap_or_default(),
|
||||||
InterpretStatus::SUCCESS,
|
tftp_ip: self.score.next_server,
|
||||||
format!("Dhcp Interpret Set next boot to {next_server}"),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
None => Outcome::noop(),
|
|
||||||
};
|
};
|
||||||
|
|
||||||
let boot_filename_outcome = match &self.score.boot_filename {
|
dhcp_server.set_pxe_options(pxe_options).await?;
|
||||||
Some(boot_filename) => {
|
|
||||||
dhcp_server.set_boot_filename(boot_filename).await?;
|
|
||||||
Outcome::new(
|
|
||||||
InterpretStatus::SUCCESS,
|
|
||||||
format!("Dhcp Interpret Set boot filename to {boot_filename}"),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
None => Outcome::noop(),
|
|
||||||
};
|
|
||||||
|
|
||||||
let filename_outcome = match &self.score.filename {
|
|
||||||
Some(filename) => {
|
|
||||||
dhcp_server.set_filename(filename).await?;
|
|
||||||
Outcome::new(
|
|
||||||
InterpretStatus::SUCCESS,
|
|
||||||
format!("Dhcp Interpret Set filename to {filename}"),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
None => Outcome::noop(),
|
|
||||||
};
|
|
||||||
|
|
||||||
let filename64_outcome = match &self.score.filename64 {
|
|
||||||
Some(filename64) => {
|
|
||||||
dhcp_server.set_filename64(filename64).await?;
|
|
||||||
Outcome::new(
|
|
||||||
InterpretStatus::SUCCESS,
|
|
||||||
format!("Dhcp Interpret Set filename64 to {filename64}"),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
None => Outcome::noop(),
|
|
||||||
};
|
|
||||||
|
|
||||||
let filenameipxe_outcome = match &self.score.filenameipxe {
|
|
||||||
Some(filenameipxe) => {
|
|
||||||
dhcp_server.set_filenameipxe(filenameipxe).await?;
|
|
||||||
Outcome::new(
|
|
||||||
InterpretStatus::SUCCESS,
|
|
||||||
format!("Dhcp Interpret Set filenameipxe to {filenameipxe}"),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
None => Outcome::noop(),
|
|
||||||
};
|
|
||||||
|
|
||||||
if next_server_outcome.status == InterpretStatus::NOOP
|
|
||||||
&& boot_filename_outcome.status == InterpretStatus::NOOP
|
|
||||||
&& filename_outcome.status == InterpretStatus::NOOP
|
|
||||||
&& filename64_outcome.status == InterpretStatus::NOOP
|
|
||||||
&& filenameipxe_outcome.status == InterpretStatus::NOOP
|
|
||||||
{
|
|
||||||
return Ok(Outcome::noop());
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(Outcome::new(
|
Ok(Outcome::new(
|
||||||
InterpretStatus::SUCCESS,
|
InterpretStatus::SUCCESS,
|
||||||
|
|||||||
@@ -55,7 +55,7 @@ impl<T: Topology + HelmCommand> Score<T> for HelmChartScore {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn name(&self) -> String {
|
fn name(&self) -> String {
|
||||||
format!("{} {} HelmChartScore", self.release_name, self.chart_name)
|
format!("{} [HelmChartScore]", self.release_name)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -225,19 +225,20 @@ impl<T: Topology + HelmCommand> Interpret<T> for HelmChartInterpret {
|
|||||||
match status {
|
match status {
|
||||||
helm_wrapper_rs::HelmDeployStatus::Deployed => Ok(Outcome::new(
|
helm_wrapper_rs::HelmDeployStatus::Deployed => Ok(Outcome::new(
|
||||||
InterpretStatus::SUCCESS,
|
InterpretStatus::SUCCESS,
|
||||||
"Helm Chart deployed".to_string(),
|
format!("Helm Chart {} deployed", self.score.release_name),
|
||||||
)),
|
)),
|
||||||
helm_wrapper_rs::HelmDeployStatus::PendingInstall => Ok(Outcome::new(
|
helm_wrapper_rs::HelmDeployStatus::PendingInstall => Ok(Outcome::new(
|
||||||
InterpretStatus::RUNNING,
|
InterpretStatus::RUNNING,
|
||||||
"Helm Chart Pending install".to_string(),
|
format!("Helm Chart {} pending install...", self.score.release_name),
|
||||||
)),
|
)),
|
||||||
helm_wrapper_rs::HelmDeployStatus::PendingUpgrade => Ok(Outcome::new(
|
helm_wrapper_rs::HelmDeployStatus::PendingUpgrade => Ok(Outcome::new(
|
||||||
InterpretStatus::RUNNING,
|
InterpretStatus::RUNNING,
|
||||||
"Helm Chart pending upgrade".to_string(),
|
format!("Helm Chart {} pending upgrade...", self.score.release_name),
|
||||||
)),
|
|
||||||
helm_wrapper_rs::HelmDeployStatus::Failed => Err(InterpretError::new(
|
|
||||||
"Failed to install helm chart".to_string(),
|
|
||||||
)),
|
)),
|
||||||
|
helm_wrapper_rs::HelmDeployStatus::Failed => Err(InterpretError::new(format!(
|
||||||
|
"Helm Chart {} installation failed",
|
||||||
|
self.score.release_name
|
||||||
|
))),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ use derive_new::new;
|
|||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
data::{Id, Version},
|
data::{FileContent, Id, Version},
|
||||||
interpret::{Interpret, InterpretError, InterpretName, InterpretStatus, Outcome},
|
interpret::{Interpret, InterpretError, InterpretName, InterpretStatus, Outcome},
|
||||||
inventory::Inventory,
|
inventory::Inventory,
|
||||||
score::Score,
|
score::Score,
|
||||||
@@ -23,7 +23,8 @@ use crate::{
|
|||||||
/// ```
|
/// ```
|
||||||
#[derive(Debug, new, Clone, Serialize)]
|
#[derive(Debug, new, Clone, Serialize)]
|
||||||
pub struct StaticFilesHttpScore {
|
pub struct StaticFilesHttpScore {
|
||||||
files_to_serve: Url,
|
pub folder_to_serve: Option<Url>,
|
||||||
|
pub files: Vec<FileContent>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<T: Topology + HttpServer> Score<T> for StaticFilesHttpScore {
|
impl<T: Topology + HttpServer> Score<T> for StaticFilesHttpScore {
|
||||||
@@ -50,12 +51,25 @@ impl<T: Topology + HttpServer> Interpret<T> for StaticFilesHttpInterpret {
|
|||||||
) -> Result<Outcome, InterpretError> {
|
) -> Result<Outcome, InterpretError> {
|
||||||
http_server.ensure_initialized().await?;
|
http_server.ensure_initialized().await?;
|
||||||
// http_server.set_ip(topology.router.get_gateway()).await?;
|
// http_server.set_ip(topology.router.get_gateway()).await?;
|
||||||
http_server.serve_files(&self.score.files_to_serve).await?;
|
if let Some(folder) = self.score.folder_to_serve.as_ref() {
|
||||||
|
http_server.serve_files(folder).await?;
|
||||||
|
}
|
||||||
|
|
||||||
|
for f in self.score.files.iter() {
|
||||||
|
http_server.serve_file_content(&f).await?
|
||||||
|
}
|
||||||
|
|
||||||
http_server.commit_config().await?;
|
http_server.commit_config().await?;
|
||||||
http_server.reload_restart().await?;
|
http_server.reload_restart().await?;
|
||||||
Ok(Outcome::success(format!(
|
Ok(Outcome::success(format!(
|
||||||
"Http Server running and serving files from {}",
|
"Http Server running and serving files from folder {:?} and content for {}",
|
||||||
self.score.files_to_serve
|
self.score.folder_to_serve,
|
||||||
|
self.score
|
||||||
|
.files
|
||||||
|
.iter()
|
||||||
|
.map(|f| f.path.to_string())
|
||||||
|
.collect::<Vec<String>>()
|
||||||
|
.join(",")
|
||||||
)))
|
)))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
72
harmony/src/modules/inventory/mod.rs
Normal file
72
harmony/src/modules/inventory/mod.rs
Normal file
@@ -0,0 +1,72 @@
|
|||||||
|
use async_trait::async_trait;
|
||||||
|
use harmony_inventory_agent::local_presence::DiscoveryEvent;
|
||||||
|
use log::info;
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
data::{Id, Version},
|
||||||
|
interpret::{Interpret, InterpretError, InterpretName, InterpretStatus, Outcome},
|
||||||
|
inventory::Inventory,
|
||||||
|
score::Score,
|
||||||
|
topology::Topology,
|
||||||
|
};
|
||||||
|
|
||||||
|
/// This launches an harmony_inventory_agent discovery process
|
||||||
|
/// This will allow us to register/update hosts running harmony_inventory_agent
|
||||||
|
/// from LAN in the Harmony inventory
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||||
|
pub struct DiscoverInventoryAgentScore {
|
||||||
|
pub discovery_timeout: Option<u64>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T: Topology> Score<T> for DiscoverInventoryAgentScore {
|
||||||
|
fn name(&self) -> String {
|
||||||
|
"DiscoverInventoryAgentScore".to_string()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn create_interpret(&self) -> Box<dyn Interpret<T>> {
|
||||||
|
Box::new(DiscoverInventoryAgentInterpret {
|
||||||
|
score: self.clone(),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
struct DiscoverInventoryAgentInterpret {
|
||||||
|
score: DiscoverInventoryAgentScore,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[async_trait]
|
||||||
|
impl<T: Topology> Interpret<T> for DiscoverInventoryAgentInterpret {
|
||||||
|
async fn execute(
|
||||||
|
&self,
|
||||||
|
inventory: &Inventory,
|
||||||
|
topology: &T,
|
||||||
|
) -> Result<Outcome, InterpretError> {
|
||||||
|
harmony_inventory_agent::local_presence::discover_agents(
|
||||||
|
self.score.discovery_timeout,
|
||||||
|
on_discover_event,
|
||||||
|
);
|
||||||
|
todo!()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_name(&self) -> InterpretName {
|
||||||
|
InterpretName::DiscoverInventoryAgent
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_version(&self) -> Version {
|
||||||
|
todo!()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_status(&self) -> InterpretStatus {
|
||||||
|
todo!()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_children(&self) -> Vec<Id> {
|
||||||
|
todo!()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn on_discover_event(event: &DiscoveryEvent) {
|
||||||
|
info!("got discovery event {event:?}");
|
||||||
|
}
|
||||||
@@ -7,7 +7,6 @@ use serde::Serialize;
|
|||||||
use crate::{
|
use crate::{
|
||||||
config::HARMONY_DATA_DIR,
|
config::HARMONY_DATA_DIR,
|
||||||
data::{Id, Version},
|
data::{Id, Version},
|
||||||
instrumentation::{self, HarmonyEvent},
|
|
||||||
interpret::{Interpret, InterpretError, InterpretName, InterpretStatus, Outcome},
|
interpret::{Interpret, InterpretError, InterpretName, InterpretStatus, Outcome},
|
||||||
inventory::Inventory,
|
inventory::Inventory,
|
||||||
score::Score,
|
score::Score,
|
||||||
@@ -30,7 +29,7 @@ impl Default for K3DInstallationScore {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl<T: Topology> Score<T> for K3DInstallationScore {
|
impl<T: Topology> Score<T> for K3DInstallationScore {
|
||||||
fn create_interpret(&self) -> Box<dyn crate::interpret::Interpret<T>> {
|
fn create_interpret(&self) -> Box<dyn Interpret<T>> {
|
||||||
Box::new(K3dInstallationInterpret {
|
Box::new(K3dInstallationInterpret {
|
||||||
score: self.clone(),
|
score: self.clone(),
|
||||||
})
|
})
|
||||||
|
|||||||
@@ -5,6 +5,7 @@ pub mod dns;
|
|||||||
pub mod dummy;
|
pub mod dummy;
|
||||||
pub mod helm;
|
pub mod helm;
|
||||||
pub mod http;
|
pub mod http;
|
||||||
|
pub mod inventory;
|
||||||
pub mod ipxe;
|
pub mod ipxe;
|
||||||
pub mod k3d;
|
pub mod k3d;
|
||||||
pub mod k8s;
|
pub mod k8s;
|
||||||
@@ -14,5 +15,6 @@ pub mod monitoring;
|
|||||||
pub mod okd;
|
pub mod okd;
|
||||||
pub mod opnsense;
|
pub mod opnsense;
|
||||||
pub mod prometheus;
|
pub mod prometheus;
|
||||||
|
pub mod storage;
|
||||||
pub mod tenant;
|
pub mod tenant;
|
||||||
pub mod tftp;
|
pub mod tftp;
|
||||||
|
|||||||
@@ -33,7 +33,10 @@ impl<T: Topology + PrometheusApplicationMonitoring<CRDPrometheus>> Score<T>
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn name(&self) -> String {
|
fn name(&self) -> String {
|
||||||
"ApplicationMonitoringScore".to_string()
|
format!(
|
||||||
|
"{} monitoring [ApplicationMonitoringScore]",
|
||||||
|
self.application.name()
|
||||||
|
)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -61,7 +64,9 @@ impl<T: Topology + PrometheusApplicationMonitoring<CRDPrometheus>> Interpret<T>
|
|||||||
|
|
||||||
match result {
|
match result {
|
||||||
Ok(outcome) => match outcome {
|
Ok(outcome) => match outcome {
|
||||||
PreparationOutcome::Success { details } => Ok(Outcome::success(details)),
|
PreparationOutcome::Success { details: _ } => {
|
||||||
|
Ok(Outcome::success("Prometheus installed".into()))
|
||||||
|
}
|
||||||
PreparationOutcome::Noop => Ok(Outcome::noop()),
|
PreparationOutcome::Noop => Ok(Outcome::noop()),
|
||||||
},
|
},
|
||||||
Err(err) => Err(InterpretError::from(err)),
|
Err(err) => Err(InterpretError::from(err)),
|
||||||
|
|||||||
@@ -1,9 +1,25 @@
|
|||||||
use non_blank_string_rs::NonBlankString;
|
use non_blank_string_rs::NonBlankString;
|
||||||
use std::str::FromStr;
|
use std::str::FromStr;
|
||||||
|
|
||||||
use crate::modules::helm::chart::{HelmChartScore, HelmRepository};
|
use crate::{modules::helm::chart::HelmChartScore, topology::DeploymentTarget};
|
||||||
|
|
||||||
|
pub fn ntfy_helm_chart_score(
|
||||||
|
namespace: String,
|
||||||
|
host: String,
|
||||||
|
target: DeploymentTarget,
|
||||||
|
) -> HelmChartScore {
|
||||||
|
// TODO not actually the correct logic, this should be fixed by using an ingresss which is the
|
||||||
|
// correct k8s standard.
|
||||||
|
//
|
||||||
|
// Another option is to delegate to the topology the ingress technology it wants to use Route,
|
||||||
|
// Ingress or other
|
||||||
|
let route_enabled = match target {
|
||||||
|
DeploymentTarget::LocalDev => false,
|
||||||
|
DeploymentTarget::Staging => true,
|
||||||
|
DeploymentTarget::Production => true,
|
||||||
|
};
|
||||||
|
let ingress_enabled = !route_enabled;
|
||||||
|
|
||||||
pub fn ntfy_helm_chart_score(namespace: String, host: String) -> HelmChartScore {
|
|
||||||
let values = format!(
|
let values = format!(
|
||||||
r#"
|
r#"
|
||||||
replicaCount: 1
|
replicaCount: 1
|
||||||
@@ -25,23 +41,14 @@ serviceAccount:
|
|||||||
|
|
||||||
service:
|
service:
|
||||||
type: ClusterIP
|
type: ClusterIP
|
||||||
port: 80
|
port: 8080
|
||||||
|
|
||||||
ingress:
|
ingress:
|
||||||
enabled: true
|
enabled: {ingress_enabled}
|
||||||
# annotations:
|
|
||||||
# kubernetes.io/ingress.class: nginx
|
|
||||||
# kubernetes.io/tls-acme: "true"
|
|
||||||
hosts:
|
|
||||||
- host: {host}
|
|
||||||
paths:
|
|
||||||
- path: /
|
|
||||||
pathType: ImplementationSpecific
|
|
||||||
tls: []
|
|
||||||
# - secretName: chart-example-tls
|
|
||||||
# hosts:
|
|
||||||
# - chart-example.local
|
|
||||||
|
|
||||||
|
route:
|
||||||
|
enabled: {route_enabled}
|
||||||
|
host: {host}
|
||||||
|
|
||||||
autoscaling:
|
autoscaling:
|
||||||
enabled: false
|
enabled: false
|
||||||
@@ -49,7 +56,7 @@ autoscaling:
|
|||||||
config:
|
config:
|
||||||
enabled: true
|
enabled: true
|
||||||
data:
|
data:
|
||||||
# base-url: "https://ntfy.something.com"
|
base-url: "https://{host}"
|
||||||
auth-file: "/var/cache/ntfy/user.db"
|
auth-file: "/var/cache/ntfy/user.db"
|
||||||
auth-default-access: "deny-all"
|
auth-default-access: "deny-all"
|
||||||
cache-file: "/var/cache/ntfy/cache.db"
|
cache-file: "/var/cache/ntfy/cache.db"
|
||||||
@@ -59,6 +66,7 @@ config:
|
|||||||
enable-signup: false
|
enable-signup: false
|
||||||
enable-login: "true"
|
enable-login: "true"
|
||||||
enable-metrics: "true"
|
enable-metrics: "true"
|
||||||
|
listen-http: ":8080"
|
||||||
|
|
||||||
persistence:
|
persistence:
|
||||||
enabled: true
|
enabled: true
|
||||||
@@ -69,16 +77,12 @@ persistence:
|
|||||||
HelmChartScore {
|
HelmChartScore {
|
||||||
namespace: Some(NonBlankString::from_str(&namespace).unwrap()),
|
namespace: Some(NonBlankString::from_str(&namespace).unwrap()),
|
||||||
release_name: NonBlankString::from_str("ntfy").unwrap(),
|
release_name: NonBlankString::from_str("ntfy").unwrap(),
|
||||||
chart_name: NonBlankString::from_str("sarab97/ntfy").unwrap(),
|
chart_name: NonBlankString::from_str("oci://hub.nationtech.io/harmony/ntfy").unwrap(),
|
||||||
chart_version: Some(NonBlankString::from_str("0.1.7").unwrap()),
|
chart_version: Some(NonBlankString::from_str("0.1.7-nationtech.1").unwrap()),
|
||||||
values_overrides: None,
|
values_overrides: None,
|
||||||
values_yaml: Some(values.to_string()),
|
values_yaml: Some(values.to_string()),
|
||||||
create_namespace: true,
|
create_namespace: true,
|
||||||
install_only: false,
|
install_only: false,
|
||||||
repository: Some(HelmRepository::new(
|
repository: None,
|
||||||
"sarab97".to_string(),
|
|
||||||
url::Url::parse("https://charts.sarabsingh.com").unwrap(),
|
|
||||||
true,
|
|
||||||
)),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
|
|
||||||
use async_trait::async_trait;
|
use async_trait::async_trait;
|
||||||
use log::debug;
|
use log::info;
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
use strum::{Display, EnumString};
|
use strum::{Display, EnumString};
|
||||||
|
|
||||||
@@ -11,7 +11,7 @@ use crate::{
|
|||||||
inventory::Inventory,
|
inventory::Inventory,
|
||||||
modules::monitoring::ntfy::helm::ntfy_helm_chart::ntfy_helm_chart_score,
|
modules::monitoring::ntfy::helm::ntfy_helm_chart::ntfy_helm_chart_score,
|
||||||
score::Score,
|
score::Score,
|
||||||
topology::{HelmCommand, K8sclient, Topology, k8s::K8sClient},
|
topology::{HelmCommand, K8sclient, MultiTargetTopology, Topology, k8s::K8sClient},
|
||||||
};
|
};
|
||||||
|
|
||||||
#[derive(Debug, Clone, Serialize)]
|
#[derive(Debug, Clone, Serialize)]
|
||||||
@@ -20,7 +20,7 @@ pub struct NtfyScore {
|
|||||||
pub host: String,
|
pub host: String,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<T: Topology + HelmCommand + K8sclient> Score<T> for NtfyScore {
|
impl<T: Topology + HelmCommand + K8sclient + MultiTargetTopology> Score<T> for NtfyScore {
|
||||||
fn create_interpret(&self) -> Box<dyn Interpret<T>> {
|
fn create_interpret(&self) -> Box<dyn Interpret<T>> {
|
||||||
Box::new(NtfyInterpret {
|
Box::new(NtfyInterpret {
|
||||||
score: self.clone(),
|
score: self.clone(),
|
||||||
@@ -28,7 +28,7 @@ impl<T: Topology + HelmCommand + K8sclient> Score<T> for NtfyScore {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn name(&self) -> String {
|
fn name(&self) -> String {
|
||||||
"NtfyScore".to_string()
|
"alert receiver [NtfyScore]".into()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -37,18 +37,6 @@ pub struct NtfyInterpret {
|
|||||||
pub score: NtfyScore,
|
pub score: NtfyScore,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, EnumString, Display)]
|
|
||||||
enum NtfyAccessMode {
|
|
||||||
#[strum(serialize = "read-write", serialize = "rw")]
|
|
||||||
ReadWrite,
|
|
||||||
#[strum(serialize = "read-only", serialize = "ro", serialize = "read")]
|
|
||||||
ReadOnly,
|
|
||||||
#[strum(serialize = "write-only", serialize = "wo", serialize = "write")]
|
|
||||||
WriteOnly,
|
|
||||||
#[strum(serialize = "deny", serialize = "none")]
|
|
||||||
Deny,
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Debug, EnumString, Display)]
|
#[derive(Debug, EnumString, Display)]
|
||||||
enum NtfyRole {
|
enum NtfyRole {
|
||||||
#[strum(serialize = "user")]
|
#[strum(serialize = "user")]
|
||||||
@@ -77,7 +65,7 @@ impl NtfyInterpret {
|
|||||||
vec![
|
vec![
|
||||||
"sh",
|
"sh",
|
||||||
"-c",
|
"-c",
|
||||||
format!("NTFY_PASSWORD={password} ntfy user add --role={role} {username}")
|
format!("NTFY_PASSWORD={password} ntfy user add --role={role} --ignore-exists {username}")
|
||||||
.as_str(),
|
.as_str(),
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
@@ -89,22 +77,27 @@ impl NtfyInterpret {
|
|||||||
|
|
||||||
/// We need a ntfy interpret to wrap the HelmChartScore in order to run the score, and then bootstrap the config inside ntfy
|
/// We need a ntfy interpret to wrap the HelmChartScore in order to run the score, and then bootstrap the config inside ntfy
|
||||||
#[async_trait]
|
#[async_trait]
|
||||||
impl<T: Topology + HelmCommand + K8sclient> Interpret<T> for NtfyInterpret {
|
impl<T: Topology + HelmCommand + K8sclient + MultiTargetTopology> Interpret<T> for NtfyInterpret {
|
||||||
async fn execute(
|
async fn execute(
|
||||||
&self,
|
&self,
|
||||||
inventory: &Inventory,
|
inventory: &Inventory,
|
||||||
topology: &T,
|
topology: &T,
|
||||||
) -> Result<Outcome, InterpretError> {
|
) -> Result<Outcome, InterpretError> {
|
||||||
ntfy_helm_chart_score(self.score.namespace.clone(), self.score.host.clone())
|
ntfy_helm_chart_score(
|
||||||
|
self.score.namespace.clone(),
|
||||||
|
self.score.host.clone(),
|
||||||
|
topology.current_target(),
|
||||||
|
)
|
||||||
.interpret(inventory, topology)
|
.interpret(inventory, topology)
|
||||||
.await?;
|
.await?;
|
||||||
|
|
||||||
debug!("installed ntfy helm chart");
|
info!("installed ntfy helm chart");
|
||||||
let client = topology
|
let client = topology
|
||||||
.k8s_client()
|
.k8s_client()
|
||||||
.await
|
.await
|
||||||
.expect("couldn't get k8s client");
|
.expect("couldn't get k8s client");
|
||||||
|
|
||||||
|
info!("deploying ntfy...");
|
||||||
client
|
client
|
||||||
.wait_until_deployment_ready(
|
.wait_until_deployment_ready(
|
||||||
"ntfy".to_string(),
|
"ntfy".to_string(),
|
||||||
@@ -112,14 +105,14 @@ impl<T: Topology + HelmCommand + K8sclient> Interpret<T> for NtfyInterpret {
|
|||||||
None,
|
None,
|
||||||
)
|
)
|
||||||
.await?;
|
.await?;
|
||||||
debug!("created k8s client");
|
info!("ntfy deployed");
|
||||||
|
|
||||||
|
info!("adding user harmony");
|
||||||
self.add_user(client, "harmony", "harmony", Some(NtfyRole::Admin))
|
self.add_user(client, "harmony", "harmony", Some(NtfyRole::Admin))
|
||||||
.await?;
|
.await?;
|
||||||
|
info!("user added");
|
||||||
|
|
||||||
debug!("exec into pod done");
|
Ok(Outcome::success("Ntfy installed".to_string()))
|
||||||
|
|
||||||
Ok(Outcome::success("installed ntfy".to_string()))
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn get_name(&self) -> InterpretName {
|
fn get_name(&self) -> InterpretName {
|
||||||
|
|||||||
148
harmony/src/modules/okd/ipxe.rs
Normal file
148
harmony/src/modules/okd/ipxe.rs
Normal file
@@ -0,0 +1,148 @@
|
|||||||
|
use askama::Template;
|
||||||
|
use async_trait::async_trait;
|
||||||
|
use derive_new::new;
|
||||||
|
use serde::Serialize;
|
||||||
|
use std::net::IpAddr;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
data::{FileContent, FilePath, Id, Version},
|
||||||
|
interpret::{Interpret, InterpretError, InterpretName, InterpretStatus, Outcome},
|
||||||
|
inventory::Inventory,
|
||||||
|
modules::{dhcp::DhcpScore, http::StaticFilesHttpScore, tftp::TftpScore},
|
||||||
|
score::Score,
|
||||||
|
topology::{DhcpServer, HttpServer, Router, TftpServer, Topology, Url},
|
||||||
|
};
|
||||||
|
|
||||||
|
#[derive(Debug, new, Clone, Serialize)]
|
||||||
|
pub struct OkdIpxeScore {
|
||||||
|
pub kickstart_filename: String,
|
||||||
|
pub harmony_inventory_agent: String,
|
||||||
|
pub cluster_pubkey_filename: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T: Topology + DhcpServer + TftpServer + HttpServer + Router> Score<T> for OkdIpxeScore {
|
||||||
|
fn create_interpret(&self) -> Box<dyn Interpret<T>> {
|
||||||
|
Box::new(IpxeInterpret::new(self.clone()))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn name(&self) -> String {
|
||||||
|
"OkdIpxeScore".to_string()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, new, Clone)]
|
||||||
|
pub struct IpxeInterpret {
|
||||||
|
score: OkdIpxeScore,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[async_trait]
|
||||||
|
impl<T: Topology + DhcpServer + TftpServer + HttpServer + Router> Interpret<T> for IpxeInterpret {
|
||||||
|
async fn execute(
|
||||||
|
&self,
|
||||||
|
inventory: &Inventory,
|
||||||
|
topology: &T,
|
||||||
|
) -> Result<Outcome, InterpretError> {
|
||||||
|
let gateway_ip = topology.get_gateway();
|
||||||
|
|
||||||
|
let scores: Vec<Box<dyn Score<T>>> = vec![
|
||||||
|
Box::new(DhcpScore {
|
||||||
|
host_binding: vec![],
|
||||||
|
next_server: Some(topology.get_gateway()),
|
||||||
|
boot_filename: None,
|
||||||
|
filename: Some("undionly.kpxe".to_string()),
|
||||||
|
filename64: Some("ipxe.efi".to_string()),
|
||||||
|
filenameipxe: Some(format!("http://{gateway_ip}:8080/boot.ipxe").to_string()),
|
||||||
|
}),
|
||||||
|
Box::new(TftpScore {
|
||||||
|
files_to_serve: Url::LocalFolder("./data/pxe/okd/tftpboot/".to_string()),
|
||||||
|
}),
|
||||||
|
Box::new(StaticFilesHttpScore {
|
||||||
|
// TODO The current russh based copy is way too slow, check for a lib update or use scp
|
||||||
|
// when available
|
||||||
|
//
|
||||||
|
// For now just run :
|
||||||
|
// scp -r data/pxe/okd/http_files/* root@192.168.1.1:/usr/local/http/
|
||||||
|
//
|
||||||
|
folder_to_serve: None,
|
||||||
|
// folder_to_serve: Some(Url::LocalFolder("./data/pxe/okd/http_files/".to_string())),
|
||||||
|
files: vec![
|
||||||
|
FileContent {
|
||||||
|
path: FilePath::Relative("boot.ipxe".to_string()),
|
||||||
|
content: BootIpxeTpl {
|
||||||
|
gateway_ip: &gateway_ip,
|
||||||
|
}
|
||||||
|
.to_string(),
|
||||||
|
},
|
||||||
|
FileContent {
|
||||||
|
path: FilePath::Relative(self.score.kickstart_filename.clone()),
|
||||||
|
content: InventoryKickstartTpl {
|
||||||
|
gateway_ip: &gateway_ip,
|
||||||
|
harmony_inventory_agent: &self.score.harmony_inventory_agent,
|
||||||
|
cluster_pubkey_filename: &self.score.cluster_pubkey_filename,
|
||||||
|
}
|
||||||
|
.to_string(),
|
||||||
|
},
|
||||||
|
FileContent {
|
||||||
|
path: FilePath::Relative("fallback.ipxe".to_string()),
|
||||||
|
content: FallbackIpxeTpl {
|
||||||
|
gateway_ip: &gateway_ip,
|
||||||
|
kickstart_filename: &self.score.kickstart_filename,
|
||||||
|
}
|
||||||
|
.to_string(),
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}),
|
||||||
|
];
|
||||||
|
|
||||||
|
for score in scores {
|
||||||
|
let result = score.interpret(inventory, topology).await;
|
||||||
|
match result {
|
||||||
|
Ok(outcome) => match outcome.status {
|
||||||
|
InterpretStatus::SUCCESS => continue,
|
||||||
|
InterpretStatus::NOOP => continue,
|
||||||
|
_ => return Err(InterpretError::new(outcome.message)),
|
||||||
|
},
|
||||||
|
Err(e) => return Err(e),
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(Outcome::success("Ipxe installed".to_string()))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_name(&self) -> InterpretName {
|
||||||
|
InterpretName::Ipxe
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_version(&self) -> Version {
|
||||||
|
todo!()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_status(&self) -> InterpretStatus {
|
||||||
|
todo!()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_children(&self) -> Vec<Id> {
|
||||||
|
todo!()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Template)]
|
||||||
|
#[template(path = "boot.ipxe.j2")]
|
||||||
|
struct BootIpxeTpl<'a> {
|
||||||
|
gateway_ip: &'a IpAddr,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Template)]
|
||||||
|
#[template(path = "fallback.ipxe.j2")]
|
||||||
|
struct FallbackIpxeTpl<'a> {
|
||||||
|
gateway_ip: &'a IpAddr,
|
||||||
|
kickstart_filename: &'a str,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Template)]
|
||||||
|
#[template(path = "inventory.kickstart.j2")]
|
||||||
|
struct InventoryKickstartTpl<'a> {
|
||||||
|
gateway_ip: &'a IpAddr,
|
||||||
|
cluster_pubkey_filename: &'a str,
|
||||||
|
harmony_inventory_agent: &'a str,
|
||||||
|
}
|
||||||
@@ -2,5 +2,6 @@ pub mod bootstrap_dhcp;
|
|||||||
pub mod bootstrap_load_balancer;
|
pub mod bootstrap_load_balancer;
|
||||||
pub mod dhcp;
|
pub mod dhcp;
|
||||||
pub mod dns;
|
pub mod dns;
|
||||||
|
pub mod ipxe;
|
||||||
pub mod load_balancer;
|
pub mod load_balancer;
|
||||||
pub mod upgrade;
|
pub mod upgrade;
|
||||||
|
|||||||
@@ -61,7 +61,7 @@ impl<T: Topology + K8sclient + PrometheusApplicationMonitoring<CRDPrometheus>> S
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn name(&self) -> String {
|
fn name(&self) -> String {
|
||||||
"CRDApplicationAlertingScore".into()
|
"prometheus alerting [CRDAlertingScore]".into()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -94,7 +94,7 @@ impl<T: Topology + K8sclient + PrometheusApplicationMonitoring<CRDPrometheus>> I
|
|||||||
self.install_monitors(self.service_monitors.clone(), &client)
|
self.install_monitors(self.service_monitors.clone(), &client)
|
||||||
.await?;
|
.await?;
|
||||||
Ok(Outcome::success(
|
Ok(Outcome::success(
|
||||||
"deployed application monitoring composants".to_string(),
|
"K8s monitoring components installed".to_string(),
|
||||||
))
|
))
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -166,7 +166,8 @@ impl K8sPrometheusCRDAlertingInterpret {
|
|||||||
|
|
||||||
let install_output = Command::new("helm")
|
let install_output = Command::new("helm")
|
||||||
.args([
|
.args([
|
||||||
"install",
|
"upgrade",
|
||||||
|
"--install",
|
||||||
&chart_name,
|
&chart_name,
|
||||||
tgz_path.to_str().unwrap(),
|
tgz_path.to_str().unwrap(),
|
||||||
"--namespace",
|
"--namespace",
|
||||||
|
|||||||
418
harmony/src/modules/storage/ceph/ceph_osd_replacement_score.rs
Normal file
418
harmony/src/modules/storage/ceph/ceph_osd_replacement_score.rs
Normal file
@@ -0,0 +1,418 @@
|
|||||||
|
use std::{
|
||||||
|
sync::Arc,
|
||||||
|
time::{Duration, Instant},
|
||||||
|
};
|
||||||
|
|
||||||
|
use async_trait::async_trait;
|
||||||
|
use log::{info, warn};
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
use tokio::time::sleep;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
data::{Id, Version},
|
||||||
|
interpret::{Interpret, InterpretError, InterpretName, InterpretStatus, Outcome},
|
||||||
|
inventory::Inventory,
|
||||||
|
score::Score,
|
||||||
|
topology::{K8sclient, Topology, k8s::K8sClient},
|
||||||
|
};
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize)]
|
||||||
|
pub struct CephRemoveOsd {
|
||||||
|
osd_deployment_name: String,
|
||||||
|
rook_ceph_namespace: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T: Topology + K8sclient> Score<T> for CephRemoveOsd {
|
||||||
|
fn name(&self) -> String {
|
||||||
|
format!("CephRemoveOsdScore")
|
||||||
|
}
|
||||||
|
|
||||||
|
#[doc(hidden)]
|
||||||
|
fn create_interpret(&self) -> Box<dyn Interpret<T>> {
|
||||||
|
Box::new(CephRemoveOsdInterpret {
|
||||||
|
score: self.clone(),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone)]
|
||||||
|
pub struct CephRemoveOsdInterpret {
|
||||||
|
score: CephRemoveOsd,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[async_trait]
|
||||||
|
impl<T: Topology + K8sclient> Interpret<T> for CephRemoveOsdInterpret {
|
||||||
|
async fn execute(
|
||||||
|
&self,
|
||||||
|
_inventory: &Inventory,
|
||||||
|
topology: &T,
|
||||||
|
) -> Result<Outcome, InterpretError> {
|
||||||
|
let client = topology.k8s_client().await.unwrap();
|
||||||
|
self.verify_ceph_toolbox_exists(client.clone()).await?;
|
||||||
|
self.scale_deployment(client.clone()).await?;
|
||||||
|
self.verify_deployment_scaled(client.clone()).await?;
|
||||||
|
self.delete_deployment(client.clone()).await?;
|
||||||
|
self.verify_deployment_deleted(client.clone()).await?;
|
||||||
|
let osd_id_full = self.get_ceph_osd_id().unwrap();
|
||||||
|
self.purge_ceph_osd(client.clone(), &osd_id_full).await?;
|
||||||
|
self.verify_ceph_osd_removal(client.clone(), &osd_id_full)
|
||||||
|
.await?;
|
||||||
|
|
||||||
|
Ok(Outcome::success(format!(
|
||||||
|
"Successfully removed OSD {} from rook-ceph cluster by deleting deployment {}",
|
||||||
|
osd_id_full, self.score.osd_deployment_name
|
||||||
|
)))
|
||||||
|
}
|
||||||
|
fn get_name(&self) -> InterpretName {
|
||||||
|
todo!()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_version(&self) -> Version {
|
||||||
|
todo!()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_status(&self) -> InterpretStatus {
|
||||||
|
todo!()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_children(&self) -> Vec<Id> {
|
||||||
|
todo!()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl CephRemoveOsdInterpret {
|
||||||
|
pub fn get_ceph_osd_id(&self) -> Result<String, InterpretError> {
|
||||||
|
let osd_id_numeric = self
|
||||||
|
.score
|
||||||
|
.osd_deployment_name
|
||||||
|
.split('-')
|
||||||
|
.nth(3)
|
||||||
|
.ok_or_else(|| {
|
||||||
|
InterpretError::new(format!(
|
||||||
|
"Could not parse OSD id from deployment name {}",
|
||||||
|
self.score.osd_deployment_name
|
||||||
|
))
|
||||||
|
})?;
|
||||||
|
let osd_id_full = format!("osd.{}", osd_id_numeric);
|
||||||
|
|
||||||
|
info!(
|
||||||
|
"Targeting Ceph OSD: {} (parsed from deployment {})",
|
||||||
|
osd_id_full, self.score.osd_deployment_name
|
||||||
|
);
|
||||||
|
|
||||||
|
Ok(osd_id_full)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn verify_ceph_toolbox_exists(
|
||||||
|
&self,
|
||||||
|
client: Arc<K8sClient>,
|
||||||
|
) -> Result<Outcome, InterpretError> {
|
||||||
|
let toolbox_dep = "rook-ceph-tools".to_string();
|
||||||
|
|
||||||
|
match client
|
||||||
|
.get_deployment(&toolbox_dep, Some(&self.score.rook_ceph_namespace))
|
||||||
|
.await
|
||||||
|
{
|
||||||
|
Ok(Some(deployment)) => {
|
||||||
|
if let Some(status) = deployment.status {
|
||||||
|
let ready_count = status.ready_replicas.unwrap_or(0);
|
||||||
|
if ready_count >= 1 {
|
||||||
|
return Ok(Outcome::success(format!(
|
||||||
|
"'{}' is ready with {} replica(s).",
|
||||||
|
&toolbox_dep, ready_count
|
||||||
|
)));
|
||||||
|
} else {
|
||||||
|
return Err(InterpretError::new(
|
||||||
|
"ceph-tool-box not ready in cluster".to_string(),
|
||||||
|
));
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
Err(InterpretError::new(format!(
|
||||||
|
"failed to get deployment status {}",
|
||||||
|
&toolbox_dep
|
||||||
|
)))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Ok(None) => Err(InterpretError::new(format!(
|
||||||
|
"Deployment '{}' not found in namespace '{}'.",
|
||||||
|
&toolbox_dep, self.score.rook_ceph_namespace
|
||||||
|
))),
|
||||||
|
Err(e) => Err(InterpretError::new(format!(
|
||||||
|
"Failed to query for deployment '{}': {}",
|
||||||
|
&toolbox_dep, e
|
||||||
|
))),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn scale_deployment(
|
||||||
|
&self,
|
||||||
|
client: Arc<K8sClient>,
|
||||||
|
) -> Result<Outcome, InterpretError> {
|
||||||
|
info!(
|
||||||
|
"Scaling down OSD deployment: {}",
|
||||||
|
self.score.osd_deployment_name
|
||||||
|
);
|
||||||
|
client
|
||||||
|
.scale_deployment(
|
||||||
|
&self.score.osd_deployment_name,
|
||||||
|
Some(&self.score.rook_ceph_namespace),
|
||||||
|
0,
|
||||||
|
)
|
||||||
|
.await?;
|
||||||
|
Ok(Outcome::success(format!(
|
||||||
|
"Scaled down deployment {}",
|
||||||
|
self.score.osd_deployment_name
|
||||||
|
)))
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn verify_deployment_scaled(
|
||||||
|
&self,
|
||||||
|
client: Arc<K8sClient>,
|
||||||
|
) -> Result<Outcome, InterpretError> {
|
||||||
|
let (timeout, interval, start) = self.build_timer();
|
||||||
|
|
||||||
|
info!("Waiting for OSD deployment to scale down to 0 replicas");
|
||||||
|
loop {
|
||||||
|
let dep = client
|
||||||
|
.get_deployment(
|
||||||
|
&self.score.osd_deployment_name,
|
||||||
|
Some(&self.score.rook_ceph_namespace),
|
||||||
|
)
|
||||||
|
.await?;
|
||||||
|
|
||||||
|
if let Some(deployment) = dep {
|
||||||
|
if let Some(status) = deployment.status {
|
||||||
|
if status.replicas.unwrap_or(1) == 0 && status.ready_replicas.unwrap_or(1) == 0
|
||||||
|
{
|
||||||
|
return Ok(Outcome::success(
|
||||||
|
"Deployment successfully scaled down.".to_string(),
|
||||||
|
));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if start.elapsed() > timeout {
|
||||||
|
return Err(InterpretError::new(format!(
|
||||||
|
"Timed out waiting for deployment {} to scale down",
|
||||||
|
self.score.osd_deployment_name
|
||||||
|
)));
|
||||||
|
}
|
||||||
|
sleep(interval).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn build_timer(&self) -> (Duration, Duration, Instant) {
|
||||||
|
let timeout = Duration::from_secs(120);
|
||||||
|
let interval = Duration::from_secs(5);
|
||||||
|
let start = Instant::now();
|
||||||
|
(timeout, interval, start)
|
||||||
|
}
|
||||||
|
pub async fn delete_deployment(
|
||||||
|
&self,
|
||||||
|
client: Arc<K8sClient>,
|
||||||
|
) -> Result<Outcome, InterpretError> {
|
||||||
|
info!(
|
||||||
|
"Deleting OSD deployment: {}",
|
||||||
|
self.score.osd_deployment_name
|
||||||
|
);
|
||||||
|
client
|
||||||
|
.delete_deployment(
|
||||||
|
&self.score.osd_deployment_name,
|
||||||
|
Some(&self.score.rook_ceph_namespace),
|
||||||
|
)
|
||||||
|
.await?;
|
||||||
|
Ok(Outcome::success(format!(
|
||||||
|
"deployment {} deleted",
|
||||||
|
self.score.osd_deployment_name
|
||||||
|
)))
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn verify_deployment_deleted(
|
||||||
|
&self,
|
||||||
|
client: Arc<K8sClient>,
|
||||||
|
) -> Result<Outcome, InterpretError> {
|
||||||
|
let (timeout, interval, start) = self.build_timer();
|
||||||
|
|
||||||
|
info!("Waiting for OSD deployment to scale down to 0 replicas");
|
||||||
|
loop {
|
||||||
|
let dep = client
|
||||||
|
.get_deployment(
|
||||||
|
&self.score.osd_deployment_name,
|
||||||
|
Some(&self.score.rook_ceph_namespace),
|
||||||
|
)
|
||||||
|
.await?;
|
||||||
|
|
||||||
|
if dep.is_none() {
|
||||||
|
info!(
|
||||||
|
"Deployment {} successfully deleted.",
|
||||||
|
self.score.osd_deployment_name
|
||||||
|
);
|
||||||
|
return Ok(Outcome::success(format!(
|
||||||
|
"Deployment {} deleted.",
|
||||||
|
self.score.osd_deployment_name
|
||||||
|
)));
|
||||||
|
}
|
||||||
|
|
||||||
|
if start.elapsed() > timeout {
|
||||||
|
return Err(InterpretError::new(format!(
|
||||||
|
"Timed out waiting for deployment {} to be deleted",
|
||||||
|
self.score.osd_deployment_name
|
||||||
|
)));
|
||||||
|
}
|
||||||
|
sleep(interval).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_osd_tree(&self, json: serde_json::Value) -> Result<CephOsdTree, InterpretError> {
|
||||||
|
let nodes = json.get("nodes").ok_or_else(|| {
|
||||||
|
InterpretError::new("Missing 'nodes' field in ceph osd tree JSON".to_string())
|
||||||
|
})?;
|
||||||
|
let tree: CephOsdTree = CephOsdTree {
|
||||||
|
nodes: serde_json::from_value(nodes.clone()).map_err(|e| {
|
||||||
|
InterpretError::new(format!("Failed to parse ceph osd tree JSON: {}", e))
|
||||||
|
})?,
|
||||||
|
};
|
||||||
|
Ok(tree)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn purge_ceph_osd(
|
||||||
|
&self,
|
||||||
|
client: Arc<K8sClient>,
|
||||||
|
osd_id_full: &str,
|
||||||
|
) -> Result<Outcome, InterpretError> {
|
||||||
|
info!(
|
||||||
|
"Purging OSD {} from Ceph cluster and removing its auth key",
|
||||||
|
osd_id_full
|
||||||
|
);
|
||||||
|
client
|
||||||
|
.exec_app_capture_output(
|
||||||
|
"rook-ceph-tools".to_string(),
|
||||||
|
"app".to_string(),
|
||||||
|
Some(&self.score.rook_ceph_namespace),
|
||||||
|
vec![
|
||||||
|
format!("ceph osd purge {osd_id_full} --yes-i-really-mean-it").as_str(),
|
||||||
|
format!("ceph auth del osd.{osd_id_full}").as_str(),
|
||||||
|
],
|
||||||
|
)
|
||||||
|
.await?;
|
||||||
|
Ok(Outcome::success(format!(
|
||||||
|
"osd id {} removed from osd tree",
|
||||||
|
osd_id_full
|
||||||
|
)))
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn verify_ceph_osd_removal(
|
||||||
|
&self,
|
||||||
|
client: Arc<K8sClient>,
|
||||||
|
osd_id_full: &str,
|
||||||
|
) -> Result<Outcome, InterpretError> {
|
||||||
|
let (timeout, interval, start) = self.build_timer();
|
||||||
|
info!(
|
||||||
|
"Verifying OSD {} has been removed from the Ceph tree...",
|
||||||
|
osd_id_full
|
||||||
|
);
|
||||||
|
loop {
|
||||||
|
let output = client
|
||||||
|
.exec_app_capture_output(
|
||||||
|
"rook-ceph-tools".to_string(),
|
||||||
|
"app".to_string(),
|
||||||
|
Some(&self.score.rook_ceph_namespace),
|
||||||
|
vec!["ceph osd tree -f json"],
|
||||||
|
)
|
||||||
|
.await?;
|
||||||
|
let tree =
|
||||||
|
self.get_osd_tree(serde_json::from_str(&output).expect("could not extract json"));
|
||||||
|
|
||||||
|
let osd_found = tree
|
||||||
|
.unwrap()
|
||||||
|
.nodes
|
||||||
|
.iter()
|
||||||
|
.any(|node| node.name == osd_id_full);
|
||||||
|
|
||||||
|
if !osd_found {
|
||||||
|
return Ok(Outcome::success(format!(
|
||||||
|
"Successfully verified that OSD {} is removed from the Ceph cluster.",
|
||||||
|
osd_id_full,
|
||||||
|
)));
|
||||||
|
}
|
||||||
|
|
||||||
|
if start.elapsed() > timeout {
|
||||||
|
return Err(InterpretError::new(format!(
|
||||||
|
"Timed out waiting for OSD {} to be removed from Ceph tree",
|
||||||
|
osd_id_full
|
||||||
|
)));
|
||||||
|
}
|
||||||
|
|
||||||
|
warn!(
|
||||||
|
"OSD {} still found in Ceph tree, retrying in {:?}...",
|
||||||
|
osd_id_full, interval
|
||||||
|
);
|
||||||
|
sleep(interval).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#[derive(Debug, Deserialize, PartialEq)]
|
||||||
|
pub struct CephOsdTree {
|
||||||
|
pub nodes: Vec<CephNode>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Deserialize, PartialEq)]
|
||||||
|
pub struct CephNode {
|
||||||
|
pub id: i32,
|
||||||
|
pub name: String,
|
||||||
|
#[serde(rename = "type")]
|
||||||
|
pub node_type: String,
|
||||||
|
pub type_id: Option<i32>,
|
||||||
|
pub children: Option<Vec<i32>>,
|
||||||
|
pub exists: Option<i32>,
|
||||||
|
pub status: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use serde_json::json;
|
||||||
|
|
||||||
|
use super::*;
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_get_osd_tree() {
|
||||||
|
let json_data = json!({
|
||||||
|
"nodes": [
|
||||||
|
{"id": 1, "name": "osd.1", "type": "osd", "primary_affinity":"1"},
|
||||||
|
{"id": 2, "name": "osd.2", "type": "osd", "crush_weight": 1.22344}
|
||||||
|
]
|
||||||
|
});
|
||||||
|
let interpret = CephRemoveOsdInterpret {
|
||||||
|
score: CephRemoveOsd {
|
||||||
|
osd_deployment_name: "osd-1".to_string(),
|
||||||
|
rook_ceph_namespace: "dummy_ns".to_string(),
|
||||||
|
},
|
||||||
|
};
|
||||||
|
let json = interpret.get_osd_tree(json_data).unwrap();
|
||||||
|
|
||||||
|
let expected = CephOsdTree {
|
||||||
|
nodes: vec![
|
||||||
|
CephNode {
|
||||||
|
id: 1,
|
||||||
|
name: "osd.1".to_string(),
|
||||||
|
node_type: "osd".to_string(),
|
||||||
|
type_id: None,
|
||||||
|
children: None,
|
||||||
|
exists: None,
|
||||||
|
status: None,
|
||||||
|
},
|
||||||
|
CephNode {
|
||||||
|
id: 2,
|
||||||
|
name: "osd.2".to_string(),
|
||||||
|
node_type: "osd".to_string(),
|
||||||
|
type_id: None,
|
||||||
|
children: None,
|
||||||
|
exists: None,
|
||||||
|
status: None,
|
||||||
|
},
|
||||||
|
],
|
||||||
|
};
|
||||||
|
|
||||||
|
assert_eq!(json, expected);
|
||||||
|
}
|
||||||
|
}
|
||||||
1
harmony/src/modules/storage/ceph/mod.rs
Normal file
1
harmony/src/modules/storage/ceph/mod.rs
Normal file
@@ -0,0 +1 @@
|
|||||||
|
pub mod ceph_osd_replacement_score;
|
||||||
1
harmony/src/modules/storage/mod.rs
Normal file
1
harmony/src/modules/storage/mod.rs
Normal file
@@ -0,0 +1 @@
|
|||||||
|
pub mod ceph;
|
||||||
@@ -28,7 +28,7 @@ impl<T: Topology + TenantManager> Score<T> for TenantScore {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn name(&self) -> String {
|
fn name(&self) -> String {
|
||||||
format!("{} TenantScore", self.config.name)
|
format!("{} [TenantScore]", self.config.name)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -47,8 +47,8 @@ impl<T: Topology + TenantManager> Interpret<T> for TenantInterpret {
|
|||||||
topology.provision_tenant(&self.tenant_config).await?;
|
topology.provision_tenant(&self.tenant_config).await?;
|
||||||
|
|
||||||
Ok(Outcome::success(format!(
|
Ok(Outcome::success(format!(
|
||||||
"Successfully provisioned tenant {} with id {}",
|
"Tenant provisioned with id '{}'",
|
||||||
self.tenant_config.name, self.tenant_config.id
|
self.tenant_config.id
|
||||||
)))
|
)))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ use crate::{
|
|||||||
|
|
||||||
#[derive(Debug, new, Clone, Serialize)]
|
#[derive(Debug, new, Clone, Serialize)]
|
||||||
pub struct TftpScore {
|
pub struct TftpScore {
|
||||||
files_to_serve: Url,
|
pub files_to_serve: Url,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<T: Topology + TftpServer + Router> Score<T> for TftpScore {
|
impl<T: Topology + TftpServer + Router> Score<T> for TftpScore {
|
||||||
|
|||||||
6
harmony/templates/boot.ipxe.j2
Normal file
6
harmony/templates/boot.ipxe.j2
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
#!ipxe
|
||||||
|
|
||||||
|
set base-url http://{{ gateway_ip }}:8080
|
||||||
|
set hostfile ${base-url}/byMAC/01-${mac:hexhyp}.ipxe
|
||||||
|
|
||||||
|
chain ${hostfile} || chain ${base-url}/fallback.ipxe
|
||||||
40
harmony/templates/fallback.ipxe.j2
Normal file
40
harmony/templates/fallback.ipxe.j2
Normal file
@@ -0,0 +1,40 @@
|
|||||||
|
#!ipxe
|
||||||
|
|
||||||
|
# =================================================================
|
||||||
|
# Harmony Discovery Agent - Default Boot Script (default.ipxe)
|
||||||
|
# =================================================================
|
||||||
|
#
|
||||||
|
# This script boots the CentOS Stream live environment for the
|
||||||
|
# purpose of hardware inventory. It loads the kernel and initramfs
|
||||||
|
# directly and passes a Kickstart URL for full automation.
|
||||||
|
#
|
||||||
|
|
||||||
|
# --- Configuration
|
||||||
|
# Set the base URL for where the CentOS kernel/initrd are hosted.
|
||||||
|
set os_base_url http://{{gateway_ip}}:8080/os/centos-stream-9
|
||||||
|
# Set the URL for the Kickstart file.
|
||||||
|
set ks_url http://{{ gateway_ip }}:8080/{{ kickstart_filename }}
|
||||||
|
|
||||||
|
# --- Boot Process
|
||||||
|
echo "Harmony: Starting automated node discovery..."
|
||||||
|
echo "Fetching kernel from ${os_base_url}/vmlinuz..."
|
||||||
|
kernel ${os_base_url}/vmlinuz
|
||||||
|
|
||||||
|
echo "Fetching initramfs from ${os_base_url}/initrd.img..."
|
||||||
|
initrd ${os_base_url}/initrd.img
|
||||||
|
|
||||||
|
echo "Configuring kernel boot arguments..."
|
||||||
|
# Kernel Arguments Explained:
|
||||||
|
# - initrd=initrd.img: Specifies the initial ramdisk to use.
|
||||||
|
# - inst.stage2: Points to the OS source. For a live boot, the base URL is sufficient.
|
||||||
|
# - inst.ks: CRITICAL: Points to our Kickstart file for automation.
|
||||||
|
# - ip=dhcp: Ensures the live environment configures its network.
|
||||||
|
# - console=...: Provides boot output on both serial and graphical consoles for debugging.
|
||||||
|
imgargs vmlinuz initrd=initrd.img inst.sshd inst.stage2=${os_base_url} inst.ks=${ks_url} ip=dhcp console=ttyS0,115200 console=tty1
|
||||||
|
|
||||||
|
echo "Booting into CentOS Stream 9 live environment..."
|
||||||
|
boot || goto failed
|
||||||
|
|
||||||
|
:failed
|
||||||
|
echo "Boot failed. Dropping to iPXE shell."
|
||||||
|
shell
|
||||||
127
harmony/templates/inventory.kickstart.j2
Normal file
127
harmony/templates/inventory.kickstart.j2
Normal file
@@ -0,0 +1,127 @@
|
|||||||
|
# --- Pre-Boot Scripting (The Main Goal) ---
|
||||||
|
# This section runs after the live environment has booted into RAM.
|
||||||
|
# It sets up SSH and downloads/runs the harmony-inventory-agent.
|
||||||
|
%pre --log=/root/ks-pre.log
|
||||||
|
|
||||||
|
echo "Harmony Kickstart: Pre-boot script started."
|
||||||
|
|
||||||
|
# 1. Configure SSH Access for Root
|
||||||
|
# Create the .ssh directory and set correct permissions.
|
||||||
|
echo " - Setting up SSH authorized_keys for root..."
|
||||||
|
mkdir -p /root/.ssh
|
||||||
|
chmod 700 /root/.ssh
|
||||||
|
|
||||||
|
# Download the public key from the provisioning server.
|
||||||
|
# The -sS flags make curl silent but show errors. -L follows redirects.
|
||||||
|
curl -vSL "http://{{ gateway_ip }}:8080/{{ cluster_pubkey_filename }}" -o /root/.ssh/authorized_keys
|
||||||
|
if [ $? -ne 0 ]; then
|
||||||
|
echo " - ERROR: Failed to download SSH public key."
|
||||||
|
else
|
||||||
|
echo " - SSH key downloaded successfully."
|
||||||
|
chmod 600 /root/.ssh/authorized_keys
|
||||||
|
fi
|
||||||
|
|
||||||
|
# 2. Download the Harmony Inventory Agent
|
||||||
|
echo " - Downloading harmony-inventory-agent..."
|
||||||
|
curl -vSL "http://{{ gateway_ip }}:8080/{{ harmony_inventory_agent }}" -o /usr/bin/harmony-inventory-agent
|
||||||
|
if [ $? -ne 0 ]; then
|
||||||
|
echo " - ERROR: Failed to download harmony_inventory_agent."
|
||||||
|
else
|
||||||
|
echo " - Agent binary downloaded successfully."
|
||||||
|
chmod +x /usr/bin/harmony-inventory-agent
|
||||||
|
fi
|
||||||
|
|
||||||
|
# 3. Create a systemd service to run the agent persistently.
|
||||||
|
# This is the most robust method to ensure the agent stays running.
|
||||||
|
echo " - Creating systemd service for the agent..."
|
||||||
|
cat > /etc/systemd/system/harmony-agent.service << EOF
|
||||||
|
[Unit]
|
||||||
|
Description=Harmony Inventory Agent
|
||||||
|
After=network-online.target
|
||||||
|
Wants=network-online.target
|
||||||
|
|
||||||
|
[Service]
|
||||||
|
Type=simple
|
||||||
|
ExecStart=/usr/bin/harmony-inventory-agent
|
||||||
|
Restart=on-failure
|
||||||
|
RestartSec=5
|
||||||
|
|
||||||
|
[Install]
|
||||||
|
WantedBy=multi-user.target
|
||||||
|
EOF
|
||||||
|
|
||||||
|
# 4. Enable and start the service
|
||||||
|
# The 'systemctl' commands will work correctly within the chroot environment of the %pre script.
|
||||||
|
echo " - Enabling and starting harmony-agent.service..."
|
||||||
|
systemctl daemon-reload
|
||||||
|
systemctl enable --now harmony-agent.service
|
||||||
|
|
||||||
|
# Check if the service started correctly
|
||||||
|
systemctl is-active --quiet harmony-agent.service
|
||||||
|
if [ $? -eq 0 ]; then
|
||||||
|
echo " - Harmony Inventory Agent service is now running."
|
||||||
|
else
|
||||||
|
echo " - ERROR: Harmony Inventory Agent service failed to start."
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "Harmony Kickstart: Pre-boot script finished. The machine is ready for inventory."
|
||||||
|
|
||||||
|
echo "Running cat - to pause system indefinitely"
|
||||||
|
cat -
|
||||||
|
|
||||||
|
%end
|
||||||
|
|
||||||
|
# =================================================================
|
||||||
|
# Harmony Discovery Agent - Kickstart File (NON-INSTALL, LIVE BOOT)
|
||||||
|
# =================================================================
|
||||||
|
#
|
||||||
|
# This file achieves a fully automated, non-interactive boot into a
|
||||||
|
# live CentOS environment. It does NOT install to disk.
|
||||||
|
#
|
||||||
|
|
||||||
|
# --- Automation and Interaction Control ---
|
||||||
|
# Perform the installation in command-line mode. This is critical for
|
||||||
|
# preventing Anaconda from starting a UI and halting for input.
|
||||||
|
cmdline
|
||||||
|
|
||||||
|
# Accept the End User License Agreement to prevent a prompt.
|
||||||
|
eula --agreed
|
||||||
|
|
||||||
|
# --- Core System Configuration (Required by Anaconda) ---
|
||||||
|
# Set keyboard and language. These are mandatory.
|
||||||
|
keyboard --vckeymap=us --xlayouts='us'
|
||||||
|
lang en_US.UTF-8
|
||||||
|
|
||||||
|
# Configure networking. This is essential for the %post script to work.
|
||||||
|
# The --activate flag ensures this device is brought up in the installer environment.
|
||||||
|
network --bootproto=dhcp --device=link --activate
|
||||||
|
|
||||||
|
# Set a locked root password. This is a mandatory command.
|
||||||
|
rootpw --lock
|
||||||
|
|
||||||
|
# Set the timezone. This is a mandatory command.
|
||||||
|
timezone UTC
|
||||||
|
|
||||||
|
# --- Disable Installation-Specific Features ---
|
||||||
|
# CRITICAL: Do not install a bootloader. The --disabled flag prevents
|
||||||
|
# this step and avoids errors about where to install it.
|
||||||
|
bootloader --disabled
|
||||||
|
|
||||||
|
# CRITICAL: Ignore all disks. This prevents Anaconda from stopping at the
|
||||||
|
# "Installation Destination" screen asking where to install.
|
||||||
|
# ignoredisk --drives /dev/sda
|
||||||
|
|
||||||
|
# Do not run the Initial Setup wizard on first boot.
|
||||||
|
firstboot --disable
|
||||||
|
|
||||||
|
# --- Package Selection ---
|
||||||
|
# We are not installing, so this section can be minimal.
|
||||||
|
# An empty %packages section is valid and ensures no time is wasted
|
||||||
|
# resolving dependencies for an installation that will not happen.
|
||||||
|
%packages
|
||||||
|
%end
|
||||||
|
|
||||||
|
|
||||||
|
# IMPORTANT: Do not include a final action command like 'reboot' or 'poweroff'.
|
||||||
|
# The default action is 'halt', which in cmdline mode will leave the system
|
||||||
|
# running in the live environment with the agent active, which is the desired state.
|
||||||
@@ -5,6 +5,10 @@ version.workspace = true
|
|||||||
readme.workspace = true
|
readme.workspace = true
|
||||||
license.workspace = true
|
license.workspace = true
|
||||||
|
|
||||||
|
[features]
|
||||||
|
default = ["tui"]
|
||||||
|
tui = ["dep:harmony_tui"]
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
assert_cmd = "2.0.17"
|
assert_cmd = "2.0.17"
|
||||||
clap = { version = "4.5.35", features = ["derive"] }
|
clap = { version = "4.5.35", features = ["derive"] }
|
||||||
@@ -18,8 +22,7 @@ indicatif = "0.18.0"
|
|||||||
lazy_static = "1.5.0"
|
lazy_static = "1.5.0"
|
||||||
log.workspace = true
|
log.workspace = true
|
||||||
indicatif-log-bridge = "0.2.3"
|
indicatif-log-bridge = "0.2.3"
|
||||||
|
chrono.workspace = true
|
||||||
|
|
||||||
|
[dev-dependencies]
|
||||||
[features]
|
harmony = { path = "../harmony", features = ["testing"] }
|
||||||
default = ["tui"]
|
|
||||||
tui = ["dep:harmony_tui"]
|
|
||||||
|
|||||||
@@ -1,187 +1,191 @@
|
|||||||
|
use chrono::Local;
|
||||||
|
use console::style;
|
||||||
use harmony::{
|
use harmony::{
|
||||||
instrumentation::{self, HarmonyEvent},
|
instrumentation::{self, HarmonyEvent},
|
||||||
|
modules::application::ApplicationFeatureStatus,
|
||||||
topology::TopologyStatus,
|
topology::TopologyStatus,
|
||||||
};
|
};
|
||||||
use indicatif::{MultiProgress, ProgressBar};
|
use log::{error, info, log_enabled};
|
||||||
use indicatif_log_bridge::LogWrapper;
|
use std::io::Write;
|
||||||
use std::{
|
use std::sync::Mutex;
|
||||||
collections::HashMap,
|
|
||||||
sync::{Arc, Mutex},
|
|
||||||
};
|
|
||||||
|
|
||||||
use crate::progress;
|
pub fn init() {
|
||||||
|
|
||||||
pub fn init() -> tokio::task::JoinHandle<()> {
|
|
||||||
configure_logger();
|
configure_logger();
|
||||||
let handle = tokio::spawn(handle_events());
|
handle_events();
|
||||||
|
|
||||||
loop {
|
|
||||||
if instrumentation::instrument(HarmonyEvent::HarmonyStarted).is_ok() {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
handle
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn configure_logger() {
|
fn configure_logger() {
|
||||||
let logger =
|
env_logger::Builder::from_env(env_logger::Env::default().default_filter_or("info"))
|
||||||
env_logger::Builder::from_env(env_logger::Env::default().default_filter_or("info")).build();
|
.format(|buf, record| {
|
||||||
let level = logger.filter();
|
let debug_mode = log_enabled!(log::Level::Debug);
|
||||||
let multi = MultiProgress::new();
|
let timestamp = Local::now().format("%Y-%m-%d %H:%M:%S");
|
||||||
LogWrapper::new(multi.clone(), logger).try_init().unwrap();
|
|
||||||
log::set_max_level(level);
|
let level = match record.level() {
|
||||||
|
log::Level::Error => style("ERROR").red(),
|
||||||
|
log::Level::Warn => style("WARN").yellow(),
|
||||||
|
log::Level::Info => style("INFO").green(),
|
||||||
|
log::Level::Debug => style("DEBUG").blue(),
|
||||||
|
log::Level::Trace => style("TRACE").magenta(),
|
||||||
|
};
|
||||||
|
if let Some(status) = record.key_values().get(log::kv::Key::from("status")) {
|
||||||
|
let status = status.to_borrowed_str().unwrap();
|
||||||
|
let emoji = match status {
|
||||||
|
"finished" => style(crate::theme::EMOJI_SUCCESS.to_string()).green(),
|
||||||
|
"skipped" => style(crate::theme::EMOJI_SKIP.to_string()).yellow(),
|
||||||
|
"failed" => style(crate::theme::EMOJI_ERROR.to_string()).red(),
|
||||||
|
_ => style("".into()),
|
||||||
|
};
|
||||||
|
if debug_mode {
|
||||||
|
writeln!(
|
||||||
|
buf,
|
||||||
|
"[{} {:<5} {}] {} {}",
|
||||||
|
timestamp,
|
||||||
|
level,
|
||||||
|
record.target(),
|
||||||
|
emoji,
|
||||||
|
record.args()
|
||||||
|
)
|
||||||
|
} else {
|
||||||
|
writeln!(buf, "[{:<5}] {} {}", level, emoji, record.args())
|
||||||
|
}
|
||||||
|
} else if let Some(emoji) = record.key_values().get(log::kv::Key::from("emoji")) {
|
||||||
|
if debug_mode {
|
||||||
|
writeln!(
|
||||||
|
buf,
|
||||||
|
"[{} {:<5} {}] {} {}",
|
||||||
|
timestamp,
|
||||||
|
level,
|
||||||
|
record.target(),
|
||||||
|
emoji,
|
||||||
|
record.args()
|
||||||
|
)
|
||||||
|
} else {
|
||||||
|
writeln!(buf, "[{:<5}] {} {}", level, emoji, record.args())
|
||||||
|
}
|
||||||
|
} else if debug_mode {
|
||||||
|
writeln!(
|
||||||
|
buf,
|
||||||
|
"[{} {:<5} {}] {}",
|
||||||
|
timestamp,
|
||||||
|
level,
|
||||||
|
record.target(),
|
||||||
|
record.args()
|
||||||
|
)
|
||||||
|
} else {
|
||||||
|
writeln!(buf, "[{:<5}] {}", level, record.args())
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.init();
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn handle_events() {
|
fn handle_events() {
|
||||||
|
let preparing_topology = Mutex::new(false);
|
||||||
|
let current_score: Mutex<Option<String>> = Mutex::new(None);
|
||||||
|
|
||||||
instrumentation::subscribe("Harmony CLI Logger", {
|
instrumentation::subscribe("Harmony CLI Logger", {
|
||||||
let sections: Arc<Mutex<HashMap<String, MultiProgress>>> =
|
|
||||||
Arc::new(Mutex::new(HashMap::new()));
|
|
||||||
let progress_bars: Arc<Mutex<HashMap<String, ProgressBar>>> =
|
|
||||||
Arc::new(Mutex::new(HashMap::new()));
|
|
||||||
|
|
||||||
move |event| {
|
move |event| {
|
||||||
let sections_clone = Arc::clone(§ions);
|
let mut preparing_topology = preparing_topology.lock().unwrap();
|
||||||
let progress_bars_clone = Arc::clone(&progress_bars);
|
let mut current_score = current_score.lock().unwrap();
|
||||||
|
|
||||||
async move {
|
|
||||||
let mut sections = sections_clone.lock().unwrap();
|
|
||||||
let mut progress_bars = progress_bars_clone.lock().unwrap();
|
|
||||||
|
|
||||||
match event {
|
match event {
|
||||||
HarmonyEvent::HarmonyStarted => {}
|
HarmonyEvent::HarmonyStarted => {}
|
||||||
|
HarmonyEvent::HarmonyFinished => {
|
||||||
|
let emoji = crate::theme::EMOJI_HARMONY.to_string();
|
||||||
|
info!(emoji = emoji.as_str(); "Harmony completed");
|
||||||
|
}
|
||||||
HarmonyEvent::TopologyStateChanged {
|
HarmonyEvent::TopologyStateChanged {
|
||||||
topology,
|
topology,
|
||||||
status,
|
status,
|
||||||
message,
|
message,
|
||||||
} => {
|
} => match status {
|
||||||
let section_key = topology_key(&topology);
|
|
||||||
|
|
||||||
match status {
|
|
||||||
TopologyStatus::Queued => {}
|
TopologyStatus::Queued => {}
|
||||||
TopologyStatus::Preparing => {
|
TopologyStatus::Preparing => {
|
||||||
let section = progress::new_section(format!(
|
let emoji = format!(
|
||||||
"{} Preparing environment: {topology}...",
|
"{}",
|
||||||
crate::theme::EMOJI_TOPOLOGY,
|
style(crate::theme::EMOJI_TOPOLOGY.to_string()).yellow()
|
||||||
));
|
);
|
||||||
(*sections).insert(section_key, section);
|
info!(emoji = emoji.as_str(); "Preparing environment: {topology}...");
|
||||||
|
(*preparing_topology) = true;
|
||||||
}
|
}
|
||||||
TopologyStatus::Success => {
|
TopologyStatus::Success => {
|
||||||
let section = (*sections).get(§ion_key).unwrap();
|
(*preparing_topology) = false;
|
||||||
let progress = progress::add_spinner(section, "".into());
|
if let Some(message) = message {
|
||||||
|
info!(status = "finished"; "{message}");
|
||||||
progress::success(
|
}
|
||||||
section,
|
|
||||||
Some(progress),
|
|
||||||
message.unwrap_or("".into()),
|
|
||||||
);
|
|
||||||
|
|
||||||
(*sections).remove(§ion_key);
|
|
||||||
}
|
}
|
||||||
TopologyStatus::Noop => {
|
TopologyStatus::Noop => {
|
||||||
let section = (*sections).get(§ion_key).unwrap();
|
(*preparing_topology) = false;
|
||||||
let progress = progress::add_spinner(section, "".into());
|
if let Some(message) = message {
|
||||||
|
info!(status = "skipped"; "{message}");
|
||||||
progress::skip(
|
}
|
||||||
section,
|
|
||||||
Some(progress),
|
|
||||||
message.unwrap_or("".into()),
|
|
||||||
);
|
|
||||||
|
|
||||||
(*sections).remove(§ion_key);
|
|
||||||
}
|
}
|
||||||
TopologyStatus::Error => {
|
TopologyStatus::Error => {
|
||||||
let section = (*sections).get(§ion_key).unwrap();
|
(*preparing_topology) = false;
|
||||||
let progress = progress::add_spinner(section, "".into());
|
if let Some(message) = message {
|
||||||
|
error!(status = "failed"; "{message}");
|
||||||
progress::error(
|
|
||||||
section,
|
|
||||||
Some(progress),
|
|
||||||
message.unwrap_or("".into()),
|
|
||||||
);
|
|
||||||
|
|
||||||
(*sections).remove(§ion_key);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
},
|
||||||
HarmonyEvent::InterpretExecutionStarted {
|
HarmonyEvent::InterpretExecutionStarted {
|
||||||
topology,
|
execution_id: _,
|
||||||
interpret,
|
topology: _,
|
||||||
|
interpret: _,
|
||||||
score,
|
score,
|
||||||
message,
|
message,
|
||||||
} => {
|
} => {
|
||||||
let section_key = if (*sections).contains_key(&topology_key(&topology)) {
|
if *preparing_topology || current_score.is_some() {
|
||||||
topology_key(&topology)
|
info!("{message}");
|
||||||
} else if (*sections).contains_key(&score_key(&score)) {
|
|
||||||
score_key(&interpret)
|
|
||||||
} else {
|
} else {
|
||||||
let key = score_key(&score);
|
(*current_score) = Some(score.clone());
|
||||||
let section = progress::new_section(format!(
|
let emoji = format!("{}", style(crate::theme::EMOJI_SCORE).blue());
|
||||||
"\n{} Interpreting score: {score}...",
|
info!(emoji = emoji.as_str(); "Interpreting score: {score}...");
|
||||||
crate::theme::EMOJI_SCORE,
|
}
|
||||||
));
|
|
||||||
(*sections).insert(key.clone(), section);
|
|
||||||
key
|
|
||||||
};
|
|
||||||
let section = (*sections).get(§ion_key).unwrap();
|
|
||||||
let progress_bar = progress::add_spinner(section, message);
|
|
||||||
|
|
||||||
(*progress_bars).insert(interpret_key(&interpret), progress_bar);
|
|
||||||
}
|
}
|
||||||
HarmonyEvent::InterpretExecutionFinished {
|
HarmonyEvent::InterpretExecutionFinished {
|
||||||
topology,
|
execution_id: _,
|
||||||
interpret,
|
topology: _,
|
||||||
|
interpret: _,
|
||||||
score,
|
score,
|
||||||
outcome,
|
outcome,
|
||||||
} => {
|
} => {
|
||||||
let has_topology = (*sections).contains_key(&topology_key(&topology));
|
if current_score.is_some() && ¤t_score.clone().unwrap() == score {
|
||||||
let section_key = if has_topology {
|
(*current_score) = None;
|
||||||
topology_key(&topology)
|
}
|
||||||
} else {
|
|
||||||
score_key(&score)
|
|
||||||
};
|
|
||||||
|
|
||||||
let section = (*sections).get(§ion_key).unwrap();
|
|
||||||
let progress_bar =
|
|
||||||
(*progress_bars).get(&interpret_key(&interpret)).cloned();
|
|
||||||
|
|
||||||
let _ = section.clear();
|
|
||||||
|
|
||||||
match outcome {
|
match outcome {
|
||||||
Ok(outcome) => match outcome.status {
|
Ok(outcome) => match outcome.status {
|
||||||
harmony::interpret::InterpretStatus::SUCCESS => {
|
harmony::interpret::InterpretStatus::SUCCESS => {
|
||||||
progress::success(section, progress_bar, outcome.message)
|
info!(status = "finished"; "{}", outcome.message);
|
||||||
}
|
}
|
||||||
harmony::interpret::InterpretStatus::NOOP => {
|
harmony::interpret::InterpretStatus::NOOP => {
|
||||||
progress::skip(section, progress_bar, outcome.message)
|
info!(status = "skipped"; "{}", outcome.message);
|
||||||
|
}
|
||||||
|
_ => {
|
||||||
|
error!(status = "failed"; "{}", outcome.message);
|
||||||
}
|
}
|
||||||
_ => progress::error(section, progress_bar, outcome.message),
|
|
||||||
},
|
},
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
progress::error(section, progress_bar, err.to_string());
|
error!(status = "failed"; "{err}");
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if !has_topology {
|
|
||||||
(*progress_bars).remove(§ion_key);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
true
|
HarmonyEvent::ApplicationFeatureStateChanged {
|
||||||
|
topology: _,
|
||||||
|
application,
|
||||||
|
feature,
|
||||||
|
status,
|
||||||
|
} => match status {
|
||||||
|
ApplicationFeatureStatus::Installing => {
|
||||||
|
info!("Installing feature '{feature}' for '{application}'...");
|
||||||
|
}
|
||||||
|
ApplicationFeatureStatus::Installed => {
|
||||||
|
info!(status = "finished"; "Feature '{feature}' installed");
|
||||||
|
}
|
||||||
|
ApplicationFeatureStatus::Failed { details } => {
|
||||||
|
error!(status = "failed"; "Feature '{feature}' installation failed: {details}");
|
||||||
|
}
|
||||||
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
})
|
});
|
||||||
.await;
|
|
||||||
}
|
|
||||||
|
|
||||||
fn topology_key(topology: &str) -> String {
|
|
||||||
format!("topology-{topology}")
|
|
||||||
}
|
|
||||||
|
|
||||||
fn score_key(score: &str) -> String {
|
|
||||||
format!("score-{score}")
|
|
||||||
}
|
|
||||||
|
|
||||||
fn interpret_key(interpret: &str) -> String {
|
|
||||||
format!("interpret-{interpret}")
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,5 +1,6 @@
|
|||||||
use clap::Parser;
|
use clap::Parser;
|
||||||
use clap::builder::ArgPredicate;
|
use clap::builder::ArgPredicate;
|
||||||
|
use harmony::instrumentation;
|
||||||
use harmony::inventory::Inventory;
|
use harmony::inventory::Inventory;
|
||||||
use harmony::maestro::Maestro;
|
use harmony::maestro::Maestro;
|
||||||
use harmony::{score::Score, topology::Topology};
|
use harmony::{score::Score, topology::Topology};
|
||||||
@@ -89,37 +90,46 @@ pub async fn run<T: Topology + Send + Sync + 'static>(
|
|||||||
topology: T,
|
topology: T,
|
||||||
scores: Vec<Box<dyn Score<T>>>,
|
scores: Vec<Box<dyn Score<T>>>,
|
||||||
args_struct: Option<Args>,
|
args_struct: Option<Args>,
|
||||||
) -> Result<(), Box<dyn std::error::Error>> {
|
|
||||||
let cli_logger_handle = cli_logger::init();
|
|
||||||
|
|
||||||
let mut maestro = Maestro::initialize(inventory, topology).await.unwrap();
|
|
||||||
maestro.register_all(scores);
|
|
||||||
|
|
||||||
let result = init(maestro, args_struct).await;
|
|
||||||
|
|
||||||
let _ = tokio::try_join!(cli_logger_handle);
|
|
||||||
result
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn init<T: Topology + Send + Sync + 'static>(
|
|
||||||
maestro: harmony::maestro::Maestro<T>,
|
|
||||||
args_struct: Option<Args>,
|
|
||||||
) -> Result<(), Box<dyn std::error::Error>> {
|
) -> Result<(), Box<dyn std::error::Error>> {
|
||||||
let args = match args_struct {
|
let args = match args_struct {
|
||||||
Some(args) => args,
|
Some(args) => args,
|
||||||
None => Args::parse(),
|
None => Args::parse(),
|
||||||
};
|
};
|
||||||
|
|
||||||
#[cfg(feature = "tui")]
|
|
||||||
if args.interactive {
|
|
||||||
return harmony_tui::init(maestro).await;
|
|
||||||
}
|
|
||||||
|
|
||||||
#[cfg(not(feature = "tui"))]
|
#[cfg(not(feature = "tui"))]
|
||||||
if args.interactive {
|
if args.interactive {
|
||||||
return Err("Not compiled with interactive support".into());
|
return Err("Not compiled with interactive support".into());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[cfg(feature = "tui")]
|
||||||
|
if args.interactive {
|
||||||
|
return harmony_tui::run(inventory, topology, scores).await;
|
||||||
|
}
|
||||||
|
|
||||||
|
run_cli(inventory, topology, scores, args).await
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn run_cli<T: Topology + Send + Sync + 'static>(
|
||||||
|
inventory: Inventory,
|
||||||
|
topology: T,
|
||||||
|
scores: Vec<Box<dyn Score<T>>>,
|
||||||
|
args: Args,
|
||||||
|
) -> Result<(), Box<dyn std::error::Error>> {
|
||||||
|
cli_logger::init();
|
||||||
|
|
||||||
|
let mut maestro = Maestro::initialize(inventory, topology).await.unwrap();
|
||||||
|
maestro.register_all(scores);
|
||||||
|
|
||||||
|
let result = init(maestro, args).await;
|
||||||
|
|
||||||
|
instrumentation::instrument(instrumentation::HarmonyEvent::HarmonyFinished).unwrap();
|
||||||
|
result
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn init<T: Topology + Send + Sync + 'static>(
|
||||||
|
maestro: harmony::maestro::Maestro<T>,
|
||||||
|
args: Args,
|
||||||
|
) -> Result<(), Box<dyn std::error::Error>> {
|
||||||
let _ = env_logger::builder().try_init();
|
let _ = env_logger::builder().try_init();
|
||||||
|
|
||||||
let scores_vec = maestro_scores_filter(&maestro, args.all, args.filter, args.number);
|
let scores_vec = maestro_scores_filter(&maestro, args.all, args.filter, args.number);
|
||||||
@@ -130,8 +140,9 @@ async fn init<T: Topology + Send + Sync + 'static>(
|
|||||||
|
|
||||||
// if list option is specified, print filtered list and exit
|
// if list option is specified, print filtered list and exit
|
||||||
if args.list {
|
if args.list {
|
||||||
println!("Available scores:");
|
let num_scores = scores_vec.len();
|
||||||
println!("{}", list_scores_with_index(&scores_vec));
|
println!("Available scores {num_scores}:");
|
||||||
|
println!("{}\n\n", list_scores_with_index(&scores_vec));
|
||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -163,7 +174,7 @@ async fn init<T: Topology + Send + Sync + 'static>(
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod test {
|
mod tests {
|
||||||
use harmony::{
|
use harmony::{
|
||||||
inventory::Inventory,
|
inventory::Inventory,
|
||||||
maestro::Maestro,
|
maestro::Maestro,
|
||||||
@@ -190,14 +201,14 @@ mod test {
|
|||||||
let maestro = init_test_maestro();
|
let maestro = init_test_maestro();
|
||||||
let res = crate::init(
|
let res = crate::init(
|
||||||
maestro,
|
maestro,
|
||||||
Some(crate::Args {
|
crate::Args {
|
||||||
yes: true,
|
yes: true,
|
||||||
filter: Some("SuccessScore".to_owned()),
|
filter: Some("SuccessScore".to_owned()),
|
||||||
interactive: false,
|
interactive: false,
|
||||||
all: true,
|
all: true,
|
||||||
number: 0,
|
number: 0,
|
||||||
list: false,
|
list: false,
|
||||||
}),
|
},
|
||||||
)
|
)
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
@@ -210,14 +221,14 @@ mod test {
|
|||||||
|
|
||||||
let res = crate::init(
|
let res = crate::init(
|
||||||
maestro,
|
maestro,
|
||||||
Some(crate::Args {
|
crate::Args {
|
||||||
yes: true,
|
yes: true,
|
||||||
filter: Some("ErrorScore".to_owned()),
|
filter: Some("ErrorScore".to_owned()),
|
||||||
interactive: false,
|
interactive: false,
|
||||||
all: true,
|
all: true,
|
||||||
number: 0,
|
number: 0,
|
||||||
list: false,
|
list: false,
|
||||||
}),
|
},
|
||||||
)
|
)
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
@@ -230,14 +241,14 @@ mod test {
|
|||||||
|
|
||||||
let res = crate::init(
|
let res = crate::init(
|
||||||
maestro,
|
maestro,
|
||||||
Some(crate::Args {
|
crate::Args {
|
||||||
yes: true,
|
yes: true,
|
||||||
filter: None,
|
filter: None,
|
||||||
interactive: false,
|
interactive: false,
|
||||||
all: false,
|
all: false,
|
||||||
number: 0,
|
number: 0,
|
||||||
list: false,
|
list: false,
|
||||||
}),
|
},
|
||||||
)
|
)
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
|
|||||||
@@ -1,50 +1,147 @@
|
|||||||
|
use indicatif::{MultiProgress, ProgressBar};
|
||||||
|
use std::collections::HashMap;
|
||||||
|
use std::sync::{Arc, Mutex};
|
||||||
use std::time::Duration;
|
use std::time::Duration;
|
||||||
|
|
||||||
use indicatif::{MultiProgress, ProgressBar};
|
pub trait ProgressTracker: Send + Sync {
|
||||||
|
fn contains_section(&self, id: &str) -> bool;
|
||||||
pub fn new_section(title: String) -> MultiProgress {
|
fn add_section(&self, id: &str, message: &str);
|
||||||
let multi_progress = MultiProgress::new();
|
fn add_task(&self, section_id: &str, task_id: &str, message: &str);
|
||||||
let _ = multi_progress.println(title);
|
fn finish_task(&self, id: &str, message: &str);
|
||||||
|
fn fail_task(&self, id: &str, message: &str);
|
||||||
multi_progress
|
fn skip_task(&self, id: &str, message: &str);
|
||||||
|
fn clear(&self);
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn add_spinner(multi_progress: &MultiProgress, message: String) -> ProgressBar {
|
struct Section {
|
||||||
let progress = multi_progress.add(ProgressBar::new_spinner());
|
header_index: usize,
|
||||||
|
task_count: usize,
|
||||||
progress.set_style(crate::theme::SPINNER_STYLE.clone());
|
pb: ProgressBar,
|
||||||
progress.set_message(message);
|
|
||||||
progress.enable_steady_tick(Duration::from_millis(100));
|
|
||||||
|
|
||||||
progress
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn success(multi_progress: &MultiProgress, progress: Option<ProgressBar>, message: String) {
|
struct IndicatifProgressTrackerState {
|
||||||
if let Some(progress) = progress {
|
sections: HashMap<String, Section>,
|
||||||
multi_progress.remove(&progress)
|
tasks: HashMap<String, ProgressBar>,
|
||||||
|
pb_count: usize,
|
||||||
}
|
}
|
||||||
|
|
||||||
let progress = multi_progress.add(ProgressBar::new_spinner());
|
#[derive(Clone)]
|
||||||
progress.set_style(crate::theme::SUCCESS_SPINNER_STYLE.clone());
|
pub struct IndicatifProgressTracker {
|
||||||
progress.finish_with_message(message);
|
mp: MultiProgress,
|
||||||
|
state: Arc<Mutex<IndicatifProgressTrackerState>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn error(multi_progress: &MultiProgress, progress: Option<ProgressBar>, message: String) {
|
impl IndicatifProgressTracker {
|
||||||
if let Some(progress) = progress {
|
pub fn new(base: MultiProgress) -> Self {
|
||||||
multi_progress.remove(&progress)
|
let sections = HashMap::new();
|
||||||
|
let tasks = HashMap::new();
|
||||||
|
|
||||||
|
let state = Arc::new(Mutex::new(IndicatifProgressTrackerState {
|
||||||
|
sections,
|
||||||
|
tasks,
|
||||||
|
pb_count: 0,
|
||||||
|
}));
|
||||||
|
|
||||||
|
Self { mp: base, state }
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
let progress = multi_progress.add(ProgressBar::new_spinner());
|
impl ProgressTracker for IndicatifProgressTracker {
|
||||||
progress.set_style(crate::theme::ERROR_SPINNER_STYLE.clone());
|
fn add_section(&self, id: &str, message: &str) {
|
||||||
progress.finish_with_message(message);
|
let mut state = self.state.lock().unwrap();
|
||||||
|
|
||||||
|
let header_pb = self
|
||||||
|
.mp
|
||||||
|
.add(ProgressBar::new(1).with_style(crate::theme::SECTION_STYLE.clone()));
|
||||||
|
header_pb.finish_with_message(message.to_string());
|
||||||
|
|
||||||
|
let header_index = state.pb_count;
|
||||||
|
state.pb_count += 1;
|
||||||
|
|
||||||
|
state.sections.insert(
|
||||||
|
id.to_string(),
|
||||||
|
Section {
|
||||||
|
header_index,
|
||||||
|
task_count: 0,
|
||||||
|
pb: header_pb,
|
||||||
|
},
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn skip(multi_progress: &MultiProgress, progress: Option<ProgressBar>, message: String) {
|
fn add_task(&self, section_id: &str, task_id: &str, message: &str) {
|
||||||
if let Some(progress) = progress {
|
let mut state = self.state.lock().unwrap();
|
||||||
multi_progress.remove(&progress)
|
|
||||||
|
let insertion_index = {
|
||||||
|
let current_section = state
|
||||||
|
.sections
|
||||||
|
.get(section_id)
|
||||||
|
.expect("Section ID not found");
|
||||||
|
current_section.header_index + current_section.task_count + 1 // +1 to insert after header
|
||||||
|
};
|
||||||
|
|
||||||
|
let pb = self.mp.insert(insertion_index, ProgressBar::new_spinner());
|
||||||
|
pb.set_style(crate::theme::SPINNER_STYLE.clone());
|
||||||
|
pb.set_prefix(" ");
|
||||||
|
pb.set_message(message.to_string());
|
||||||
|
pb.enable_steady_tick(Duration::from_millis(80));
|
||||||
|
|
||||||
|
state.pb_count += 1;
|
||||||
|
|
||||||
|
let section = state
|
||||||
|
.sections
|
||||||
|
.get_mut(section_id)
|
||||||
|
.expect("Section ID not found");
|
||||||
|
section.task_count += 1;
|
||||||
|
|
||||||
|
// We inserted a new progress bar, so we must update the header_index
|
||||||
|
// for all subsequent sections.
|
||||||
|
for (id, s) in state.sections.iter_mut() {
|
||||||
|
if id != section_id && s.header_index >= insertion_index {
|
||||||
|
s.header_index += 1;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
let progress = multi_progress.add(ProgressBar::new_spinner());
|
state.tasks.insert(task_id.to_string(), pb);
|
||||||
progress.set_style(crate::theme::SKIP_SPINNER_STYLE.clone());
|
}
|
||||||
progress.finish_with_message(message);
|
|
||||||
|
fn finish_task(&self, id: &str, message: &str) {
|
||||||
|
let state = self.state.lock().unwrap();
|
||||||
|
if let Some(pb) = state.tasks.get(id) {
|
||||||
|
pb.set_style(crate::theme::SUCCESS_SPINNER_STYLE.clone());
|
||||||
|
pb.finish_with_message(message.to_string());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn fail_task(&self, id: &str, message: &str) {
|
||||||
|
let state = self.state.lock().unwrap();
|
||||||
|
if let Some(pb) = state.tasks.get(id) {
|
||||||
|
pb.set_style(crate::theme::ERROR_SPINNER_STYLE.clone());
|
||||||
|
pb.finish_with_message(message.to_string());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn skip_task(&self, id: &str, message: &str) {
|
||||||
|
let state = self.state.lock().unwrap();
|
||||||
|
if let Some(pb) = state.tasks.get(id) {
|
||||||
|
pb.set_style(crate::theme::SKIP_SPINNER_STYLE.clone());
|
||||||
|
pb.finish_with_message(message.to_string());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn contains_section(&self, id: &str) -> bool {
|
||||||
|
let state = self.state.lock().unwrap();
|
||||||
|
state.sections.contains_key(id)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn clear(&self) {
|
||||||
|
let mut state = self.state.lock().unwrap();
|
||||||
|
|
||||||
|
state.tasks.values().for_each(|p| self.mp.remove(p));
|
||||||
|
state.tasks.clear();
|
||||||
|
state.sections.values().for_each(|s| self.mp.remove(&s.pb));
|
||||||
|
state.sections.clear();
|
||||||
|
state.pb_count = 0;
|
||||||
|
|
||||||
|
let _ = self.mp.clear();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -11,17 +11,24 @@ pub static EMOJI_TOPOLOGY: Emoji<'_, '_> = Emoji("📦", "");
|
|||||||
pub static EMOJI_SCORE: Emoji<'_, '_> = Emoji("🎶", "");
|
pub static EMOJI_SCORE: Emoji<'_, '_> = Emoji("🎶", "");
|
||||||
|
|
||||||
lazy_static! {
|
lazy_static! {
|
||||||
|
pub static ref SECTION_STYLE: ProgressStyle = ProgressStyle::default_spinner()
|
||||||
|
.template("{wide_msg:.bold}")
|
||||||
|
.unwrap();
|
||||||
pub static ref SPINNER_STYLE: ProgressStyle = ProgressStyle::default_spinner()
|
pub static ref SPINNER_STYLE: ProgressStyle = ProgressStyle::default_spinner()
|
||||||
.template(" {spinner:.green} {msg}")
|
.template(" {spinner:.green} {wide_msg}")
|
||||||
.unwrap()
|
.unwrap()
|
||||||
.tick_strings(&["⠋", "⠙", "⠹", "⠸", "⠼", "⠴", "⠦", "⠧", "⠇", "⠏"]);
|
.tick_strings(&["⠋", "⠙", "⠹", "⠸", "⠼", "⠴", "⠦", "⠧", "⠇", "⠏"]);
|
||||||
pub static ref SUCCESS_SPINNER_STYLE: ProgressStyle = SPINNER_STYLE
|
pub static ref SUCCESS_SPINNER_STYLE: ProgressStyle = SPINNER_STYLE
|
||||||
.clone()
|
.clone()
|
||||||
.tick_strings(&[format!("{}", EMOJI_SUCCESS).as_str()]);
|
.tick_strings(&[format!("{}", EMOJI_SUCCESS).as_str()]);
|
||||||
pub static ref SKIP_SPINNER_STYLE: ProgressStyle = SPINNER_STYLE
|
pub static ref SKIP_SPINNER_STYLE: ProgressStyle = ProgressStyle::default_spinner()
|
||||||
|
.template(" {spinner:.orange} {wide_msg}")
|
||||||
|
.unwrap()
|
||||||
.clone()
|
.clone()
|
||||||
.tick_strings(&[format!("{}", EMOJI_SKIP).as_str()]);
|
.tick_strings(&[format!("{}", EMOJI_SKIP).as_str()]);
|
||||||
pub static ref ERROR_SPINNER_STYLE: ProgressStyle = SPINNER_STYLE
|
pub static ref ERROR_SPINNER_STYLE: ProgressStyle = ProgressStyle::default_spinner()
|
||||||
|
.template(" {spinner:.red} {wide_msg}")
|
||||||
|
.unwrap()
|
||||||
.clone()
|
.clone()
|
||||||
.tick_strings(&[format!("{}", EMOJI_ERROR).as_str()]);
|
.tick_strings(&[format!("{}", EMOJI_ERROR).as_str()]);
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,112 +1,66 @@
|
|||||||
use indicatif::{MultiProgress, ProgressBar};
|
use harmony_cli::progress::{IndicatifProgressTracker, ProgressTracker};
|
||||||
use indicatif_log_bridge::LogWrapper;
|
use indicatif::MultiProgress;
|
||||||
use log::error;
|
|
||||||
use std::{
|
|
||||||
collections::HashMap,
|
|
||||||
sync::{Arc, Mutex},
|
|
||||||
};
|
|
||||||
|
|
||||||
use crate::instrumentation::{self, HarmonyComposerEvent};
|
use crate::instrumentation::{self, HarmonyComposerEvent};
|
||||||
|
|
||||||
pub fn init() -> tokio::task::JoinHandle<()> {
|
pub fn init() {
|
||||||
configure_logger();
|
configure_logger();
|
||||||
let handle = tokio::spawn(handle_events());
|
handle_events();
|
||||||
|
|
||||||
loop {
|
|
||||||
if instrumentation::instrument(HarmonyComposerEvent::HarmonyComposerStarted).is_ok() {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
handle
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn configure_logger() {
|
fn configure_logger() {
|
||||||
let logger =
|
|
||||||
env_logger::Builder::from_env(env_logger::Env::default().default_filter_or("info")).build();
|
env_logger::Builder::from_env(env_logger::Env::default().default_filter_or("info")).build();
|
||||||
let level = logger.filter();
|
|
||||||
let multi = MultiProgress::new();
|
|
||||||
LogWrapper::new(multi.clone(), logger).try_init().unwrap();
|
|
||||||
log::set_max_level(level);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn handle_events() {
|
pub fn handle_events() {
|
||||||
const PROGRESS_SETUP: &str = "project-initialization";
|
let progress_tracker = IndicatifProgressTracker::new(MultiProgress::new());
|
||||||
|
|
||||||
|
const SETUP_SECTION: &str = "project-initialization";
|
||||||
|
const COMPILTATION_TASK: &str = "compilation";
|
||||||
const PROGRESS_DEPLOYMENT: &str = "deployment";
|
const PROGRESS_DEPLOYMENT: &str = "deployment";
|
||||||
|
|
||||||
instrumentation::subscribe("Harmony Composer Logger", {
|
instrumentation::subscribe("Harmony Composer Logger", {
|
||||||
let progresses: Arc<Mutex<HashMap<String, MultiProgress>>> =
|
move |event| match event {
|
||||||
Arc::new(Mutex::new(HashMap::new()));
|
|
||||||
let compilation_progress = Arc::new(Mutex::new(None::<ProgressBar>));
|
|
||||||
|
|
||||||
move |event| {
|
|
||||||
let progresses_clone = Arc::clone(&progresses);
|
|
||||||
let compilation_progress_clone = Arc::clone(&compilation_progress);
|
|
||||||
|
|
||||||
async move {
|
|
||||||
let mut progresses_guard = progresses_clone.lock().unwrap();
|
|
||||||
let mut compilation_progress_guard = compilation_progress_clone.lock().unwrap();
|
|
||||||
|
|
||||||
match event {
|
|
||||||
HarmonyComposerEvent::HarmonyComposerStarted => {}
|
HarmonyComposerEvent::HarmonyComposerStarted => {}
|
||||||
HarmonyComposerEvent::ProjectInitializationStarted => {
|
HarmonyComposerEvent::ProjectInitializationStarted => {
|
||||||
let multi_progress = harmony_cli::progress::new_section(format!(
|
progress_tracker.add_section(
|
||||||
|
SETUP_SECTION,
|
||||||
|
&format!(
|
||||||
"{} Initializing Harmony project...",
|
"{} Initializing Harmony project...",
|
||||||
harmony_cli::theme::EMOJI_HARMONY,
|
harmony_cli::theme::EMOJI_HARMONY,
|
||||||
));
|
),
|
||||||
(*progresses_guard).insert(PROGRESS_SETUP.to_string(), multi_progress);
|
);
|
||||||
}
|
}
|
||||||
HarmonyComposerEvent::ProjectInitialized => println!("\n"),
|
HarmonyComposerEvent::ProjectInitialized => {}
|
||||||
HarmonyComposerEvent::ProjectCompilationStarted { details } => {
|
HarmonyComposerEvent::ProjectCompilationStarted { details } => {
|
||||||
let initialization_progress =
|
progress_tracker.add_task(SETUP_SECTION, COMPILTATION_TASK, details);
|
||||||
(*progresses_guard).get(PROGRESS_SETUP).unwrap();
|
|
||||||
let _ = initialization_progress.clear();
|
|
||||||
|
|
||||||
let progress =
|
|
||||||
harmony_cli::progress::add_spinner(initialization_progress, details);
|
|
||||||
*compilation_progress_guard = Some(progress);
|
|
||||||
}
|
}
|
||||||
HarmonyComposerEvent::ProjectCompiled => {
|
HarmonyComposerEvent::ProjectCompiled => {
|
||||||
let initialization_progress =
|
progress_tracker.finish_task(COMPILTATION_TASK, "project compiled");
|
||||||
(*progresses_guard).get(PROGRESS_SETUP).unwrap();
|
|
||||||
|
|
||||||
harmony_cli::progress::success(
|
|
||||||
initialization_progress,
|
|
||||||
(*compilation_progress_guard).take(),
|
|
||||||
"project compiled".to_string(),
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
HarmonyComposerEvent::ProjectCompilationFailed { details } => {
|
HarmonyComposerEvent::ProjectCompilationFailed { details } => {
|
||||||
let initialization_progress =
|
progress_tracker.fail_task(
|
||||||
(*progresses_guard).get(PROGRESS_SETUP).unwrap();
|
COMPILTATION_TASK,
|
||||||
|
&format!("failed to compile project:\n{details}"),
|
||||||
harmony_cli::progress::error(
|
|
||||||
initialization_progress,
|
|
||||||
(*compilation_progress_guard).take(),
|
|
||||||
"failed to compile project".to_string(),
|
|
||||||
);
|
);
|
||||||
|
|
||||||
error!("{details}");
|
|
||||||
}
|
}
|
||||||
HarmonyComposerEvent::DeploymentStarted { target } => {
|
HarmonyComposerEvent::DeploymentStarted { target, profile } => {
|
||||||
let multi_progress = harmony_cli::progress::new_section(format!(
|
progress_tracker.add_section(
|
||||||
"{} Starting deployment to {target}...\n\n",
|
PROGRESS_DEPLOYMENT,
|
||||||
harmony_cli::theme::EMOJI_DEPLOY
|
&format!(
|
||||||
));
|
"\n{} Deploying project on target '{target}' with profile '{profile}'...\n",
|
||||||
(*progresses_guard).insert(PROGRESS_DEPLOYMENT.to_string(), multi_progress);
|
harmony_cli::theme::EMOJI_DEPLOY,
|
||||||
|
),
|
||||||
|
);
|
||||||
}
|
}
|
||||||
HarmonyComposerEvent::DeploymentCompleted => println!("\n"),
|
HarmonyComposerEvent::DeploymentCompleted => {
|
||||||
HarmonyComposerEvent::Shutdown => {
|
progress_tracker.clear();
|
||||||
for (_, progresses) in (*progresses_guard).iter() {
|
|
||||||
progresses.clear().unwrap();
|
|
||||||
}
|
}
|
||||||
|
HarmonyComposerEvent::DeploymentFailed { details } => {
|
||||||
return false;
|
progress_tracker.add_task(PROGRESS_DEPLOYMENT, "deployment-failed", "");
|
||||||
}
|
progress_tracker.fail_task("deployment-failed", details);
|
||||||
}
|
|
||||||
true
|
|
||||||
}
|
}
|
||||||
|
HarmonyComposerEvent::Shutdown => {}
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
.await
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,53 +1,68 @@
|
|||||||
use log::debug;
|
|
||||||
use once_cell::sync::Lazy;
|
use once_cell::sync::Lazy;
|
||||||
use tokio::sync::broadcast;
|
use std::{collections::HashMap, sync::Mutex};
|
||||||
|
|
||||||
|
use crate::{HarmonyProfile, HarmonyTarget};
|
||||||
|
|
||||||
#[derive(Debug, Clone)]
|
#[derive(Debug, Clone)]
|
||||||
pub enum HarmonyComposerEvent {
|
pub enum HarmonyComposerEvent {
|
||||||
HarmonyComposerStarted,
|
HarmonyComposerStarted,
|
||||||
ProjectInitializationStarted,
|
ProjectInitializationStarted,
|
||||||
ProjectInitialized,
|
ProjectInitialized,
|
||||||
ProjectCompilationStarted { details: String },
|
ProjectCompilationStarted {
|
||||||
|
details: String,
|
||||||
|
},
|
||||||
ProjectCompiled,
|
ProjectCompiled,
|
||||||
ProjectCompilationFailed { details: String },
|
ProjectCompilationFailed {
|
||||||
DeploymentStarted { target: String },
|
details: String,
|
||||||
|
},
|
||||||
|
DeploymentStarted {
|
||||||
|
target: HarmonyTarget,
|
||||||
|
profile: HarmonyProfile,
|
||||||
|
},
|
||||||
DeploymentCompleted,
|
DeploymentCompleted,
|
||||||
|
DeploymentFailed {
|
||||||
|
details: String,
|
||||||
|
},
|
||||||
Shutdown,
|
Shutdown,
|
||||||
}
|
}
|
||||||
|
|
||||||
static HARMONY_COMPOSER_EVENT_BUS: Lazy<broadcast::Sender<HarmonyComposerEvent>> =
|
type Subscriber = Box<dyn Fn(&HarmonyComposerEvent) + Send + Sync>;
|
||||||
Lazy::new(|| {
|
|
||||||
// TODO: Adjust channel capacity
|
|
||||||
let (tx, _rx) = broadcast::channel(16);
|
|
||||||
tx
|
|
||||||
});
|
|
||||||
|
|
||||||
pub fn instrument(event: HarmonyComposerEvent) -> Result<(), &'static str> {
|
static SUBSCRIBERS: Lazy<Mutex<HashMap<String, Subscriber>>> =
|
||||||
match HARMONY_COMPOSER_EVENT_BUS.send(event) {
|
Lazy::new(|| Mutex::new(HashMap::new()));
|
||||||
Ok(_) => Ok(()),
|
|
||||||
Err(_) => Err("send error: no subscribers"),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
pub async fn subscribe<F, Fut>(name: &str, mut handler: F)
|
/// Subscribes a listener to all instrumentation events.
|
||||||
|
///
|
||||||
|
/// Simply provide a unique name and a closure to run when an event happens.
|
||||||
|
///
|
||||||
|
/// # Example
|
||||||
|
/// ```
|
||||||
|
/// instrumentation::subscribe("my_logger", |event| {
|
||||||
|
/// println!("Event occurred: {:?}", event);
|
||||||
|
/// });
|
||||||
|
/// ```
|
||||||
|
pub fn subscribe<F>(name: &str, callback: F)
|
||||||
where
|
where
|
||||||
F: FnMut(HarmonyComposerEvent) -> Fut + Send + 'static,
|
F: Fn(&HarmonyComposerEvent) + Send + Sync + 'static,
|
||||||
Fut: Future<Output = bool> + Send,
|
|
||||||
{
|
{
|
||||||
let mut rx = HARMONY_COMPOSER_EVENT_BUS.subscribe();
|
let mut subs = SUBSCRIBERS.lock().unwrap();
|
||||||
debug!("[{name}] Service started. Listening for events...");
|
subs.insert(name.to_string(), Box::new(callback));
|
||||||
loop {
|
|
||||||
match rx.recv().await {
|
|
||||||
Ok(event) => {
|
|
||||||
if !handler(event).await {
|
|
||||||
debug!("[{name}] Handler requested exit.");
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Err(broadcast::error::RecvError::Lagged(n)) => {
|
|
||||||
debug!("[{name}] Lagged behind by {n} messages.");
|
|
||||||
}
|
|
||||||
Err(_) => break,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Instruments an event, notifying all subscribers.
|
||||||
|
///
|
||||||
|
/// This will call every closure that was registered with `subscribe`.
|
||||||
|
///
|
||||||
|
/// # Example
|
||||||
|
/// ```
|
||||||
|
/// instrumentation::instrument(HarmonyEvent::HarmonyStarted);
|
||||||
|
/// ```
|
||||||
|
pub fn instrument(event: HarmonyComposerEvent) -> Result<(), &'static str> {
|
||||||
|
let subs = SUBSCRIBERS.lock().unwrap();
|
||||||
|
|
||||||
|
for callback in subs.values() {
|
||||||
|
callback(&event);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -49,14 +49,11 @@ struct CheckArgs {
|
|||||||
|
|
||||||
#[derive(Args, Clone, Debug)]
|
#[derive(Args, Clone, Debug)]
|
||||||
struct DeployArgs {
|
struct DeployArgs {
|
||||||
#[arg(long, default_value_t = false)]
|
#[arg(long = "target", short = 't', default_value = "local")]
|
||||||
staging: bool,
|
harmony_target: HarmonyTarget,
|
||||||
|
|
||||||
#[arg(long, default_value_t = false)]
|
#[arg(long = "profile", short = 'p', default_value = "dev")]
|
||||||
prod: bool,
|
harmony_profile: HarmonyProfile,
|
||||||
|
|
||||||
#[arg(long, default_value_t = false)]
|
|
||||||
smoke_test: bool,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Args, Clone, Debug)]
|
#[derive(Args, Clone, Debug)]
|
||||||
@@ -68,9 +65,41 @@ struct AllArgs {
|
|||||||
deploy: DeployArgs,
|
deploy: DeployArgs,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Clone, Debug, clap::ValueEnum)]
|
||||||
|
enum HarmonyTarget {
|
||||||
|
Local,
|
||||||
|
Remote,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl std::fmt::Display for HarmonyTarget {
|
||||||
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||||
|
match self {
|
||||||
|
HarmonyTarget::Local => f.write_str("local"),
|
||||||
|
HarmonyTarget::Remote => f.write_str("remote"),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Clone, Debug, clap::ValueEnum)]
|
||||||
|
enum HarmonyProfile {
|
||||||
|
Dev,
|
||||||
|
Staging,
|
||||||
|
Production,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl std::fmt::Display for HarmonyProfile {
|
||||||
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||||
|
match self {
|
||||||
|
HarmonyProfile::Dev => f.write_str("dev"),
|
||||||
|
HarmonyProfile::Staging => f.write_str("staging"),
|
||||||
|
HarmonyProfile::Production => f.write_str("production"),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#[tokio::main]
|
#[tokio::main]
|
||||||
async fn main() {
|
async fn main() {
|
||||||
let hc_logger_handle = harmony_composer_logger::init();
|
harmony_composer_logger::init();
|
||||||
let cli_args = GlobalArgs::parse();
|
let cli_args = GlobalArgs::parse();
|
||||||
|
|
||||||
let harmony_path = Path::new(&cli_args.harmony_path)
|
let harmony_path = Path::new(&cli_args.harmony_path)
|
||||||
@@ -122,26 +151,39 @@ async fn main() {
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
Commands::Deploy(args) => {
|
Commands::Deploy(args) => {
|
||||||
let deploy = if args.staging {
|
|
||||||
instrumentation::instrument(HarmonyComposerEvent::DeploymentStarted {
|
instrumentation::instrument(HarmonyComposerEvent::DeploymentStarted {
|
||||||
target: "staging".to_string(),
|
target: args.harmony_target.clone(),
|
||||||
|
profile: args.harmony_profile.clone(),
|
||||||
})
|
})
|
||||||
.unwrap();
|
.unwrap();
|
||||||
todo!("implement staging deployment")
|
|
||||||
} else if args.prod {
|
if matches!(args.harmony_profile, HarmonyProfile::Dev)
|
||||||
instrumentation::instrument(HarmonyComposerEvent::DeploymentStarted {
|
&& !matches!(args.harmony_target, HarmonyTarget::Local)
|
||||||
target: "prod".to_string(),
|
{
|
||||||
})
|
instrumentation::instrument(HarmonyComposerEvent::DeploymentFailed {
|
||||||
.unwrap();
|
details: format!(
|
||||||
todo!("implement prod deployment")
|
"Cannot run profile '{}' on target '{}'. Profile '{}' can run locally only.",
|
||||||
} else {
|
args.harmony_profile, args.harmony_target, args.harmony_profile
|
||||||
instrumentation::instrument(HarmonyComposerEvent::DeploymentStarted {
|
),
|
||||||
target: "dev".to_string(),
|
}).unwrap();
|
||||||
})
|
return;
|
||||||
.unwrap();
|
|
||||||
Command::new(harmony_bin_path).arg("-y").arg("-a").spawn()
|
|
||||||
}
|
}
|
||||||
.expect("failed to run harmony deploy");
|
|
||||||
|
let use_local_k3d = match args.harmony_target {
|
||||||
|
HarmonyTarget::Local => true,
|
||||||
|
HarmonyTarget::Remote => false,
|
||||||
|
};
|
||||||
|
|
||||||
|
let mut command = Command::new(harmony_bin_path);
|
||||||
|
command
|
||||||
|
.env("HARMONY_USE_LOCAL_K3D", format!("{use_local_k3d}"))
|
||||||
|
.env("HARMONY_PROFILE", format!("{}", args.harmony_profile))
|
||||||
|
.arg("-y")
|
||||||
|
.arg("-a");
|
||||||
|
|
||||||
|
info!("{:?}", command);
|
||||||
|
|
||||||
|
let deploy = command.spawn().expect("failed to run harmony deploy");
|
||||||
|
|
||||||
let deploy_output = deploy.wait_with_output().unwrap();
|
let deploy_output = deploy.wait_with_output().unwrap();
|
||||||
debug!("{}", String::from_utf8(deploy_output.stdout).unwrap());
|
debug!("{}", String::from_utf8(deploy_output.stdout).unwrap());
|
||||||
@@ -157,8 +199,6 @@ async fn main() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
instrumentation::instrument(HarmonyComposerEvent::Shutdown).unwrap();
|
instrumentation::instrument(HarmonyComposerEvent::Shutdown).unwrap();
|
||||||
|
|
||||||
let _ = tokio::try_join!(hc_logger_handle);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug, clap::ValueEnum)]
|
#[derive(Clone, Debug, clap::ValueEnum)]
|
||||||
|
|||||||
17
harmony_inventory_agent/Cargo.toml
Normal file
17
harmony_inventory_agent/Cargo.toml
Normal file
@@ -0,0 +1,17 @@
|
|||||||
|
[package]
|
||||||
|
name = "harmony_inventory_agent"
|
||||||
|
version = "0.1.0"
|
||||||
|
edition = "2024"
|
||||||
|
|
||||||
|
[dependencies]
|
||||||
|
actix-web = "4.4"
|
||||||
|
sysinfo = "0.30"
|
||||||
|
serde.workspace = true
|
||||||
|
serde_json.workspace = true
|
||||||
|
log.workspace = true
|
||||||
|
env_logger.workspace = true
|
||||||
|
tokio.workspace = true
|
||||||
|
thiserror.workspace = true
|
||||||
|
# mdns-sd = "0.14.1"
|
||||||
|
mdns-sd = { git = "https://github.com/jggc/mdns-sd.git", branch = "patch-1" }
|
||||||
|
local-ip-address = "0.6.5"
|
||||||
22
harmony_inventory_agent/README.md
Normal file
22
harmony_inventory_agent/README.md
Normal file
@@ -0,0 +1,22 @@
|
|||||||
|
## Compiling :
|
||||||
|
|
||||||
|
```bash
|
||||||
|
cargo build -p harmony_inventory_agent --release --target x86_64-unknown-linux-musl
|
||||||
|
```
|
||||||
|
|
||||||
|
This will create a statically linked binary that can run on pretty much any x86_64 system.
|
||||||
|
|
||||||
|
This requires installation of the target
|
||||||
|
|
||||||
|
```
|
||||||
|
rustup target add x86_64-unknown-linux-musl
|
||||||
|
```
|
||||||
|
|
||||||
|
And installation of the musl tools too.
|
||||||
|
|
||||||
|
On Archlinux, they can be installed with :
|
||||||
|
|
||||||
|
```
|
||||||
|
pacman -S musl
|
||||||
|
```
|
||||||
|
|
||||||
840
harmony_inventory_agent/src/hwinfo.rs
Normal file
840
harmony_inventory_agent/src/hwinfo.rs
Normal file
@@ -0,0 +1,840 @@
|
|||||||
|
use log::{debug, warn};
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
use serde_json::Value;
|
||||||
|
use std::fs;
|
||||||
|
use std::path::Path;
|
||||||
|
use std::process::Command;
|
||||||
|
use sysinfo::System;
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
|
pub struct PhysicalHost {
|
||||||
|
pub storage_drives: Vec<StorageDrive>,
|
||||||
|
pub storage_controller: StorageController,
|
||||||
|
pub memory_modules: Vec<MemoryModule>,
|
||||||
|
pub cpus: Vec<CPU>,
|
||||||
|
pub chipset: Chipset,
|
||||||
|
pub network_interfaces: Vec<NetworkInterface>,
|
||||||
|
pub management_interface: Option<ManagementInterface>,
|
||||||
|
pub host_uuid: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
|
pub struct StorageDrive {
|
||||||
|
pub name: String,
|
||||||
|
pub model: String,
|
||||||
|
pub serial: String,
|
||||||
|
pub size_bytes: u64,
|
||||||
|
pub logical_block_size: u32,
|
||||||
|
pub physical_block_size: u32,
|
||||||
|
pub rotational: bool,
|
||||||
|
pub wwn: Option<String>,
|
||||||
|
pub interface_type: String,
|
||||||
|
pub smart_status: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
|
pub struct StorageController {
|
||||||
|
pub name: String,
|
||||||
|
pub driver: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
|
pub struct MemoryModule {
|
||||||
|
pub size_bytes: u64,
|
||||||
|
pub speed_mhz: Option<u32>,
|
||||||
|
pub manufacturer: Option<String>,
|
||||||
|
pub part_number: Option<String>,
|
||||||
|
pub serial_number: Option<String>,
|
||||||
|
pub rank: Option<u8>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
|
pub struct CPU {
|
||||||
|
pub model: String,
|
||||||
|
pub vendor: String,
|
||||||
|
pub cores: u32,
|
||||||
|
pub threads: u32,
|
||||||
|
pub frequency_mhz: u64,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
|
pub struct Chipset {
|
||||||
|
pub name: String,
|
||||||
|
pub vendor: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
|
pub struct NetworkInterface {
|
||||||
|
pub name: String,
|
||||||
|
pub mac_address: String,
|
||||||
|
pub speed_mbps: Option<u32>,
|
||||||
|
pub is_up: bool,
|
||||||
|
pub mtu: u32,
|
||||||
|
pub ipv4_addresses: Vec<String>,
|
||||||
|
pub ipv6_addresses: Vec<String>,
|
||||||
|
pub driver: String,
|
||||||
|
pub firmware_version: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
|
pub struct ManagementInterface {
|
||||||
|
pub kind: String,
|
||||||
|
pub address: Option<String>,
|
||||||
|
pub firmware: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl PhysicalHost {
|
||||||
|
pub fn gather() -> Result<Self, String> {
|
||||||
|
let mut sys = System::new_all();
|
||||||
|
sys.refresh_all();
|
||||||
|
|
||||||
|
Self::all_tools_available()?;
|
||||||
|
|
||||||
|
Ok(Self {
|
||||||
|
storage_drives: Self::gather_storage_drives()?,
|
||||||
|
storage_controller: Self::gather_storage_controller()?,
|
||||||
|
memory_modules: Self::gather_memory_modules()?,
|
||||||
|
cpus: Self::gather_cpus(&sys)?,
|
||||||
|
chipset: Self::gather_chipset()?,
|
||||||
|
network_interfaces: Self::gather_network_interfaces()?,
|
||||||
|
management_interface: Self::gather_management_interface()?,
|
||||||
|
host_uuid: Self::get_host_uuid()?,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
fn all_tools_available() -> Result<(), String> {
|
||||||
|
let required_tools = [
|
||||||
|
("lsblk", Some("--version")),
|
||||||
|
("lspci", Some("--version")),
|
||||||
|
("lsmod", None),
|
||||||
|
("dmidecode", Some("--version")),
|
||||||
|
("smartctl", Some("--version")),
|
||||||
|
("ip", Some("route")), // No version flag available
|
||||||
|
];
|
||||||
|
|
||||||
|
let mut missing_tools = Vec::new();
|
||||||
|
|
||||||
|
debug!("Looking for required_tools {required_tools:?}");
|
||||||
|
for (tool, tool_arg) in required_tools.iter() {
|
||||||
|
// First check if tool exists in PATH using which(1)
|
||||||
|
let mut exists = if let Ok(output) = Command::new("which").arg(tool).output() {
|
||||||
|
output.status.success()
|
||||||
|
} else {
|
||||||
|
false
|
||||||
|
};
|
||||||
|
|
||||||
|
if !exists {
|
||||||
|
// Fallback: manual PATH search if which(1) is unavailable
|
||||||
|
debug!("Looking for {tool} in path");
|
||||||
|
if let Ok(path_var) = std::env::var("PATH") {
|
||||||
|
debug!("PATH is {path_var}");
|
||||||
|
exists = path_var.split(':').any(|dir| {
|
||||||
|
let tool_path = std::path::Path::new(dir).join(tool);
|
||||||
|
tool_path.exists() && Self::is_executable(&tool_path)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if !exists {
|
||||||
|
warn!("Unable to find tool {tool} from PATH");
|
||||||
|
missing_tools.push(*tool);
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Verify tool is functional by checking version/help output
|
||||||
|
let mut cmd = Command::new(tool);
|
||||||
|
if let Some(tool_arg) = tool_arg {
|
||||||
|
cmd.arg(tool_arg);
|
||||||
|
}
|
||||||
|
cmd.stdout(std::process::Stdio::null());
|
||||||
|
cmd.stderr(std::process::Stdio::null());
|
||||||
|
|
||||||
|
if let Ok(status) = cmd.status() {
|
||||||
|
if !status.success() {
|
||||||
|
warn!("Unable to test {tool} status failed");
|
||||||
|
missing_tools.push(*tool);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
warn!("Unable to test {tool}");
|
||||||
|
missing_tools.push(*tool);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if !missing_tools.is_empty() {
|
||||||
|
let missing_str = missing_tools
|
||||||
|
.iter()
|
||||||
|
.map(|s| s.to_string())
|
||||||
|
.collect::<Vec<String>>()
|
||||||
|
.join(", ");
|
||||||
|
return Err(format!(
|
||||||
|
"The following required tools are not available: {}. Please install these tools to use PhysicalHost::gather()",
|
||||||
|
missing_str
|
||||||
|
));
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(unix)]
|
||||||
|
fn is_executable(path: &std::path::Path) -> bool {
|
||||||
|
debug!("Checking if {} is executable", path.to_string_lossy());
|
||||||
|
use std::os::unix::fs::PermissionsExt;
|
||||||
|
|
||||||
|
match std::fs::metadata(path) {
|
||||||
|
Ok(meta) => meta.permissions().mode() & 0o111 != 0,
|
||||||
|
Err(_) => false,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(not(unix))]
|
||||||
|
fn is_executable(_path: &std::path::Path) -> bool {
|
||||||
|
// On non-Unix systems, we assume existence implies executability
|
||||||
|
true
|
||||||
|
}
|
||||||
|
|
||||||
|
fn gather_storage_drives() -> Result<Vec<StorageDrive>, String> {
|
||||||
|
let mut drives = Vec::new();
|
||||||
|
|
||||||
|
// Use lsblk with JSON output for robust parsing
|
||||||
|
let output = Command::new("lsblk")
|
||||||
|
.args([
|
||||||
|
"-d",
|
||||||
|
"-o",
|
||||||
|
"NAME,MODEL,SERIAL,SIZE,ROTA,WWN",
|
||||||
|
"-n",
|
||||||
|
"-e",
|
||||||
|
"7",
|
||||||
|
"--json",
|
||||||
|
])
|
||||||
|
.output()
|
||||||
|
.map_err(|e| format!("Failed to execute lsblk: {}", e))?;
|
||||||
|
|
||||||
|
if !output.status.success() {
|
||||||
|
return Err(format!(
|
||||||
|
"lsblk command failed: {}",
|
||||||
|
String::from_utf8_lossy(&output.stderr)
|
||||||
|
));
|
||||||
|
}
|
||||||
|
|
||||||
|
let json: Value = serde_json::from_slice(&output.stdout)
|
||||||
|
.map_err(|e| format!("Failed to parse lsblk JSON output: {}", e))?;
|
||||||
|
|
||||||
|
let blockdevices = json
|
||||||
|
.get("blockdevices")
|
||||||
|
.and_then(|v| v.as_array())
|
||||||
|
.ok_or("Invalid lsblk JSON: missing 'blockdevices' array")?;
|
||||||
|
|
||||||
|
for device in blockdevices {
|
||||||
|
let name = device
|
||||||
|
.get("name")
|
||||||
|
.and_then(|v| v.as_str())
|
||||||
|
.ok_or("Missing 'name' in lsblk device")?
|
||||||
|
.to_string();
|
||||||
|
|
||||||
|
if name.is_empty() {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
let model = device
|
||||||
|
.get("model")
|
||||||
|
.and_then(|v| v.as_str())
|
||||||
|
.map(|s| s.trim().to_string())
|
||||||
|
.unwrap_or_default();
|
||||||
|
|
||||||
|
let serial = device
|
||||||
|
.get("serial")
|
||||||
|
.and_then(|v| v.as_str())
|
||||||
|
.map(|s| s.trim().to_string())
|
||||||
|
.unwrap_or_default();
|
||||||
|
|
||||||
|
let size_str = device
|
||||||
|
.get("size")
|
||||||
|
.and_then(|v| v.as_str())
|
||||||
|
.ok_or("Missing 'size' in lsblk device")?;
|
||||||
|
let size_bytes = Self::parse_size(size_str)?;
|
||||||
|
|
||||||
|
let rotational = device
|
||||||
|
.get("rota")
|
||||||
|
.and_then(|v| v.as_bool())
|
||||||
|
.ok_or("Missing 'rota' in lsblk device")?;
|
||||||
|
|
||||||
|
let wwn = device
|
||||||
|
.get("wwn")
|
||||||
|
.and_then(|v| v.as_str())
|
||||||
|
.map(|s| s.trim().to_string())
|
||||||
|
.filter(|s| !s.is_empty() && s != "null");
|
||||||
|
|
||||||
|
let device_path = Path::new("/sys/block").join(&name);
|
||||||
|
|
||||||
|
let logical_block_size = Self::read_sysfs_u32(
|
||||||
|
&device_path.join("queue/logical_block_size"),
|
||||||
|
)
|
||||||
|
.map_err(|e| format!("Failed to read logical block size for {}: {}", name, e))?;
|
||||||
|
|
||||||
|
let physical_block_size = Self::read_sysfs_u32(
|
||||||
|
&device_path.join("queue/physical_block_size"),
|
||||||
|
)
|
||||||
|
.map_err(|e| format!("Failed to read physical block size for {}: {}", name, e))?;
|
||||||
|
|
||||||
|
let interface_type = Self::get_interface_type(&name, &device_path)?;
|
||||||
|
let smart_status = Self::get_smart_status(&name)?;
|
||||||
|
|
||||||
|
let mut drive = StorageDrive {
|
||||||
|
name: name.clone(),
|
||||||
|
model,
|
||||||
|
serial,
|
||||||
|
size_bytes,
|
||||||
|
logical_block_size,
|
||||||
|
physical_block_size,
|
||||||
|
rotational,
|
||||||
|
wwn,
|
||||||
|
interface_type,
|
||||||
|
smart_status,
|
||||||
|
};
|
||||||
|
|
||||||
|
// Enhance with additional sysfs info if available
|
||||||
|
if device_path.exists() {
|
||||||
|
if drive.model.is_empty() {
|
||||||
|
drive.model = Self::read_sysfs_string(&device_path.join("device/model"))
|
||||||
|
.unwrap_or(format!("Failed to read model for {}", name));
|
||||||
|
}
|
||||||
|
if drive.serial.is_empty() {
|
||||||
|
drive.serial = Self::read_sysfs_string(&device_path.join("device/serial"))
|
||||||
|
.unwrap_or(format!("Failed to read serial for {}", name));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
drives.push(drive);
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(drives)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn gather_storage_controller() -> Result<StorageController, String> {
|
||||||
|
let mut controller = StorageController {
|
||||||
|
name: "Unknown".to_string(),
|
||||||
|
driver: "Unknown".to_string(),
|
||||||
|
};
|
||||||
|
|
||||||
|
// Use lspci with JSON output if available
|
||||||
|
let output = Command::new("lspci")
|
||||||
|
.args(["-nn", "-d", "::0100", "-J"]) // Storage controllers class with JSON
|
||||||
|
.output()
|
||||||
|
.map_err(|e| format!("Failed to execute lspci: {}", e))?;
|
||||||
|
|
||||||
|
if output.status.success() {
|
||||||
|
let json: Value = serde_json::from_slice(&output.stdout)
|
||||||
|
.map_err(|e| format!("Failed to parse lspci JSON output: {}", e))?;
|
||||||
|
|
||||||
|
if let Some(devices) = json.as_array() {
|
||||||
|
for device in devices {
|
||||||
|
if let Some(device_info) = device.as_object()
|
||||||
|
&& let Some(name) = device_info
|
||||||
|
.get("device")
|
||||||
|
.and_then(|v| v.as_object())
|
||||||
|
.and_then(|v| v.get("name"))
|
||||||
|
.and_then(|v| v.as_str())
|
||||||
|
{
|
||||||
|
controller.name = name.to_string();
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Fallback to text output if JSON fails or no device found
|
||||||
|
if controller.name == "Unknown" {
|
||||||
|
let output = Command::new("lspci")
|
||||||
|
.args(["-nn", "-d", "::0100"]) // Storage controllers class
|
||||||
|
.output()
|
||||||
|
.map_err(|e| format!("Failed to execute lspci (fallback): {}", e))?;
|
||||||
|
|
||||||
|
if output.status.success() {
|
||||||
|
let output_str = String::from_utf8_lossy(&output.stdout);
|
||||||
|
if let Some(line) = output_str.lines().next() {
|
||||||
|
let parts: Vec<&str> = line.split(':').collect();
|
||||||
|
if parts.len() > 2 {
|
||||||
|
controller.name = parts[2].trim().to_string();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Try to get driver info from lsmod
|
||||||
|
let output = Command::new("lsmod")
|
||||||
|
.output()
|
||||||
|
.map_err(|e| format!("Failed to execute lsmod: {}", e))?;
|
||||||
|
|
||||||
|
if output.status.success() {
|
||||||
|
let output_str = String::from_utf8_lossy(&output.stdout);
|
||||||
|
for line in output_str.lines() {
|
||||||
|
if line.contains("ahci")
|
||||||
|
|| line.contains("nvme")
|
||||||
|
|| line.contains("megaraid")
|
||||||
|
|| line.contains("mpt3sas")
|
||||||
|
{
|
||||||
|
let parts: Vec<&str> = line.split_whitespace().collect();
|
||||||
|
if !parts.is_empty() {
|
||||||
|
controller.driver = parts[0].to_string();
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(controller)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn gather_memory_modules() -> Result<Vec<MemoryModule>, String> {
|
||||||
|
let mut modules = Vec::new();
|
||||||
|
|
||||||
|
let output = Command::new("dmidecode")
|
||||||
|
.arg("--type")
|
||||||
|
.arg("17")
|
||||||
|
.output()
|
||||||
|
.map_err(|e| format!("Failed to execute dmidecode: {}", e))?;
|
||||||
|
|
||||||
|
if !output.status.success() {
|
||||||
|
return Err(format!(
|
||||||
|
"dmidecode command failed: {}",
|
||||||
|
String::from_utf8_lossy(&output.stderr)
|
||||||
|
));
|
||||||
|
}
|
||||||
|
|
||||||
|
let output_str = String::from_utf8(output.stdout)
|
||||||
|
.map_err(|e| format!("Failed to parse dmidecode output: {}", e))?;
|
||||||
|
|
||||||
|
let sections: Vec<&str> = output_str.split("Memory Device").collect();
|
||||||
|
|
||||||
|
for section in sections.into_iter().skip(1) {
|
||||||
|
let mut module = MemoryModule {
|
||||||
|
size_bytes: 0,
|
||||||
|
speed_mhz: None,
|
||||||
|
manufacturer: None,
|
||||||
|
part_number: None,
|
||||||
|
serial_number: None,
|
||||||
|
rank: None,
|
||||||
|
};
|
||||||
|
|
||||||
|
for line in section.lines() {
|
||||||
|
let line = line.trim();
|
||||||
|
if let Some(size_str) = line.strip_prefix("Size: ") {
|
||||||
|
if size_str != "No Module Installed"
|
||||||
|
&& let Some((num, unit)) = size_str.split_once(' ')
|
||||||
|
&& let Ok(num) = num.parse::<u64>()
|
||||||
|
{
|
||||||
|
module.size_bytes = match unit {
|
||||||
|
"MB" => num * 1024 * 1024,
|
||||||
|
"GB" => num * 1024 * 1024 * 1024,
|
||||||
|
"KB" => num * 1024,
|
||||||
|
_ => 0,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
} else if let Some(speed_str) = line.strip_prefix("Speed: ") {
|
||||||
|
if let Some((num, _unit)) = speed_str.split_once(' ') {
|
||||||
|
module.speed_mhz = num.parse().ok();
|
||||||
|
}
|
||||||
|
} else if let Some(man) = line.strip_prefix("Manufacturer: ") {
|
||||||
|
module.manufacturer = Some(man.to_string());
|
||||||
|
} else if let Some(part) = line.strip_prefix("Part Number: ") {
|
||||||
|
module.part_number = Some(part.to_string());
|
||||||
|
} else if let Some(serial) = line.strip_prefix("Serial Number: ") {
|
||||||
|
module.serial_number = Some(serial.to_string());
|
||||||
|
} else if let Some(rank) = line.strip_prefix("Rank: ") {
|
||||||
|
module.rank = rank.parse().ok();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if module.size_bytes > 0 {
|
||||||
|
modules.push(module);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(modules)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn gather_cpus(sys: &System) -> Result<Vec<CPU>, String> {
|
||||||
|
let mut cpus = Vec::new();
|
||||||
|
let global_cpu = sys.global_cpu_info();
|
||||||
|
|
||||||
|
cpus.push(CPU {
|
||||||
|
model: global_cpu.brand().to_string(),
|
||||||
|
vendor: global_cpu.vendor_id().to_string(),
|
||||||
|
cores: sys.physical_core_count().unwrap_or(1) as u32,
|
||||||
|
threads: sys.cpus().len() as u32,
|
||||||
|
frequency_mhz: global_cpu.frequency(),
|
||||||
|
});
|
||||||
|
|
||||||
|
Ok(cpus)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn gather_chipset() -> Result<Chipset, String> {
|
||||||
|
Ok(Chipset {
|
||||||
|
name: Self::read_dmi("baseboard-product-name")?,
|
||||||
|
vendor: Self::read_dmi("baseboard-manufacturer")?,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
fn gather_network_interfaces() -> Result<Vec<NetworkInterface>, String> {
|
||||||
|
let mut interfaces = Vec::new();
|
||||||
|
let sys_net_path = Path::new("/sys/class/net");
|
||||||
|
|
||||||
|
let entries = fs::read_dir(sys_net_path)
|
||||||
|
.map_err(|e| format!("Failed to read /sys/class/net: {}", e))?;
|
||||||
|
|
||||||
|
for entry in entries {
|
||||||
|
let entry = entry.map_err(|e| format!("Failed to read directory entry: {}", e))?;
|
||||||
|
let iface_name = entry
|
||||||
|
.file_name()
|
||||||
|
.into_string()
|
||||||
|
.map_err(|_| "Invalid UTF-8 in interface name")?;
|
||||||
|
let iface_path = entry.path();
|
||||||
|
|
||||||
|
// Skip virtual interfaces
|
||||||
|
if iface_name.starts_with("lo")
|
||||||
|
|| iface_name.starts_with("docker")
|
||||||
|
|| iface_name.starts_with("virbr")
|
||||||
|
|| iface_name.starts_with("veth")
|
||||||
|
|| iface_name.starts_with("br-")
|
||||||
|
|| iface_name.starts_with("tun")
|
||||||
|
|| iface_name.starts_with("wg")
|
||||||
|
{
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check if it's a physical interface by looking for device directory
|
||||||
|
if !iface_path.join("device").exists() {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
let mac_address = Self::read_sysfs_string(&iface_path.join("address"))
|
||||||
|
.map_err(|e| format!("Failed to read MAC address for {}: {}", iface_name, e))?;
|
||||||
|
|
||||||
|
let speed_mbps = if iface_path.join("speed").exists() {
|
||||||
|
match Self::read_sysfs_u32(&iface_path.join("speed")) {
|
||||||
|
Ok(speed) => Some(speed),
|
||||||
|
Err(e) => {
|
||||||
|
debug!(
|
||||||
|
"Failed to read speed for {}: {} . This is expected to fail on wifi interfaces.",
|
||||||
|
iface_name, e
|
||||||
|
);
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
};
|
||||||
|
|
||||||
|
let operstate = Self::read_sysfs_string(&iface_path.join("operstate"))
|
||||||
|
.map_err(|e| format!("Failed to read operstate for {}: {}", iface_name, e))?;
|
||||||
|
|
||||||
|
let mtu = Self::read_sysfs_u32(&iface_path.join("mtu"))
|
||||||
|
.map_err(|e| format!("Failed to read MTU for {}: {}", iface_name, e))?;
|
||||||
|
|
||||||
|
let driver =
|
||||||
|
Self::read_sysfs_symlink_basename(&iface_path.join("device/driver/module"))
|
||||||
|
.map_err(|e| format!("Failed to read driver for {}: {}", iface_name, e))?;
|
||||||
|
|
||||||
|
let firmware_version = Self::read_sysfs_opt_string(
|
||||||
|
&iface_path.join("device/firmware_version"),
|
||||||
|
)
|
||||||
|
.map_err(|e| format!("Failed to read firmware version for {}: {}", iface_name, e))?;
|
||||||
|
|
||||||
|
// Get IP addresses using ip command with JSON output
|
||||||
|
let (ipv4_addresses, ipv6_addresses) = Self::get_interface_ips_json(&iface_name)
|
||||||
|
.map_err(|e| format!("Failed to get IP addresses for {}: {}", iface_name, e))?;
|
||||||
|
|
||||||
|
interfaces.push(NetworkInterface {
|
||||||
|
name: iface_name,
|
||||||
|
mac_address,
|
||||||
|
speed_mbps,
|
||||||
|
is_up: operstate == "up",
|
||||||
|
mtu,
|
||||||
|
ipv4_addresses,
|
||||||
|
ipv6_addresses,
|
||||||
|
driver,
|
||||||
|
firmware_version,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(interfaces)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn gather_management_interface() -> Result<Option<ManagementInterface>, String> {
|
||||||
|
if Path::new("/dev/ipmi0").exists() {
|
||||||
|
Ok(Some(ManagementInterface {
|
||||||
|
kind: "IPMI".to_string(),
|
||||||
|
address: None,
|
||||||
|
firmware: Some(Self::read_dmi("bios-version")?),
|
||||||
|
}))
|
||||||
|
} else if Path::new("/sys/class/misc/mei").exists() {
|
||||||
|
Ok(Some(ManagementInterface {
|
||||||
|
kind: "Intel ME".to_string(),
|
||||||
|
address: None,
|
||||||
|
firmware: None,
|
||||||
|
}))
|
||||||
|
} else {
|
||||||
|
Ok(None)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_host_uuid() -> Result<String, String> {
|
||||||
|
Self::read_dmi("system-uuid")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Helper methods
|
||||||
|
fn read_sysfs_string(path: &Path) -> Result<String, String> {
|
||||||
|
fs::read_to_string(path)
|
||||||
|
.map(|s| s.trim().to_string())
|
||||||
|
.map_err(|e| format!("Failed to read {}: {}", path.display(), e))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn read_sysfs_opt_string(path: &Path) -> Result<Option<String>, String> {
|
||||||
|
match fs::read_to_string(path) {
|
||||||
|
Ok(s) => {
|
||||||
|
let s = s.trim().to_string();
|
||||||
|
Ok(if s.is_empty() { None } else { Some(s) })
|
||||||
|
}
|
||||||
|
Err(e) if e.kind() == std::io::ErrorKind::NotFound => Ok(None),
|
||||||
|
Err(e) => Err(format!("Failed to read {}: {}", path.display(), e)),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn read_sysfs_u32(path: &Path) -> Result<u32, String> {
|
||||||
|
fs::read_to_string(path)
|
||||||
|
.map_err(|e| format!("Failed to read {}: {}", path.display(), e))?
|
||||||
|
.trim()
|
||||||
|
.parse()
|
||||||
|
.map_err(|e| format!("Failed to parse {}: {}", path.display(), e))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn read_sysfs_symlink_basename(path: &Path) -> Result<String, String> {
|
||||||
|
match fs::read_link(path) {
|
||||||
|
Ok(target_path) => match target_path.file_name() {
|
||||||
|
Some(name_osstr) => match name_osstr.to_str() {
|
||||||
|
Some(name_str) => Ok(name_str.to_string()),
|
||||||
|
None => Err(format!(
|
||||||
|
"Symlink target basename is not valid UTF-8: {}",
|
||||||
|
target_path.display()
|
||||||
|
)),
|
||||||
|
},
|
||||||
|
None => Err(format!(
|
||||||
|
"Symlink target has no basename: {} -> {}",
|
||||||
|
path.display(),
|
||||||
|
target_path.display()
|
||||||
|
)),
|
||||||
|
},
|
||||||
|
Err(e) if e.kind() == std::io::ErrorKind::NotFound => Err(format!(
|
||||||
|
"Could not resolve symlink for path : {}",
|
||||||
|
path.display()
|
||||||
|
)),
|
||||||
|
Err(e) => Err(format!("Failed to read symlink {}: {}", path.display(), e)),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn read_dmi(field: &str) -> Result<String, String> {
|
||||||
|
let output = Command::new("dmidecode")
|
||||||
|
.arg("-s")
|
||||||
|
.arg(field)
|
||||||
|
.output()
|
||||||
|
.map_err(|e| format!("Failed to execute dmidecode for field {}: {}", field, e))?;
|
||||||
|
|
||||||
|
if !output.status.success() {
|
||||||
|
return Err(format!(
|
||||||
|
"dmidecode command failed for field {}: {}",
|
||||||
|
field,
|
||||||
|
String::from_utf8_lossy(&output.stderr)
|
||||||
|
));
|
||||||
|
}
|
||||||
|
|
||||||
|
String::from_utf8(output.stdout)
|
||||||
|
.map(|s| s.trim().to_string())
|
||||||
|
.map_err(|e| {
|
||||||
|
format!(
|
||||||
|
"Failed to parse dmidecode output for field {}: {}",
|
||||||
|
field, e
|
||||||
|
)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_interface_type(device_name: &str, device_path: &Path) -> Result<String, String> {
|
||||||
|
if device_name.starts_with("nvme") {
|
||||||
|
Ok("NVMe".to_string())
|
||||||
|
} else if device_name.starts_with("sd") {
|
||||||
|
Ok("SATA".to_string())
|
||||||
|
} else if device_name.starts_with("hd") {
|
||||||
|
Ok("IDE".to_string())
|
||||||
|
} else if device_name.starts_with("vd") {
|
||||||
|
Ok("VirtIO".to_string())
|
||||||
|
} else if device_name.starts_with("sr") {
|
||||||
|
Ok("CDROM".to_string())
|
||||||
|
} else if device_name.starts_with("zram") {
|
||||||
|
Ok("Ramdisk".to_string())
|
||||||
|
} else {
|
||||||
|
// Try to determine from device path
|
||||||
|
let subsystem = Self::read_sysfs_string(&device_path.join("device/subsystem"))?;
|
||||||
|
Ok(subsystem
|
||||||
|
.split('/')
|
||||||
|
.next_back()
|
||||||
|
.unwrap_or("Unknown")
|
||||||
|
.to_string())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_smart_status(device_name: &str) -> Result<Option<String>, String> {
|
||||||
|
let output = Command::new("smartctl")
|
||||||
|
.arg("-H")
|
||||||
|
.arg(format!("/dev/{}", device_name))
|
||||||
|
.output()
|
||||||
|
.map_err(|e| format!("Failed to execute smartctl for {}: {}", device_name, e))?;
|
||||||
|
|
||||||
|
if !output.status.success() {
|
||||||
|
return Ok(None);
|
||||||
|
}
|
||||||
|
|
||||||
|
let stdout = String::from_utf8(output.stdout)
|
||||||
|
.map_err(|e| format!("Failed to parse smartctl output for {}: {}", device_name, e))?;
|
||||||
|
|
||||||
|
for line in stdout.lines() {
|
||||||
|
if line.contains("SMART overall-health self-assessment") {
|
||||||
|
if let Some(status) = line.split(':').nth(1) {
|
||||||
|
return Ok(Some(status.trim().to_string()));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(None)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn parse_size(size_str: &str) -> Result<u64, String> {
|
||||||
|
debug!("Parsing size_str '{size_str}'");
|
||||||
|
let size;
|
||||||
|
if size_str.ends_with('T') {
|
||||||
|
size = size_str[..size_str.len() - 1]
|
||||||
|
.parse::<f64>()
|
||||||
|
.map(|t| t * 1024.0 * 1024.0 * 1024.0 * 1024.0)
|
||||||
|
.map_err(|e| format!("Failed to parse T size '{}': {}", size_str, e))
|
||||||
|
} else if size_str.ends_with('G') {
|
||||||
|
size = size_str[..size_str.len() - 1]
|
||||||
|
.parse::<f64>()
|
||||||
|
.map(|g| g * 1024.0 * 1024.0 * 1024.0)
|
||||||
|
.map_err(|e| format!("Failed to parse G size '{}': {}", size_str, e))
|
||||||
|
} else if size_str.ends_with('M') {
|
||||||
|
size = size_str[..size_str.len() - 1]
|
||||||
|
.parse::<f64>()
|
||||||
|
.map(|m| m * 1024.0 * 1024.0)
|
||||||
|
.map_err(|e| format!("Failed to parse M size '{}': {}", size_str, e))
|
||||||
|
} else if size_str.ends_with('K') {
|
||||||
|
size = size_str[..size_str.len() - 1]
|
||||||
|
.parse::<f64>()
|
||||||
|
.map(|k| k * 1024.0)
|
||||||
|
.map_err(|e| format!("Failed to parse K size '{}': {}", size_str, e))
|
||||||
|
} else if size_str.ends_with('B') {
|
||||||
|
size = size_str[..size_str.len() - 1]
|
||||||
|
.parse::<f64>()
|
||||||
|
.map_err(|e| format!("Failed to parse B size '{}': {}", size_str, e))
|
||||||
|
} else {
|
||||||
|
size = size_str
|
||||||
|
.parse::<f64>()
|
||||||
|
.map_err(|e| format!("Failed to parse size '{}': {}", size_str, e))
|
||||||
|
}
|
||||||
|
|
||||||
|
size.map(|s| s as u64)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_interface_ips_json(iface_name: &str) -> Result<(Vec<String>, Vec<String>), String> {
|
||||||
|
let mut ipv4 = Vec::new();
|
||||||
|
let mut ipv6 = Vec::new();
|
||||||
|
|
||||||
|
// Get IPv4 addresses using JSON output
|
||||||
|
let output = Command::new("ip")
|
||||||
|
.args(["-j", "-4", "addr", "show", iface_name])
|
||||||
|
.output()
|
||||||
|
.map_err(|e| {
|
||||||
|
format!(
|
||||||
|
"Failed to execute ip command for IPv4 on {}: {}",
|
||||||
|
iface_name, e
|
||||||
|
)
|
||||||
|
})?;
|
||||||
|
|
||||||
|
if !output.status.success() {
|
||||||
|
return Err(format!(
|
||||||
|
"ip command for IPv4 on {} failed: {}",
|
||||||
|
iface_name,
|
||||||
|
String::from_utf8_lossy(&output.stderr)
|
||||||
|
));
|
||||||
|
}
|
||||||
|
|
||||||
|
let json: Value = serde_json::from_slice(&output.stdout).map_err(|e| {
|
||||||
|
format!(
|
||||||
|
"Failed to parse ip JSON output for IPv4 on {}: {}",
|
||||||
|
iface_name, e
|
||||||
|
)
|
||||||
|
})?;
|
||||||
|
|
||||||
|
if let Some(addrs) = json.as_array() {
|
||||||
|
for addr_info in addrs {
|
||||||
|
if let Some(addr_info_obj) = addr_info.as_object()
|
||||||
|
&& let Some(addr_info) =
|
||||||
|
addr_info_obj.get("addr_info").and_then(|v| v.as_array())
|
||||||
|
{
|
||||||
|
for addr in addr_info {
|
||||||
|
if let Some(addr_obj) = addr.as_object()
|
||||||
|
&& let Some(ip) = addr_obj.get("local").and_then(|v| v.as_str())
|
||||||
|
{
|
||||||
|
ipv4.push(ip.to_string());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Get IPv6 addresses using JSON output
|
||||||
|
let output = Command::new("ip")
|
||||||
|
.args(["-j", "-6", "addr", "show", iface_name])
|
||||||
|
.output()
|
||||||
|
.map_err(|e| {
|
||||||
|
format!(
|
||||||
|
"Failed to execute ip command for IPv6 on {}: {}",
|
||||||
|
iface_name, e
|
||||||
|
)
|
||||||
|
})?;
|
||||||
|
|
||||||
|
if !output.status.success() {
|
||||||
|
return Err(format!(
|
||||||
|
"ip command for IPv6 on {} failed: {}",
|
||||||
|
iface_name,
|
||||||
|
String::from_utf8_lossy(&output.stderr)
|
||||||
|
));
|
||||||
|
}
|
||||||
|
|
||||||
|
let json: Value = serde_json::from_slice(&output.stdout).map_err(|e| {
|
||||||
|
format!(
|
||||||
|
"Failed to parse ip JSON output for IPv6 on {}: {}",
|
||||||
|
iface_name, e
|
||||||
|
)
|
||||||
|
})?;
|
||||||
|
|
||||||
|
if let Some(addrs) = json.as_array() {
|
||||||
|
for addr_info in addrs {
|
||||||
|
if let Some(addr_info_obj) = addr_info.as_object()
|
||||||
|
&& let Some(addr_info) =
|
||||||
|
addr_info_obj.get("addr_info").and_then(|v| v.as_array())
|
||||||
|
{
|
||||||
|
for addr in addr_info {
|
||||||
|
if let Some(addr_obj) = addr.as_object()
|
||||||
|
&& let Some(ip) = addr_obj.get("local").and_then(|v| v.as_str())
|
||||||
|
{
|
||||||
|
// Skip link-local addresses
|
||||||
|
if !ip.starts_with("fe80::") {
|
||||||
|
ipv6.push(ip.to_string());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok((ipv4, ipv6))
|
||||||
|
}
|
||||||
|
}
|
||||||
2
harmony_inventory_agent/src/lib.rs
Normal file
2
harmony_inventory_agent/src/lib.rs
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
mod hwinfo;
|
||||||
|
pub mod local_presence;
|
||||||
90
harmony_inventory_agent/src/local_presence/advertise.rs
Normal file
90
harmony_inventory_agent/src/local_presence/advertise.rs
Normal file
@@ -0,0 +1,90 @@
|
|||||||
|
use log::{debug, error, info, warn};
|
||||||
|
use mdns_sd::{ServiceDaemon, ServiceInfo};
|
||||||
|
use std::collections::HashMap;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
hwinfo::PhysicalHost,
|
||||||
|
local_presence::{PresenceError, SERVICE_NAME, VERSION},
|
||||||
|
};
|
||||||
|
|
||||||
|
/// Advertises the agent's presence on the local network.
|
||||||
|
///
|
||||||
|
/// This function is synchronous and non-blocking. It spawns a background Tokio task
|
||||||
|
/// to handle the mDNS advertisement for the lifetime of the application.
|
||||||
|
pub fn advertise(service_port: u16) -> Result<(), PresenceError> {
|
||||||
|
let host_id = match PhysicalHost::gather() {
|
||||||
|
Ok(host) => Some(host.host_uuid),
|
||||||
|
Err(e) => {
|
||||||
|
error!("Could not build physical host, harmony presence id will be unavailable : {e}");
|
||||||
|
None
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
let instance_name = format!(
|
||||||
|
"inventory-agent-{}",
|
||||||
|
host_id.clone().unwrap_or("unknown".to_string())
|
||||||
|
);
|
||||||
|
|
||||||
|
let spawned_msg = format!("Spawned local presence advertisement task for '{instance_name}'.");
|
||||||
|
|
||||||
|
tokio::spawn(async move {
|
||||||
|
info!(
|
||||||
|
"Local presence task started. Advertising as '{}'.",
|
||||||
|
instance_name
|
||||||
|
);
|
||||||
|
|
||||||
|
// The ServiceDaemon must live for the entire duration of the advertisement.
|
||||||
|
// If it's dropped, the advertisement stops.
|
||||||
|
let mdns = match ServiceDaemon::new() {
|
||||||
|
Ok(daemon) => daemon,
|
||||||
|
Err(e) => {
|
||||||
|
warn!("Failed to create mDNS daemon: {}. Task shutting down.", e);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
let mut props = HashMap::new();
|
||||||
|
if let Some(host_id) = host_id {
|
||||||
|
props.insert("id".to_string(), host_id);
|
||||||
|
}
|
||||||
|
props.insert("version".to_string(), VERSION.to_string());
|
||||||
|
|
||||||
|
let local_ip: Box<dyn mdns_sd::AsIpAddrs> = match local_ip_address::local_ip() {
|
||||||
|
Ok(ip) => Box::new(ip),
|
||||||
|
Err(e) => {
|
||||||
|
error!(
|
||||||
|
"Could not figure out local ip, mdns will have to try to figure it out by itself : {e}"
|
||||||
|
);
|
||||||
|
Box::new(())
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
debug!("Using local ip {:?}", local_ip.as_ip_addrs());
|
||||||
|
|
||||||
|
let service_info = ServiceInfo::new(
|
||||||
|
SERVICE_NAME,
|
||||||
|
&instance_name,
|
||||||
|
&format!("{}.local.", instance_name),
|
||||||
|
local_ip,
|
||||||
|
service_port,
|
||||||
|
Some(props),
|
||||||
|
)
|
||||||
|
.expect("ServiceInfo creation should not fail with valid inputs");
|
||||||
|
|
||||||
|
// The registration handle must also be kept alive.
|
||||||
|
let _registration_handle = match mdns.register(service_info) {
|
||||||
|
Ok(handle) => {
|
||||||
|
info!("Service successfully registered on the local network.");
|
||||||
|
handle
|
||||||
|
}
|
||||||
|
Err(e) => {
|
||||||
|
warn!("Failed to register service: {}. Task shutting down.", e);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
});
|
||||||
|
|
||||||
|
info!("{spawned_msg}");
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
34
harmony_inventory_agent/src/local_presence/discover.rs
Normal file
34
harmony_inventory_agent/src/local_presence/discover.rs
Normal file
@@ -0,0 +1,34 @@
|
|||||||
|
use mdns_sd::{ServiceDaemon, ServiceEvent};
|
||||||
|
|
||||||
|
use crate::local_presence::SERVICE_NAME;
|
||||||
|
|
||||||
|
pub type DiscoveryEvent = ServiceEvent;
|
||||||
|
|
||||||
|
pub fn discover_agents(timeout: Option<u64>, on_event: fn(&DiscoveryEvent)) {
|
||||||
|
// Create a new mDNS daemon.
|
||||||
|
let mdns = ServiceDaemon::new().expect("Failed to create mDNS daemon");
|
||||||
|
|
||||||
|
// Start browsing for the service type.
|
||||||
|
// The receiver will be a stream of events.
|
||||||
|
let receiver = mdns.browse(SERVICE_NAME).expect("Failed to browse");
|
||||||
|
|
||||||
|
std::thread::spawn(move || {
|
||||||
|
while let Ok(event) = receiver.recv() {
|
||||||
|
on_event(&event);
|
||||||
|
match event {
|
||||||
|
ServiceEvent::ServiceResolved(resolved) => {
|
||||||
|
println!("Resolved a new service: {}", resolved.fullname);
|
||||||
|
}
|
||||||
|
other_event => {
|
||||||
|
println!("Received other event: {:?}", &other_event);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
if let Some(timeout) = timeout {
|
||||||
|
// Gracefully shutdown the daemon.
|
||||||
|
std::thread::sleep(std::time::Duration::from_secs(timeout));
|
||||||
|
mdns.shutdown().unwrap();
|
||||||
|
}
|
||||||
|
}
|
||||||
16
harmony_inventory_agent/src/local_presence/mod.rs
Normal file
16
harmony_inventory_agent/src/local_presence/mod.rs
Normal file
@@ -0,0 +1,16 @@
|
|||||||
|
mod discover;
|
||||||
|
pub use discover::*;
|
||||||
|
mod advertise;
|
||||||
|
pub use advertise::*;
|
||||||
|
|
||||||
|
pub const SERVICE_NAME: &str = "_harmony._tcp.local.";
|
||||||
|
const VERSION: &str = env!("CARGO_PKG_VERSION");
|
||||||
|
|
||||||
|
// A specific error type for our module enhances clarity and usability.
|
||||||
|
#[derive(thiserror::Error, Debug)]
|
||||||
|
pub enum PresenceError {
|
||||||
|
#[error("Failed to create mDNS daemon")]
|
||||||
|
DaemonCreationFailed(#[from] mdns_sd::Error),
|
||||||
|
#[error("The shutdown signal has already been sent")]
|
||||||
|
ShutdownFailed,
|
||||||
|
}
|
||||||
47
harmony_inventory_agent/src/main.rs
Normal file
47
harmony_inventory_agent/src/main.rs
Normal file
@@ -0,0 +1,47 @@
|
|||||||
|
// src/main.rs
|
||||||
|
use actix_web::{App, HttpServer, Responder, get};
|
||||||
|
use log::error;
|
||||||
|
use std::env;
|
||||||
|
|
||||||
|
use crate::hwinfo::PhysicalHost;
|
||||||
|
|
||||||
|
mod hwinfo;
|
||||||
|
mod local_presence;
|
||||||
|
|
||||||
|
#[get("/inventory")]
|
||||||
|
async fn inventory() -> impl Responder {
|
||||||
|
log::info!("Received inventory request");
|
||||||
|
let host = PhysicalHost::gather();
|
||||||
|
match host {
|
||||||
|
Ok(host) => {
|
||||||
|
log::info!("Inventory data gathered successfully");
|
||||||
|
actix_web::HttpResponse::Ok().json(host)
|
||||||
|
}
|
||||||
|
Err(error) => {
|
||||||
|
log::error!("Inventory data gathering FAILED");
|
||||||
|
actix_web::HttpResponse::InternalServerError().json(error)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[actix_web::main]
|
||||||
|
async fn main() -> std::io::Result<()> {
|
||||||
|
env_logger::init();
|
||||||
|
|
||||||
|
let port = env::var("HARMONY_INVENTORY_AGENT_PORT").unwrap_or_else(|_| "8080".to_string());
|
||||||
|
let port = port
|
||||||
|
.parse::<u16>()
|
||||||
|
.expect(&format!("Invalid port number, cannot parse to u16 {port}"));
|
||||||
|
let bind_addr = format!("0.0.0.0:{}", port);
|
||||||
|
|
||||||
|
log::info!("Starting inventory agent on {}", bind_addr);
|
||||||
|
|
||||||
|
if let Err(e) = local_presence::advertise(port) {
|
||||||
|
error!("Could not start advertise local presence : {e}");
|
||||||
|
}
|
||||||
|
|
||||||
|
HttpServer::new(|| App::new().service(inventory))
|
||||||
|
.bind(&bind_addr)?
|
||||||
|
.run()
|
||||||
|
.await
|
||||||
|
}
|
||||||
23
harmony_secret/Cargo.toml
Normal file
23
harmony_secret/Cargo.toml
Normal file
@@ -0,0 +1,23 @@
|
|||||||
|
[package]
|
||||||
|
name = "harmony_secret"
|
||||||
|
edition = "2024"
|
||||||
|
version.workspace = true
|
||||||
|
readme.workspace = true
|
||||||
|
license.workspace = true
|
||||||
|
|
||||||
|
[dependencies]
|
||||||
|
harmony_secret_derive = { version = "0.1.0", path = "../harmony_secret_derive" }
|
||||||
|
serde = { version = "1.0.209", features = ["derive", "rc"] }
|
||||||
|
serde_json = "1.0.127"
|
||||||
|
thiserror.workspace = true
|
||||||
|
lazy_static.workspace = true
|
||||||
|
directories.workspace = true
|
||||||
|
log.workspace = true
|
||||||
|
infisical = "0.0.2"
|
||||||
|
tokio.workspace = true
|
||||||
|
async-trait.workspace = true
|
||||||
|
http.workspace = true
|
||||||
|
|
||||||
|
[dev-dependencies]
|
||||||
|
pretty_assertions.workspace = true
|
||||||
|
tempfile.workspace = true
|
||||||
18
harmony_secret/src/config.rs
Normal file
18
harmony_secret/src/config.rs
Normal file
@@ -0,0 +1,18 @@
|
|||||||
|
use lazy_static::lazy_static;
|
||||||
|
|
||||||
|
lazy_static! {
|
||||||
|
pub static ref SECRET_NAMESPACE: String =
|
||||||
|
std::env::var("HARMONY_SECRET_NAMESPACE").expect("HARMONY_SECRET_NAMESPACE environment variable is required, it should contain the name of the project you are working on to access its secrets");
|
||||||
|
pub static ref SECRET_STORE: Option<String> =
|
||||||
|
std::env::var("HARMONY_SECRET_STORE").ok();
|
||||||
|
pub static ref INFISICAL_URL: Option<String> =
|
||||||
|
std::env::var("HARMONY_SECRET_INFISICAL_URL").ok();
|
||||||
|
pub static ref INFISICAL_PROJECT_ID: Option<String> =
|
||||||
|
std::env::var("HARMONY_SECRET_INFISICAL_PROJECT_ID").ok();
|
||||||
|
pub static ref INFISICAL_ENVIRONMENT: Option<String> =
|
||||||
|
std::env::var("HARMONY_SECRET_INFISICAL_ENVIRONMENT").ok();
|
||||||
|
pub static ref INFISICAL_CLIENT_ID: Option<String> =
|
||||||
|
std::env::var("HARMONY_SECRET_INFISICAL_CLIENT_ID").ok();
|
||||||
|
pub static ref INFISICAL_CLIENT_SECRET: Option<String> =
|
||||||
|
std::env::var("HARMONY_SECRET_INFISICAL_CLIENT_SECRET").ok();
|
||||||
|
}
|
||||||
167
harmony_secret/src/lib.rs
Normal file
167
harmony_secret/src/lib.rs
Normal file
@@ -0,0 +1,167 @@
|
|||||||
|
pub mod config;
|
||||||
|
mod store;
|
||||||
|
|
||||||
|
use crate::config::SECRET_NAMESPACE;
|
||||||
|
use async_trait::async_trait;
|
||||||
|
use config::INFISICAL_CLIENT_ID;
|
||||||
|
use config::INFISICAL_CLIENT_SECRET;
|
||||||
|
use config::INFISICAL_ENVIRONMENT;
|
||||||
|
use config::INFISICAL_PROJECT_ID;
|
||||||
|
use config::INFISICAL_URL;
|
||||||
|
use config::SECRET_STORE;
|
||||||
|
use serde::{Serialize, de::DeserializeOwned};
|
||||||
|
use std::fmt;
|
||||||
|
use store::InfisicalSecretStore;
|
||||||
|
use store::LocalFileSecretStore;
|
||||||
|
use thiserror::Error;
|
||||||
|
use tokio::sync::OnceCell;
|
||||||
|
|
||||||
|
pub use harmony_secret_derive::Secret;
|
||||||
|
|
||||||
|
// The Secret trait remains the same.
|
||||||
|
pub trait Secret: Serialize + DeserializeOwned + Sized {
|
||||||
|
const KEY: &'static str;
|
||||||
|
}
|
||||||
|
|
||||||
|
// The error enum remains the same.
|
||||||
|
#[derive(Debug, Error)]
|
||||||
|
pub enum SecretStoreError {
|
||||||
|
#[error("Secret not found for key '{key}' in namespace '{namespace}'")]
|
||||||
|
NotFound { namespace: String, key: String },
|
||||||
|
#[error("Failed to deserialize secret for key '{key}': {source}")]
|
||||||
|
Deserialization {
|
||||||
|
key: String,
|
||||||
|
source: serde_json::Error,
|
||||||
|
},
|
||||||
|
#[error("Failed to serialize secret for key '{key}': {source}")]
|
||||||
|
Serialization {
|
||||||
|
key: String,
|
||||||
|
source: serde_json::Error,
|
||||||
|
},
|
||||||
|
#[error("Underlying storage error: {0}")]
|
||||||
|
Store(#[from] Box<dyn std::error::Error + Send + Sync>),
|
||||||
|
}
|
||||||
|
|
||||||
|
// The trait is now async!
|
||||||
|
#[async_trait]
|
||||||
|
pub trait SecretStore: fmt::Debug + Send + Sync {
|
||||||
|
async fn get_raw(&self, namespace: &str, key: &str) -> Result<Vec<u8>, SecretStoreError>;
|
||||||
|
async fn set_raw(
|
||||||
|
&self,
|
||||||
|
namespace: &str,
|
||||||
|
key: &str,
|
||||||
|
value: &[u8],
|
||||||
|
) -> Result<(), SecretStoreError>;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Use OnceCell for async-friendly, one-time initialization.
|
||||||
|
static SECRET_MANAGER: OnceCell<SecretManager> = OnceCell::const_new();
|
||||||
|
|
||||||
|
/// Initializes and returns a reference to the global SecretManager.
|
||||||
|
async fn get_secret_manager() -> &'static SecretManager {
|
||||||
|
SECRET_MANAGER.get_or_init(init_secret_manager).await
|
||||||
|
}
|
||||||
|
|
||||||
|
/// The async initialization function for the SecretManager.
|
||||||
|
async fn init_secret_manager() -> SecretManager {
|
||||||
|
let default_secret_score = "infisical".to_string();
|
||||||
|
let store_type = SECRET_STORE.as_ref().unwrap_or(&default_secret_score);
|
||||||
|
|
||||||
|
let store: Box<dyn SecretStore> = match store_type.as_str() {
|
||||||
|
"file" => Box::new(LocalFileSecretStore::default()),
|
||||||
|
"infisical" | _ => {
|
||||||
|
let store = InfisicalSecretStore::new(
|
||||||
|
INFISICAL_URL.clone().expect("Infisical url must be set, see harmony_secret config for ways to provide it. You can try with HARMONY_SECRET_INFISICAL_URL"),
|
||||||
|
INFISICAL_PROJECT_ID.clone().expect("Infisical project id must be set, see harmony_secret config for ways to provide it. You can try with HARMONY_SECRET_INFISICAL_PROJECT_ID"),
|
||||||
|
INFISICAL_ENVIRONMENT.clone().expect("Infisical environment must be set, see harmony_secret config for ways to provide it. You can try with HARMONY_SECRET_INFISICAL_ENVIRONMENT"),
|
||||||
|
INFISICAL_CLIENT_ID.clone().expect("Infisical client id must be set, see harmony_secret config for ways to provide it. You can try with HARMONY_SECRET_INFISICAL_CLIENT_ID"),
|
||||||
|
INFISICAL_CLIENT_SECRET.clone().expect("Infisical client secret must be set, see harmony_secret config for ways to provide it. You can try with HARMONY_SECRET_INFISICAL_CLIENT_SECRET"),
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
.expect("Failed to initialize Infisical secret store");
|
||||||
|
Box::new(store)
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
SecretManager::new(SECRET_NAMESPACE.clone(), store)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Manages the lifecycle of secrets, providing a simple static API.
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub struct SecretManager {
|
||||||
|
namespace: String,
|
||||||
|
store: Box<dyn SecretStore>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl SecretManager {
|
||||||
|
fn new(namespace: String, store: Box<dyn SecretStore>) -> Self {
|
||||||
|
Self { namespace, store }
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Retrieves and deserializes a secret.
|
||||||
|
pub async fn get<T: Secret>() -> Result<T, SecretStoreError> {
|
||||||
|
let manager = get_secret_manager().await;
|
||||||
|
let raw_value = manager.store.get_raw(&manager.namespace, T::KEY).await?;
|
||||||
|
serde_json::from_slice(&raw_value).map_err(|e| SecretStoreError::Deserialization {
|
||||||
|
key: T::KEY.to_string(),
|
||||||
|
source: e,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Serializes and stores a secret.
|
||||||
|
pub async fn set<T: Secret>(secret: &T) -> Result<(), SecretStoreError> {
|
||||||
|
let manager = get_secret_manager().await;
|
||||||
|
let raw_value =
|
||||||
|
serde_json::to_vec(secret).map_err(|e| SecretStoreError::Serialization {
|
||||||
|
key: T::KEY.to_string(),
|
||||||
|
source: e,
|
||||||
|
})?;
|
||||||
|
manager
|
||||||
|
.store
|
||||||
|
.set_raw(&manager.namespace, T::KEY, &raw_value)
|
||||||
|
.await
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod test {
|
||||||
|
use super::*;
|
||||||
|
#[cfg(secrete2etest)]
|
||||||
|
use pretty_assertions::assert_eq;
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug, PartialEq)]
|
||||||
|
struct TestUserMeta {
|
||||||
|
labels: Vec<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Secret, Serialize, Deserialize, Debug, PartialEq)]
|
||||||
|
struct TestSecret {
|
||||||
|
user: String,
|
||||||
|
password: String,
|
||||||
|
metadata: TestUserMeta,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(secrete2etest)]
|
||||||
|
#[tokio::test]
|
||||||
|
async fn set_and_retrieve_secret() {
|
||||||
|
let secret = TestSecret {
|
||||||
|
user: String::from("user"),
|
||||||
|
password: String::from("password"),
|
||||||
|
metadata: TestUserMeta {
|
||||||
|
labels: vec![
|
||||||
|
String::from("label1"),
|
||||||
|
String::from("label2"),
|
||||||
|
String::from(
|
||||||
|
"some longet label with \" special @#%$)(udiojcia[]]] \"'asdij'' characters Nдs はにほへとちり าฟันพัฒนา yağız şoföre ç <20> <20> <20> <20> <20> <20> <20> <20> <20> <20> <20> <20> <20> 👩👩👧👦 /span> 👩👧👦 and why not emojis ",
|
||||||
|
),
|
||||||
|
],
|
||||||
|
},
|
||||||
|
};
|
||||||
|
|
||||||
|
SecretManager::set(&secret).await.unwrap();
|
||||||
|
let value = SecretManager::get::<TestSecret>().await.unwrap();
|
||||||
|
|
||||||
|
assert_eq!(value, secret);
|
||||||
|
}
|
||||||
|
}
|
||||||
129
harmony_secret/src/store/infisical.rs
Normal file
129
harmony_secret/src/store/infisical.rs
Normal file
@@ -0,0 +1,129 @@
|
|||||||
|
use crate::{SecretStore, SecretStoreError};
|
||||||
|
use async_trait::async_trait;
|
||||||
|
use infisical::{
|
||||||
|
AuthMethod, InfisicalError,
|
||||||
|
client::Client,
|
||||||
|
secrets::{CreateSecretRequest, GetSecretRequest, UpdateSecretRequest},
|
||||||
|
};
|
||||||
|
use log::{info, warn};
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub struct InfisicalSecretStore {
|
||||||
|
client: Client,
|
||||||
|
project_id: String,
|
||||||
|
environment: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl InfisicalSecretStore {
|
||||||
|
/// Creates a new, authenticated Infisical client.
|
||||||
|
pub async fn new(
|
||||||
|
base_url: String,
|
||||||
|
project_id: String,
|
||||||
|
environment: String,
|
||||||
|
client_id: String,
|
||||||
|
client_secret: String,
|
||||||
|
) -> Result<Self, InfisicalError> {
|
||||||
|
info!("INFISICAL_STORE: Initializing client for URL: {base_url}");
|
||||||
|
|
||||||
|
// The builder and login logic remains the same.
|
||||||
|
let mut client = Client::builder().base_url(base_url).build().await?;
|
||||||
|
let auth_method = AuthMethod::new_universal_auth(client_id, client_secret);
|
||||||
|
client.login(auth_method).await?;
|
||||||
|
|
||||||
|
info!("INFISICAL_STORE: Client authenticated successfully.");
|
||||||
|
Ok(Self {
|
||||||
|
client,
|
||||||
|
project_id,
|
||||||
|
environment,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[async_trait]
|
||||||
|
impl SecretStore for InfisicalSecretStore {
|
||||||
|
async fn get_raw(&self, _environment: &str, key: &str) -> Result<Vec<u8>, SecretStoreError> {
|
||||||
|
let environment = &self.environment;
|
||||||
|
info!("INFISICAL_STORE: Getting key '{key}' from environment '{environment}'");
|
||||||
|
|
||||||
|
let request = GetSecretRequest::builder(key, &self.project_id, environment).build();
|
||||||
|
|
||||||
|
match self.client.secrets().get(request).await {
|
||||||
|
Ok(secret) => Ok(secret.secret_value.into_bytes()),
|
||||||
|
Err(e) => {
|
||||||
|
// Correctly match against the actual InfisicalError enum.
|
||||||
|
match e {
|
||||||
|
// The specific case for a 404 Not Found error.
|
||||||
|
InfisicalError::HttpError { status, .. }
|
||||||
|
if status == http::StatusCode::NOT_FOUND =>
|
||||||
|
{
|
||||||
|
Err(SecretStoreError::NotFound {
|
||||||
|
namespace: environment.to_string(),
|
||||||
|
key: key.to_string(),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
// For all other errors, wrap them in our generic Store error.
|
||||||
|
_ => Err(SecretStoreError::Store(Box::new(e))),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn set_raw(
|
||||||
|
&self,
|
||||||
|
_environment: &str,
|
||||||
|
key: &str,
|
||||||
|
val: &[u8],
|
||||||
|
) -> Result<(), SecretStoreError> {
|
||||||
|
info!(
|
||||||
|
"INFISICAL_STORE: Setting key '{key}' in environment '{}'",
|
||||||
|
self.environment
|
||||||
|
);
|
||||||
|
let value_str =
|
||||||
|
String::from_utf8(val.to_vec()).map_err(|e| SecretStoreError::Store(Box::new(e)))?;
|
||||||
|
|
||||||
|
// --- Upsert Logic ---
|
||||||
|
// First, attempt to update the secret.
|
||||||
|
let update_req = UpdateSecretRequest::builder(key, &self.project_id, &self.environment)
|
||||||
|
.secret_value(&value_str)
|
||||||
|
.build();
|
||||||
|
|
||||||
|
match self.client.secrets().update(update_req).await {
|
||||||
|
Ok(_) => {
|
||||||
|
info!("INFISICAL_STORE: Successfully updated secret '{key}'.");
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
Err(e) => {
|
||||||
|
// If the update failed, check if it was because the secret doesn't exist.
|
||||||
|
match e {
|
||||||
|
InfisicalError::HttpError { status, .. }
|
||||||
|
if status == http::StatusCode::NOT_FOUND =>
|
||||||
|
{
|
||||||
|
// The secret was not found, so we create it instead.
|
||||||
|
warn!(
|
||||||
|
"INFISICAL_STORE: Secret '{key}' not found for update, attempting to create it."
|
||||||
|
);
|
||||||
|
let create_req = CreateSecretRequest::builder(
|
||||||
|
key,
|
||||||
|
&value_str,
|
||||||
|
&self.project_id,
|
||||||
|
&self.environment,
|
||||||
|
)
|
||||||
|
.build();
|
||||||
|
|
||||||
|
// Handle potential errors during creation.
|
||||||
|
self.client
|
||||||
|
.secrets()
|
||||||
|
.create(create_req)
|
||||||
|
.await
|
||||||
|
.map_err(|create_err| SecretStoreError::Store(Box::new(create_err)))?;
|
||||||
|
|
||||||
|
info!("INFISICAL_STORE: Successfully created secret '{key}'.");
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
// Any other error during update is a genuine failure.
|
||||||
|
_ => Err(SecretStoreError::Store(Box::new(e))),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
104
harmony_secret/src/store/local_file.rs
Normal file
104
harmony_secret/src/store/local_file.rs
Normal file
@@ -0,0 +1,104 @@
|
|||||||
|
use async_trait::async_trait;
|
||||||
|
use log::info;
|
||||||
|
use std::path::{Path, PathBuf};
|
||||||
|
|
||||||
|
use crate::{SecretStore, SecretStoreError};
|
||||||
|
|
||||||
|
#[derive(Debug, Default)]
|
||||||
|
pub struct LocalFileSecretStore;
|
||||||
|
|
||||||
|
impl LocalFileSecretStore {
|
||||||
|
/// Helper to consistently generate the secret file path.
|
||||||
|
fn get_file_path(base_dir: &Path, ns: &str, key: &str) -> PathBuf {
|
||||||
|
base_dir.join(format!("{ns}_{key}.json"))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[async_trait]
|
||||||
|
impl SecretStore for LocalFileSecretStore {
|
||||||
|
async fn get_raw(&self, ns: &str, key: &str) -> Result<Vec<u8>, SecretStoreError> {
|
||||||
|
let data_dir = directories::BaseDirs::new()
|
||||||
|
.expect("Could not find a valid home directory")
|
||||||
|
.data_dir()
|
||||||
|
.join("harmony")
|
||||||
|
.join("secrets");
|
||||||
|
|
||||||
|
let file_path = Self::get_file_path(&data_dir, ns, key);
|
||||||
|
info!(
|
||||||
|
"LOCAL_STORE: Getting key '{key}' from namespace '{ns}' at {}",
|
||||||
|
file_path.display()
|
||||||
|
);
|
||||||
|
|
||||||
|
tokio::fs::read(&file_path)
|
||||||
|
.await
|
||||||
|
.map_err(|_| SecretStoreError::NotFound {
|
||||||
|
namespace: ns.to_string(),
|
||||||
|
key: key.to_string(),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn set_raw(&self, ns: &str, key: &str, val: &[u8]) -> Result<(), SecretStoreError> {
|
||||||
|
let data_dir = directories::BaseDirs::new()
|
||||||
|
.expect("Could not find a valid home directory")
|
||||||
|
.data_dir()
|
||||||
|
.join("harmony")
|
||||||
|
.join("secrets");
|
||||||
|
|
||||||
|
let file_path = Self::get_file_path(&data_dir, ns, key);
|
||||||
|
info!(
|
||||||
|
"LOCAL_STORE: Setting key '{key}' in namespace '{ns}' at {}",
|
||||||
|
file_path.display()
|
||||||
|
);
|
||||||
|
|
||||||
|
if let Some(parent_dir) = file_path.parent() {
|
||||||
|
tokio::fs::create_dir_all(parent_dir)
|
||||||
|
.await
|
||||||
|
.map_err(|e| SecretStoreError::Store(Box::new(e)))?;
|
||||||
|
}
|
||||||
|
|
||||||
|
tokio::fs::write(&file_path, val)
|
||||||
|
.await
|
||||||
|
.map_err(|e| SecretStoreError::Store(Box::new(e)))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use super::*;
|
||||||
|
use tempfile::tempdir;
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn test_set_and_get_raw_successfully() {
|
||||||
|
let dir = tempdir().unwrap();
|
||||||
|
let ns = "test-ns";
|
||||||
|
let key = "test-key";
|
||||||
|
let value = b"{\"data\":\"test-value\"}";
|
||||||
|
|
||||||
|
// To test the store directly, we override the base directory logic.
|
||||||
|
// For this test, we'll manually construct the path within our temp dir.
|
||||||
|
let file_path = LocalFileSecretStore::get_file_path(dir.path(), ns, key);
|
||||||
|
|
||||||
|
// Manually write to the temp path to simulate the store's behavior
|
||||||
|
tokio::fs::create_dir_all(file_path.parent().unwrap())
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
tokio::fs::write(&file_path, value).await.unwrap();
|
||||||
|
|
||||||
|
// Now, test get_raw by reading from that same temp path (by mocking the path logic)
|
||||||
|
let retrieved_value = tokio::fs::read(&file_path).await.unwrap();
|
||||||
|
assert_eq!(retrieved_value, value);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn test_get_raw_not_found() {
|
||||||
|
let dir = tempdir().unwrap();
|
||||||
|
let ns = "test-ns";
|
||||||
|
let key = "non-existent-key";
|
||||||
|
|
||||||
|
// We need to check if reading a non-existent file gives the correct error
|
||||||
|
let file_path = LocalFileSecretStore::get_file_path(dir.path(), ns, key);
|
||||||
|
let result = tokio::fs::read(&file_path).await;
|
||||||
|
|
||||||
|
assert!(matches!(result, Err(_)));
|
||||||
|
}
|
||||||
|
}
|
||||||
4
harmony_secret/src/store/mod.rs
Normal file
4
harmony_secret/src/store/mod.rs
Normal file
@@ -0,0 +1,4 @@
|
|||||||
|
mod infisical;
|
||||||
|
mod local_file;
|
||||||
|
pub use infisical::*;
|
||||||
|
pub use local_file::*;
|
||||||
8
harmony_secret/test_harmony_secret_infisical.sh
Normal file
8
harmony_secret/test_harmony_secret_infisical.sh
Normal file
@@ -0,0 +1,8 @@
|
|||||||
|
export HARMONY_SECRET_NAMESPACE=harmony_test_secrets
|
||||||
|
export HARMONY_SECRET_INFISICAL_URL=http://localhost
|
||||||
|
export HARMONY_SECRET_INFISICAL_PROJECT_ID=eb4723dc-eede-44d7-98cc-c8e0caf29ccb
|
||||||
|
export HARMONY_SECRET_INFISICAL_ENVIRONMENT=dev
|
||||||
|
export HARMONY_SECRET_INFISICAL_CLIENT_ID=dd16b07f-0e38-4090-a1d0-922de9f44d91
|
||||||
|
export HARMONY_SECRET_INFISICAL_CLIENT_SECRET=bd2ae054e7759b11ca2e908494196337cc800bab138cb1f59e8d9b15ca3f286f
|
||||||
|
|
||||||
|
cargo test
|
||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user