fix: unjank the demo #85
@ -1,6 +1,5 @@
|
||||
use std::{path::PathBuf, sync::Arc};
|
||||
|
||||
use base64::{Engine as _, engine::general_purpose};
|
||||
use harmony::{
|
||||
data::Id,
|
||||
inventory::Inventory,
|
||||
@ -10,17 +9,11 @@ use harmony::{
|
||||
ApplicationScore, RustWebFramework, RustWebapp,
|
||||
features::{ContinuousDelivery, Monitoring},
|
||||
},
|
||||
monitoring::{
|
||||
alert_channel::webhook_receiver::WebhookReceiver,
|
||||
kube_prometheus::helm_prometheus_alert_score::HelmPrometheusAlertingScore,
|
||||
ntfy::ntfy::NtfyScore,
|
||||
},
|
||||
tenant::TenantScore,
|
||||
},
|
||||
score::Score,
|
||||
topology::{
|
||||
K8sAnywhereTopology, Url,
|
||||
tenant::{ResourceLimits, TenantConfig, TenantManager, TenantNetworkPolicy},
|
||||
tenant::{ResourceLimits, TenantConfig, TenantNetworkPolicy},
|
||||
},
|
||||
};
|
||||
|
||||
|
@ -4,7 +4,7 @@ use k8s_openapi::{
|
||||
ClusterResourceScope, NamespaceResourceScope,
|
||||
api::{
|
||||
apps::v1::Deployment,
|
||||
core::v1::{ObjectReference, Pod},
|
||||
core::v1::Pod,
|
||||
},
|
||||
};
|
||||
use kube::{
|
||||
|
@ -2,7 +2,6 @@ use log::debug;
|
||||
use serde::Serialize;
|
||||
use serde_with::skip_serializing_none;
|
||||
use serde_yaml::Value;
|
||||
use url::Url;
|
||||
|
||||
use crate::modules::application::features::CDApplicationConfig;
|
||||
|
||||
@ -257,7 +256,7 @@ spec:
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use pretty_assertions::assert_eq;
|
||||
use url::Url;
|
||||
|
||||
|
||||
use crate::modules::application::features::{
|
||||
ArgoApplication, Automated, Backoff, Helm, Retry, Source, SyncPolicy,
|
||||
|
@ -14,7 +14,7 @@ use crate::{
|
||||
features::{ArgoApplication, ArgoHelmScore},
|
||||
},
|
||||
score::Score,
|
||||
topology::{DeploymentTarget, HelmCommand, K8sclient, MultiTargetTopology, Topology, Url},
|
||||
topology::{DeploymentTarget, HelmCommand, K8sclient, MultiTargetTopology, Topology},
|
||||
};
|
||||
|
||||
/// ContinuousDelivery in Harmony provides this functionality :
|
||||
|
@ -1,5 +1,4 @@
|
||||
use async_trait::async_trait;
|
||||
use kube::api::{ApiResource, GroupVersionKind};
|
||||
use log::error;
|
||||
use non_blank_string_rs::NonBlankString;
|
||||
use serde::Serialize;
|
||||
|
@ -1,11 +1,9 @@
|
||||
use std::fs;
|
||||
use std::io::Read;
|
||||
use std::path::PathBuf;
|
||||
use std::process;
|
||||
use std::sync::Arc;
|
||||
|
||||
use async_trait::async_trait;
|
||||
use bollard::image::PushImageOptions;
|
||||
use bollard::query_parameters::PushImageOptionsBuilder;
|
||||
use bollard::{Docker, body_full};
|
||||
wjro marked this conversation as resolved
|
||||
use dockerfile_builder::Dockerfile;
|
||||
@ -15,7 +13,6 @@ use futures_util::StreamExt;
|
||||
use log::{debug, error, info};
|
||||
use serde::Serialize;
|
||||
use tar::Archive;
|
||||
use tempfile::tempfile;
|
||||
|
||||
use crate::config::{REGISTRY_PROJECT, REGISTRY_URL};
|
||||
use crate::{
|
||||
|
Loading…
Reference in New Issue
Block a user
add a comment to justify using bollard vs docker via the cli
The reason I used
bollard
was because I find it preferable to access the docker daemon programmatically over an API over using a (albeit relatively stable) CLI interface