Compare commits

...

3 commits

Author SHA1 Message Date
Qyriad
ccd37ee169 daemon tests for gotosocial 2026-03-25 19:21:37 +01:00
Qyriad
7f4a5a35ca daemon w/ tests PoC 2026-03-24 15:07:52 +01:00
Qyriad
68fc04a6d2 nix: output docs in 2026-03-23 17:38:16 +01:00
15 changed files with 670 additions and 173 deletions

10
Cargo.lock generated
View file

@ -223,6 +223,7 @@ dependencies = [
"tracing", "tracing",
"tracing-human-layer", "tracing-human-layer",
"tracing-subscriber", "tracing-subscriber",
"which",
] ]
[[package]] [[package]]
@ -851,6 +852,15 @@ version = "0.11.1+wasi-snapshot-preview1"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "ccf3ec651a847eb01de73ccad15eb7d99f80485de043efb2f370cd654f4ea44b" checksum = "ccf3ec651a847eb01de73ccad15eb7d99f80485de043efb2f370cd654f4ea44b"
[[package]]
name = "which"
version = "8.0.2"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "81995fafaaaf6ae47a7d0cc83c67caf92aeb7e5331650ae6ff856f7c0c60c459"
dependencies = [
"libc",
]
[[package]] [[package]]
name = "windows" name = "windows"
version = "0.61.3" version = "0.61.3"

View file

@ -47,6 +47,7 @@ tap = "1.0.1"
tracing = { version = "0.1.44", features = ["attributes"] } tracing = { version = "0.1.44", features = ["attributes"] }
tracing-human-layer = "0.2.1" tracing-human-layer = "0.2.1"
tracing-subscriber = { version = "0.3.22", default-features = false, features = ["std", "env-filter", "fmt", "ansi", "registry", "parking_lot"] } tracing-subscriber = { version = "0.3.22", default-features = false, features = ["std", "env-filter", "fmt", "ansi", "registry", "parking_lot"] }
which = "8.0.2"
[profile.dev] [profile.dev]
opt-level = 1 opt-level = 1
@ -63,3 +64,21 @@ opt-level = 1
[profile.release.package."*"] [profile.release.package."*"]
debug = true debug = true
debug-assertions = true debug-assertions = true
[lints.clippy]
#arithmetic_side_effects = "warn"
as_ptr_cast_mut = "warn"
assigning_clones = "warn"
borrow_as_ptr = "warn"
#cargo_common_metadata = "warn"
cast_lossless = "warn"
#cast_possible_truncation = "warn"
cast_possible_wrap = "warn"
cast_ptr_alignment = "warn"
cast_sign_loss = "warn"
clear_with_drain = "warn"
coerce_container_to_any = "warn"
derive_partial_eq_without_eq = "warn"
doc_broken_link = "warn"
doc_comment_double_space_linebreaks = "warn"
doc_markdown = "warn"

View file

@ -9,18 +9,28 @@
in import src { inherit pkgs; }, in import src { inherit pkgs; },
}: let }: let
inherit (qpkgs) lib; inherit (qpkgs) lib;
dynix = (qpkgs.callPackage ./package.nix { })
# Use LLD for faster link times.
defaultStdenv = pkgs.clangStdenv.override {
cc = pkgs.clangStdenv.cc.override {
bintools = pkgs.wrapBintoolsWith { inherit (pkgs.llvmPackages) bintools; };
};
};
dynix = (qpkgs.callPackage ./package.nix { clangStdenv = defaultStdenv; })
.overrideAttrs (final: prev: { .overrideAttrs (final: prev: {
dynixCommand = qpkgs.stdlib.mkStdenvPretty prev.dynixCommand; dynixCommand = qpkgs.stdlib.mkStdenvPretty prev.dynixCommand;
dynixModules = qpkgs.stdlib.mkStdenvPretty prev.dynixModules; dynixModules = qpkgs.stdlib.mkStdenvPretty prev.dynixModules;
}) })
|> qpkgs.stdlib.mkStdenvPretty; |> qpkgs.stdlib.mkStdenvPretty;
byStdenv = lib.mapAttrs (stdenvName: stdenv: let byStdenv = lib.mapAttrs (stdenvName: stdenv: let
withStdenv = dynix.override { clangStdenv = stdenv; }; withStdenv = dynix.override { clangStdenv = stdenv; };
dynix' = withStdenv.overrideAttrs (prev: { dynix' = withStdenv.overrideAttrs (prev: {
pname = "${prev.pname}-${stdenvName}"; pname = "${prev.pname}-${stdenvName}";
}); });
in dynix') qpkgs.validStdenvs; in dynix') qpkgs.validStdenvs;
in dynix.overrideAttrs (prev: lib.recursiveUpdate prev { in dynix.overrideAttrs (prev: lib.recursiveUpdate prev {
passthru = { inherit byStdenv; }; passthru = { inherit byStdenv; };
}) })

View file

@ -7,23 +7,15 @@
clangStdenv, clangStdenv,
callPackage, callPackage,
linkFarm, linkFarm,
llvmPackages,
rustHooks, rustHooks,
rustPackages, rustPackages,
versionCheckHook, versionCheckHook,
wrapBintoolsWith, writeScript,
}: lib.callWith' rustPackages ({ }: lib.callWith' rustPackages ({
rustPlatform, rustPlatform,
cargo, cargo,
}: let }: let
# Use LLD for faster link times. stdenv = clangStdenv;
stdenv = clangStdenv.override {
cc = clangStdenv.cc.override {
bintools = wrapBintoolsWith {
bintools = llvmPackages.bintools;
};
};
};
cargoToml = lib.importTOML ./Cargo.toml; cargoToml = lib.importTOML ./Cargo.toml;
cargoPackage = cargoToml.package; cargoPackage = cargoToml.package;
in stdenv.mkDerivation (finalAttrs: let in stdenv.mkDerivation (finalAttrs: let
@ -46,18 +38,37 @@ in {
cp -r --reflink=auto "$dynixCommand/"* "$out/" cp -r --reflink=auto "$dynixCommand/"* "$out/"
mkdir -p "$modules" mkdir -p "$modules"
cp -r --reflink=auto "$dynixModules/"* "$modules/" cp -r --reflink=auto "$dynixModules/"* "$modules/"
install -Dm a=rx "$dynixTestingClient" "$out/libexec/dynix-testing-client.py"
''; '';
# #
# SUB-DERIVATONS # SUB-DERIVATONS
# #
dynixTestingClient = writeScript "dynix-testing-client.py" ''
#!/usr/bin/env python3
import socket, sys, os, json
try:
sockpath = sys.argv[1]
except IndexError:
sockpath = f"{os.environ['XDG_RUNTIME_DIR']}/dynix.sock"
sock = socket.socket(family=socket.AF_UNIX)
sock.connect(sockpath)
sock.sendall(sys.stdin.buffer.read())
sock.settimeout(20)
reply = json.loads(sock.recv(256).decode("utf-8"))
print(json.dumps(reply, indent=2))
sys.exit(reply["status"])
'';
dynixCommand = stdenv.mkDerivation { dynixCommand = stdenv.mkDerivation {
pname = "${self.pname}-command"; pname = "${self.pname}-command";
inherit (self) version; inherit (self) version;
inherit (self) strictDeps __structuredAttrs; inherit (self) strictDeps __structuredAttrs;
inherit (self) doCheck doInstallCheck; inherit (self) doCheck doInstallCheck;
outputs = [ "out" "doc" ];
src = lib.fileset.toSource { src = lib.fileset.toSource {
root = ./.; root = ./.;
fileset = lib.fileset.unions [ fileset = lib.fileset.unions [
@ -71,6 +82,12 @@ in {
lockFile = ./Cargo.lock; lockFile = ./Cargo.lock;
}; };
postInstall = ''
cargo doc --document-private-items
mkdir -p "$doc"
cp -r ./target/doc/* "$doc/"
'';
nativeBuildInputs = rustHooks.asList ++ [ nativeBuildInputs = rustHooks.asList ++ [
cargo cargo
]; ];

View file

@ -4,3 +4,13 @@
match_block_trailing_comma = true match_block_trailing_comma = true
merge_derives = false merge_derives = false
# Unstable options.
blank_lines_upper_bound = 3
format_code_in_doc_comments = true
format_macro_matchers = true
# When structs, slices, arrays, and block/array-like macros are used as the last argument in an expression list,
# allow them to overflow (like blocks/closures) instead of being indented on a new line.
overflow_delimited_expr = true
# Put `type` and `const` items before methods.
reorder_impl_items = true

View file

@ -16,7 +16,21 @@
fenixLib ? let fenixLib ? let
src = fetchTarball "https://github.com/nix-community/fenix/archive/main.tar.gz"; src = fetchTarball "https://github.com/nix-community/fenix/archive/main.tar.gz";
in import src { inherit pkgs; }, in import src { inherit pkgs; },
fenixToolchain ? fenixLib.latest.toolchain, fenixBaseToolchain ? fenixLib.stable.withComponents [
"cargo"
"rustc"
"llvm-tools"
"rust-std"
"rust-docs"
"rust-src"
"rustc-dev"
"clippy"
],
fenixToolchain ? fenixLib.combine [
fenixBaseToolchain
# Rustfmt is very handy to have as nightly.
fenixLib.latest.rustfmt
],
}: let }: let
inherit (pkgs) lib; inherit (pkgs) lib;

View file

@ -1,6 +1,7 @@
use std::{ use std::{
env, io, env, io,
os::fd::{AsFd, BorrowedFd, IntoRawFd, OwnedFd, RawFd}, os::fd::{AsFd, BorrowedFd, IntoRawFd, OwnedFd, RawFd},
process::{Command, Stdio},
sync::{ sync::{
Arc, LazyLock, Arc, LazyLock,
atomic::{AtomicUsize, Ordering}, atomic::{AtomicUsize, Ordering},
@ -12,7 +13,20 @@ use iddqd::{BiHashMap, IdOrdMap};
use mio::{Events, Interest, Poll, Token, event::Event, net::UnixListener, unix::SourceFd}; use mio::{Events, Interest, Poll, Token, event::Event, net::UnixListener, unix::SourceFd};
use rustix::{buffer::spare_capacity, net::SocketFlags, process::Uid}; use rustix::{
buffer::spare_capacity,
net::SocketFlags,
process::{Pid, PidfdFlags, Uid, WaitId, WaitIdOptions},
};
mod rustix {
pub use rustix::process::{getuid, pidfd_open, waitid};
pub use rustix::*;
}
//mod rustix_prelude {
// pub use rustix::process::{getuid, pidfd_open, waitid};
//}
use serde_json::StreamDeserializer; use serde_json::StreamDeserializer;
@ -21,10 +35,7 @@ use crate::prelude::*;
pub mod api; pub mod api;
use api::DaemonCmd; use api::DaemonCmd;
use crate::{ use crate::daemon_tokfd::{FdInfo, FdKind};
SourceFile, SourceLine,
daemon_tokfd::{FdInfo, FdKind},
};
use crate::{OwnedFdWithFlags, TokenFd}; use crate::{OwnedFdWithFlags, TokenFd};
@ -45,6 +56,22 @@ pub static TMPDIR: LazyLock<&'static Path> = LazyLock::new(|| {
Box::leak(dir) Box::leak(dir)
}); });
pub static NIXOS_REBUILD: LazyLock<&'static Path> = LazyLock::new(|| {
which::which("nixos-rebuild")
.inspect_err(|e| error!("couldn't find `nixos-rebuild` in PATH: {e}"))
.map(PathBuf::into_boxed_path)
.map(|boxed| &*Box::leak(boxed))
.unwrap_or(Path::new("/run/current-system/sw/bin/nixos-rebuild"))
});
pub static NIX: LazyLock<&'static Path> = LazyLock::new(|| {
which::which("nix")
.inspect_err(|e| error!("couldn't find `nix` in PATH: {e}"))
.map(PathBuf::into_boxed_path)
.map(|boxed| &*Box::leak(boxed))
.unwrap_or(Path::new("/run/current-system/sw/bin/nix"))
});
const TIMEOUT_NEVER: Option<Duration> = None; const TIMEOUT_NEVER: Option<Duration> = None;
static NEXT_TOKEN_NUMBER: AtomicUsize = AtomicUsize::new(1); static NEXT_TOKEN_NUMBER: AtomicUsize = AtomicUsize::new(1);
@ -60,6 +87,42 @@ fn next_token() -> Token {
Token(tok) Token(tok)
} }
trait EventExt {
type Display;
fn display(&self) -> Self::Display;
}
#[derive(Copy)]
#[derive(Debug, Clone, PartialEq, Eq, PartialOrd, Ord, Hash)]
struct EventDisplay {
token: Token,
error: bool,
writable: bool,
write_closed: bool,
readable: bool,
read_closed: bool,
}
impl EventExt for Event {
type Display = EventDisplay;
fn display(&self) -> Self::Display {
EventDisplay {
token: self.token(),
error: self.is_error(),
writable: self.is_writable(),
write_closed: self.is_write_closed(),
readable: self.is_readable(),
read_closed: self.is_read_closed(),
}
}
}
impl Display for EventDisplay {
fn fmt(&self, f: &mut Formatter) -> FmtResult {
todo!()
}
}
#[derive(Debug)] #[derive(Debug)]
pub struct Daemon { pub struct Daemon {
config_path: Arc<Path>, config_path: Arc<Path>,
@ -132,6 +195,33 @@ impl Daemon {
token token
} }
#[expect(dead_code)]
fn register_with_name<S>(&mut self, fd: RawFd, kind: FdKind, name: Box<OsStr>) -> Token {
let token = next_token();
debug!(
"Registering new {} FdInfo for {fd} ({}) with token {token:?}",
name.to_string_lossy(),
kind.name_str(),
);
self.fd_info
.insert_unique(FdInfo::new_with_name(fd, kind, name))
.unwrap();
self.tokfd
.insert_unique(TokenFd { token, fd })
.unwrap_or_else(|e| todo!("{e}"));
let mut source = SourceFd(&fd);
self.poller
.registry()
.register(&mut source, token, Interest::READABLE)
.unwrap_or_else(|e| unreachable!("registering {fd:?} with poller failed: {e}"));
token
}
fn deregister(&mut self, fd: RawFd) { fn deregister(&mut self, fd: RawFd) {
let info = self let info = self
.fd_info .fd_info
@ -279,6 +369,32 @@ const DAEMON: Token = Token(0);
/// Private helpers. /// Private helpers.
impl Daemon { impl Daemon {
fn proxy_stdio(&mut self, fd: &BorrowedFd) -> Result<(), IoError> {
let info = self.fd_info.get(&fd.as_raw_fd()).unwrap();
let label = match info.kind {
FdKind::ChildStdout => "stdout",
FdKind::ChildStderr => "stderr",
other => unreachable!("child stdio cannot have kind {other:?}"),
};
// FIXME: don't use a new allocation every time.
let mut buffer: Vec<u8> = Vec::with_capacity(1024);
// FIXME: handle line buffering correctly.
loop {
let count = rustix::io::read(fd, spare_capacity(&mut buffer))
.inspect_err(|e| error!("read() on child stdio fd {fd:?} failed: {e}"))?;
if count == 0 {
break;
}
for line in buffer.lines() {
info!("[child {label}]: {}", line.as_bstr())
}
}
Ok(())
}
fn read_cmd(&mut self, fd: &BorrowedFd) -> Result<(), IoError> { fn read_cmd(&mut self, fd: &BorrowedFd) -> Result<(), IoError> {
// FIXME: don't use a new allocation every time. // FIXME: don't use a new allocation every time.
let mut cmd_buffer: Vec<u8> = Vec::with_capacity(1024); let mut cmd_buffer: Vec<u8> = Vec::with_capacity(1024);
@ -317,42 +433,82 @@ impl Daemon {
} }
fn dispatch_cmd(&mut self, cmd: DaemonCmd) -> Result<(), IoError> { fn dispatch_cmd(&mut self, cmd: DaemonCmd) -> Result<(), IoError> {
// Write the new file...
let (name, value) = match cmd { let (name, value) = match cmd {
DaemonCmd::Append { name, value } => (name, value), DaemonCmd::Append { name, value } => (name, value),
}; };
let mut opts = File::options(); let source_file = crate::open_source_file(self.config_path.clone())?;
opts.read(true)
.write(true)
.create(false)
.custom_flags(libc::O_CLOEXEC);
let source_file = SourceFile::open_from(self.config_path.clone(), opts)?;
let pri = crate::get_where(source_file.clone()).unwrap_or_else(|e| todo!("{e}")); let pri = crate::get_where(source_file.clone()).unwrap_or_else(|e| todo!("{e}"));
let new_pri = pri - 1; let new_pri = pri - 1;
//let new_pri_line =
// crate::get_next_prio_line(source_file.clone(), Arc::from(name), Arc::from(value));
// Get next priority line. // Get next priority line.
let source_lines = source_file.lines()?; let opt_name = name.to_nix_decl();
let penultimate = source_lines.get(source_lines.len() - 2); let new_line = crate::get_next_prio_line(
// FIXME: don't rely on whitespace lol source_file.clone(),
debug_assert_eq!(penultimate.map(SourceLine::text).as_deref(), Some(" ];")); &opt_name,
let penultimate = penultimate.unwrap();
let new_generation = 0 - new_pri;
let new_line = SourceLine {
line: penultimate.line,
path: source_file.path(),
text: Arc::from(format!(
" {} = lib.mkOverride ({}) ({}); # DYNIX GENERATION {}",
name.to_nix_decl(),
new_pri, new_pri,
value, &value.to_nix_source(),
new_generation, )
)), .unwrap_or_else(|e| panic!("someone is holding a reference to source.lines(): {e}"));
};
drop(source_lines);
crate::write_next_prio(source_file, new_line).unwrap_or_else(|e| todo!("{e}")); crate::write_next_prio(source_file, new_line).unwrap_or_else(|e| todo!("{e}"));
// Rebuild and switch.
// FIXME: allow passing additional args.
//let child = Command::new(*NIXOS_REBUILD)
// .arg("switch")
// .arg("--log-format")
// .arg("raw-with-logs")
// .arg("--no-reexec")
// .arg("-v")
// .stdout(Stdio::piped())
// .stderr(Stdio::piped())
// .spawn()
// .inspect_err(|e| {
// error!("failed to spawn `nixos-rebuild` command: {e}");
// })?;
let expr = "(import <nixpkgs/nixos> { }).config.dynamicism.applyDynamicConfiguration { }";
let child = Command::new(*NIX)
.arg("run")
.arg("--show-trace")
.arg("--log-format")
.arg("raw-with-logs")
.arg("--impure")
.arg("-E")
.arg(expr)
.stdout(Stdio::piped())
.stderr(Stdio::piped())
.spawn()
.inspect_err(|e| error!("failed to spawn `nix run` command: {e}"))?;
debug!("Spanwed child process {}", child.id());
let pid = Pid::from_child(&child);
let stdout = child.stdout.unwrap_or_else(|| {
unreachable!("`child` is given `.stdout(Stdio::piped())`");
});
let stderr = child.stderr.unwrap_or_else(|| {
unreachable!("`child` is given `.stderr(Stdio::piped())`");
});
let _token = self.register(stdout.into_raw_fd(), FdKind::ChildStdout);
let _token = self.register(stderr.into_raw_fd(), FdKind::ChildStderr);
match rustix::process::pidfd_open(pid, PidfdFlags::NONBLOCK) {
Ok(pidfd) => {
debug!("Opened pidfd {pidfd:?}, for process {pid}");
self.register(pidfd.into_raw_fd(), FdKind::Pid(pid));
},
Err(e) if e.kind() == IoErrorKind::NotFound => {
warn!("child {pid} not found; died before we could open it?");
},
Err(e) => {
error!("Error opening pidfd for child {pid}: {e}");
return Err(e)?;
},
}
Ok(()) Ok(())
} }
@ -386,9 +542,9 @@ impl Daemon {
loop { loop {
if tracing::enabled!(tracing::Level::DEBUG) { if tracing::enabled!(tracing::Level::DEBUG) {
trace!("Daemon loop iteration, with file descriptors: "); debug!("Daemon loop iteration, with file descriptors: ");
for info in &self.fd_info { for info in &self.fd_info {
trace!("- {}", info.display()); debug!("- {}", info.display());
} }
} }
@ -441,7 +597,7 @@ impl Daemon {
} }
fn handle_event(&mut self, event: &Event) -> Result<(), IoError> { fn handle_event(&mut self, event: &Event) -> Result<(), IoError> {
trace!("Handling event {event:?}"); trace!("Handling event {event:#?}");
match event.token() { match event.token() {
DAEMON => { DAEMON => {
@ -487,16 +643,92 @@ impl Daemon {
}, },
other_token => { other_token => {
// This must be a stream fd. // This must be a stream fd.
let stream_fd = self.fd_for_token(other_token).unwrap_or_else(|| { let fd = self.fd_for_token(other_token).unwrap_or_else(|| {
unreachable!("tried to get fd for non-existent token? {other_token:?}") unreachable!("tried to get fd for non-existent token? {other_token:?}")
}); });
let Some(info) = self.fd_info.get(&fd) else {
panic!("Received an event on an unregistered fd {fd}; IO-safety violation?");
};
let either_available = event.is_readable() || event.is_writable();
if !either_available {
info!(
"File descriptor {} r:{}, w:{}",
info.display(),
event.is_readable(),
event.is_writable(),
);
// FIXME: code duplication
if event.is_read_closed() {
self.deregister(fd);
return Ok(());
}
}
match info.kind {
FdKind::Pid(pid) => {
debug!("Reaping child process {pid}");
// SAFETY: `fd` cannot have been closed yet, since that's what we do here.
let pidfd = unsafe { BorrowedFd::borrow_raw(fd) };
let status = rustix::waitid(WaitId::PidFd(pidfd), WaitIdOptions::EXITED)
.unwrap_or_else(|e| {
todo!("waitid() can fail? on pid {pid}: {e}");
})
.unwrap_or_else(|| {
todo!("waitid() returned None? for pid {pid}");
});
debug!("waitid() for pid {pid} returned status: {status:?}");
let is_dead = status.exited() || status.killed() || status.dumped();
if !is_dead {
todo!("Handle process {pid} events that aren't death: {status:?}");
}
let Some(exit_code) = status.exit_status() else {
unreachable!("Process {pid} died with no exit code at all? {status:?}");
};
debug!("Child process {pid} exited with code {exit_code}");
// Close the pidfd.
self.deregister(fd);
let stream = self
.fd_info
.iter()
.find_map(|info| (info.kind == FdKind::SockStream).then_some(info));
if let Some(stream) = stream {
// SAFETY: fixme.
let stream_fd = unsafe { BorrowedFd::borrow_raw(stream.fd) };
let payload = format!("{{ \"status\": {exit_code} }}\n");
if let Err(e) = rustix::io::write(stream_fd, payload.as_bytes()) {
error!("couldn't write reply to stream fd {stream_fd:?}: {e}");
}
}
},
FdKind::ChildStdout => {
warn!("got stdout");
// SAFETY: oh boy.
let stdout = unsafe { BorrowedFd::borrow_raw(fd) };
self.proxy_stdio(&stdout)
.unwrap_or_else(|e| error!("failed to proxy child stdout: {e}"));
},
FdKind::ChildStderr => {
warn!("got stderr");
// SAFETY: oh boy.
let stderr = unsafe { BorrowedFd::borrow_raw(fd) };
self.proxy_stdio(&stderr)
.unwrap_or_else(|e| error!("failed to proxy child stderr: {e}"));
},
FdKind::SockStream => {
// SAFETY: oh boy.
let stream_fd = unsafe { BorrowedFd::borrow_raw(fd) };
self.read_cmd(&stream_fd).unwrap();
},
kind => todo!("{kind:?}"),
};
if event.is_read_closed() { if event.is_read_closed() {
self.deregister(stream_fd); self.deregister(fd);
} else { return Ok(());
// SAFETY: oh boy.
let stream_fd = unsafe { BorrowedFd::borrow_raw(stream_fd) };
self.read_cmd(&stream_fd).unwrap();
} }
}, },
} }

View file

@ -47,6 +47,24 @@ impl ConvenientAttrPath {
} }
} }
#[derive(Debug, Clone, PartialEq)]
#[derive(Deserialize, Serialize)]
#[serde(untagged)]
pub enum NixLiteral {
String(String),
Number(f64),
// FIXME: add the rest =P
}
impl NixLiteral {
pub fn to_nix_source(&self) -> String {
match self {
NixLiteral::String(s) => format!("\"{s}\""),
NixLiteral::Number(n) => n.to_string(),
}
}
}
#[derive(Debug, Clone, PartialEq)] #[derive(Debug, Clone, PartialEq)]
#[derive(serde::Deserialize, serde::Serialize)] #[derive(serde::Deserialize, serde::Serialize)]
#[serde(tag = "action", content = "args", rename_all = "snake_case")] #[serde(tag = "action", content = "args", rename_all = "snake_case")]
@ -54,6 +72,6 @@ impl ConvenientAttrPath {
pub enum DaemonCmd { pub enum DaemonCmd {
Append { Append {
name: ConvenientAttrPath, name: ConvenientAttrPath,
value: Box<str>, value: Box<NixLiteral>,
}, },
} }

View file

@ -3,6 +3,7 @@ use std::{os::fd::RawFd, sync::OnceLock};
use circular_buffer::CircularBuffer; use circular_buffer::CircularBuffer;
use iddqd::{BiHashItem, IdOrdItem}; use iddqd::{BiHashItem, IdOrdItem};
use mio::Token; use mio::Token;
use rustix::process::Pid;
use crate::prelude::*; use crate::prelude::*;
@ -76,11 +77,11 @@ impl FdInfo {
impl IdOrdItem for FdInfo { impl IdOrdItem for FdInfo {
type Key<'a> = &'a RawFd; type Key<'a> = &'a RawFd;
iddqd::id_upcast!();
fn key(&self) -> &RawFd { fn key(&self) -> &RawFd {
&self.fd &self.fd
} }
iddqd::id_upcast!();
} }
#[derive(Debug)] #[derive(Debug)]
@ -105,13 +106,16 @@ impl<'a> Display for FdInfoDisplay<'a> {
} }
#[derive(Copy)] #[derive(Copy)]
#[derive(Debug, Clone, PartialEq, Eq, PartialOrd, Ord, Hash, Default)] #[derive(Debug, Clone, PartialEq, Eq, Hash, Default)]
#[non_exhaustive] #[non_exhaustive]
pub enum FdKind { pub enum FdKind {
File, File,
Socket, Socket,
SockStream, SockStream,
Poller, Poller,
ChildStdout,
ChildStderr,
Pid(Pid),
#[default] #[default]
Unknown, Unknown,
} }
@ -124,6 +128,9 @@ impl FdKind {
Socket => "socket", Socket => "socket",
SockStream => "socket stream", SockStream => "socket stream",
Poller => "poller", Poller => "poller",
ChildStdout => "child stdout",
ChildStderr => "child stderr",
Pid(_) => "pidfd",
Unknown => "«unknown»", Unknown => "«unknown»",
} }
} }
@ -140,6 +147,8 @@ impl BiHashItem for TokenFd {
type K1<'a> = Token; type K1<'a> = Token;
type K2<'a> = RawFd; type K2<'a> = RawFd;
iddqd::bi_upcast!();
fn key1(&self) -> Token { fn key1(&self) -> Token {
self.token self.token
} }
@ -147,8 +156,6 @@ impl BiHashItem for TokenFd {
fn key2(&self) -> RawFd { fn key2(&self) -> RawFd {
self.fd self.fd
} }
iddqd::bi_upcast!();
} }
impl From<TokenFd> for (Token, RawFd) { impl From<TokenFd> for (Token, RawFd) {

View file

@ -99,6 +99,22 @@ static MK_OVERRIDE_RE: LazyLock<Regex> = LazyLock::new(|| {
Regex::new(r"(?-u)\bmkOverride\s+\((?<priority>[\d-]+)\)").unwrap() Regex::new(r"(?-u)\bmkOverride\s+\((?<priority>[\d-]+)\)").unwrap()
}); });
pub(crate) fn open_source_file(path: Arc<Path>) -> Result<SourceFile, IoError> {
let mut opts = File::options();
opts.read(true)
.write(true)
.create(false)
.custom_flags(libc::O_CLOEXEC);
SourceFile::open_from(Arc::clone(&path), opts)
.tap_err(|e| error!("couldn't open source file at {}: {e}", path.display()))
}
pub(crate) fn get_line_to_insert() -> SourceLine {
//
todo!();
}
#[tracing::instrument(level = "debug")] #[tracing::instrument(level = "debug")]
pub fn do_append(args: Arc<Args>, append_args: AppendCmd) -> Result<(), BoxDynError> { pub fn do_append(args: Arc<Args>, append_args: AppendCmd) -> Result<(), BoxDynError> {
let filepath = Path::new(&args.file); let filepath = Path::new(&args.file);
@ -110,22 +126,16 @@ pub fn do_append(args: Arc<Args>, append_args: AppendCmd) -> Result<(), BoxDynEr
filepath.to_path_buf() filepath.to_path_buf()
}; };
let mut opts = File::options(); let source_file = open_source_file(Arc::from(filepath))?;
opts.read(true)
.write(true)
.create(false)
.custom_flags(libc::O_CLOEXEC);
let source_file = SourceFile::open_from(Arc::from(filepath), opts)?;
let pri = get_where(source_file.clone())?; let pri = get_where(source_file.clone())?;
let new_pri = pri - 1; let new_pri = pri - 1;
let new_pri_line = get_next_prio_line( let new_pri_line = get_next_prio_line(
source_file.clone(), source_file.clone(),
append_args.name, &append_args.name,
new_pri, new_pri,
append_args.value, &append_args.value,
)?; )?;
debug!("new_pri_line={new_pri_line}"); debug!("new_pri_line={new_pri_line}");
@ -210,9 +220,9 @@ pub fn get_where(dynamic_nix: SourceFile) -> Result<i64, BoxDynError> {
pub fn get_next_prio_line( pub fn get_next_prio_line(
source: SourceFile, source: SourceFile,
option_name: Arc<str>, option_name: &str,
new_prio: i64, new_prio: i64,
new_value: Arc<str>, new_value: &str,
) -> Result<SourceLine, BoxDynError> { ) -> Result<SourceLine, BoxDynError> {
let source_lines = source.lines()?; let source_lines = source.lines()?;
let penultimate = source_lines.get(source_lines.len() - 2); let penultimate = source_lines.get(source_lines.len() - 2);
@ -221,6 +231,7 @@ pub fn get_next_prio_line(
let penultimate = penultimate.unwrap(); let penultimate = penultimate.unwrap();
let new_generation = 0 - new_prio; let new_generation = 0 - new_prio;
info!("setting '{option_name}' to '{new_value}' for generation '{new_generation}'");
let new_line = SourceLine { let new_line = SourceLine {
line: penultimate.line, line: penultimate.line,
@ -246,6 +257,8 @@ pub fn write_next_prio(mut source: SourceFile, new_line: SourceLine) -> Result<(
text: Arc::from(" }"), text: Arc::from(" }"),
}; };
debug!("writing new source line: {new_line}");
source.insert_lines(&[new_mod_start, new_line, new_mod_end])?; source.insert_lines(&[new_mod_start, new_line, new_mod_end])?;
Ok(()) Ok(())

View file

@ -6,6 +6,7 @@ use std::{
cell::{Ref, RefCell}, cell::{Ref, RefCell},
hash::Hash, hash::Hash,
io::{BufRead, BufReader, BufWriter}, io::{BufRead, BufReader, BufWriter},
mem::{self, MaybeUninit},
ops::Deref, ops::Deref,
ptr, ptr,
sync::{Arc, Mutex, OnceLock}, sync::{Arc, Mutex, OnceLock},
@ -103,6 +104,28 @@ pub struct SourceFile {
lines: Arc<OnceLock<RefCell<Vec<SourceLine>>>>, lines: Arc<OnceLock<RefCell<Vec<SourceLine>>>>,
} }
#[derive(Debug)]
#[repr(transparent)]
pub struct OpaqueDerefSourceLines<'s>(Ref<'s, [SourceLine]>);
impl<'s> Deref for OpaqueDerefSourceLines<'s> {
type Target = [SourceLine];
fn deref(&self) -> &[SourceLine] {
&*self.0
}
}
#[derive(Debug)]
#[repr(transparent)]
pub struct OpaqueDerefSourceLine<'s>(Ref<'s, SourceLine>);
impl<'s> Deref for OpaqueDerefSourceLine<'s> {
type Target = SourceLine;
fn deref(&self) -> &SourceLine {
&*self.0
}
}
impl SourceFile { impl SourceFile {
/// Panics if `path` is a directory path instead of a file path. /// Panics if `path` is a directory path instead of a file path.
pub fn open_from(path: Arc<Path>, options: OpenOptions) -> Result<Self, IoError> { pub fn open_from(path: Arc<Path>, options: OpenOptions) -> Result<Self, IoError> {
@ -159,15 +182,17 @@ impl SourceFile {
Ok(self._lines_slice()) Ok(self._lines_slice())
} }
pub fn lines(&self) -> Result<impl Deref<Target = [SourceLine]> + '_, IoError> { pub fn lines(&self) -> Result<OpaqueDerefSourceLines<'_>, IoError> {
self._lines() let lines = self._lines()?;
Ok(OpaqueDerefSourceLines(lines))
} }
pub fn line(&self, line: Line) -> Result<impl Deref<Target = SourceLine> + '_, IoError> { pub fn line(&self, line: Line) -> Result<OpaqueDerefSourceLine<'_>, IoError> {
let lines_lock = self._lines()?; let lines_lock = self._lines()?;
let line = Ref::map(lines_lock, |lines| &lines[line.index() as usize]); let line = Ref::map(lines_lock, |lines| &lines[line.index() as usize]);
Ok(line) Ok(OpaqueDerefSourceLine(line))
} }
/// `lines` but already be initialized. /// `lines` but already be initialized.

View file

@ -50,16 +50,11 @@
requiredBy = [ "multi-user.target" ]; requiredBy = [ "multi-user.target" ];
after = [ "default.target" ]; after = [ "default.target" ];
script = '' script = ''
if [[ -e /etc/nixos/hardware-configuration.nix ]]; then
echo "install-dynix: configuration already copied; nothing to do"
exit 0
fi
nix profile install -vv "${dynix.drvPath}^*" # " nix profile install -vv "${dynix.drvPath}^*" # "
mkdir -vp /etc/nixos mkdir -vp /etc/nixos
nixos-generate-config nixos-generate-config
cp -rv --dereference /run/current-system/sw/share/nixos/*.nix /etc/nixos/ cp -rvf --dereference /run/current-system/sw/share/nixos/*.nix /etc/nixos/
if ! [[ -e /etc/nixos/dynix-vm-configuration.nix ]]; then if ! [[ -e /etc/nixos/dynix-vm-configuration.nix ]]; then
echo "FAILURE" echo "FAILURE"
echo "FAILURE" >&2 echo "FAILURE" >&2

View file

@ -2,7 +2,7 @@
# #
# SPDX-License-Identifier: EUPL-1.1 # SPDX-License-Identifier: EUPL-1.1
{ pkgs, lib, modulesPath, ... }: { pkgs, lib, modulesPath, config, ... }:
let let
moduleList = import (modulesPath + "/module-list.nix"); moduleList = import (modulesPath + "/module-list.nix");
@ -44,11 +44,40 @@ in
}; };
}; };
# Setup XDG base directories for me.
security.pam.services.login = {
rules.session.xdg = {
enable = true;
control = "optional";
modulePath = "${pkgs.pam_xdg}/lib/security/pam_xdg.so";
args = [ ];
order = 10500;
};
};
environment.pathsToLink = [ "/share" ]; environment.pathsToLink = [ "/share" ];
environment.extraOutputsToInstall = [ "modules" ]; environment.extraOutputsToInstall = [ "modules" ];
environment.variables = { environment.variables = {
"NIXOS_CONFIG" = "/etc/nixos/configuration.nix"; "NIXOS_CONFIG" = "/etc/nixos/configuration.nix";
}; };
environment.sessionVariables = {
"NIXOS_CONFIG" = "/etc/nixos/configuration.nix";
};
systemd.services.dynix-daemon = {
enable = true;
path = [ config.nix.package ];
serviceConfig = {
Environment = [
"RUST_LOG=trace"
];
ExecSearchPath = [ "/run/current-system/sw/bin" ];
SuccessExitStatus = [ "0" "2" ];
# `bash -l` so XDG_RUNTIME_DIR is set correctly. lol.
ExecStart = "bash -l -c 'exec /root/.nix-profile/bin/dynix daemon --color=always'";
SyslogIdentifier = "dynix-daemon";
};
};
environment.shellAliases = { environment.shellAliases = {
ls = "eza --long --header --group --group-directories-first --classify --binary"; ls = "eza --long --header --group --group-directories-first --classify --binary";
@ -58,5 +87,8 @@ in
eza eza
fd fd
ripgrep ripgrep
netcat.nc
socat
python3
]; ];
} }

View file

@ -5,7 +5,7 @@
from pathlib import Path from pathlib import Path
import shlex import shlex
import textwrap import textwrap
from typing import cast, TYPE_CHECKING from typing import Any, cast, TYPE_CHECKING
from beartype import beartype from beartype import beartype
@ -18,6 +18,13 @@ if TYPE_CHECKING:
assert machine.shell is not None assert machine.shell is not None
ls = "eza -lah --color=always --group-directories-first" ls = "eza -lah --color=always --group-directories-first"
testing_client = "/root/.nix-profile/libexec/dynix-testing-client.py"
ANSI_RESET = "\x1b[0m"
ANSI_BOLD = "\x1b[1m"
ANSI_NOBOLD = "\x1b[22m"
ANSI_RED = "\x1b[31m"
ANSI_GREEN = "\x1b[32m"
@beartype @beartype
def run_log(machine: Machine, *commands: str, timeout: int | None = 60) -> str: def run_log(machine: Machine, *commands: str, timeout: int | None = 60) -> str:
@ -54,71 +61,108 @@ def get_config_file() -> str:
config_file_path.unlink() config_file_path.unlink()
machine.logger.info(f"{ANSI_GREEN}INFO{ANSI_RESET}: got config file:")
machine.logger.info(textwrap.indent(data, " "))
return data return data
@beartype @beartype
def dynix_append(option: str, value: str): def dynix_append_cli(option: str, value: Any):
value = f'"{value}"' if isinstance(value, str) else value
machine.succeed(f''' machine.succeed(f'''
dynix append {shlex.quote(option)} {shlex.quote(value)} dynix append {shlex.quote(option)} {shlex.quote(value)}
'''.strip()) '''.strip())
@beartype
def do_apply():
expr = textwrap.dedent(""" expr = textwrap.dedent("""
(import <nixpkgs/nixos> { }).config.dynamicism.applyDynamicConfiguration { } (import <nixpkgs/nixos> { }).config.dynamicism.applyDynamicConfiguration { }
""").strip() """).strip()
machine.succeed(textwrap.dedent(rf"""
machine.succeed(rf"""
nix run --show-trace --log-format raw-with-logs --impure -E {shlex.quote(expr)} nix run --show-trace --log-format raw-with-logs --impure -E {shlex.quote(expr)}
""".strip()) """).strip())
@beartype
def dynix_append_daemon(option: str, value: Any):
import json
payload = json.dumps(dict(
action="append",
args=dict(
name=option,
value=value,
),
))
machine.succeed(f"echo '{payload}' | {testing_client} /run/user/0/dynix.sock")
@beartype
def run_all_tests(machine: Machine, *, use_daemon: bool):
dynix_append = dynix_append_daemon if use_daemon else dynix_append_cli
dynix_out = machine.succeed("dynix --version")
assert "dynix" in dynix_out, f"dynix not in {dynix_out=}"
machine.succeed("systemctl start user@0.service")
machine.wait_for_unit("user@0.service")
machine.succeed("systemctl start dynix-daemon.service")
machine.wait_for_unit("dynix-daemon.service")
machine.log("REBUILDING configuration inside VM")
machine.succeed("env PAGER= nixos-rebuild switch --log-format raw-with-logs --no-reexec --fallback")
machine.wait_for_unit("gotosocial.service")
# Make sure the config before any dynamic changes is what we expect.
config_text = get_config_file()
lines = config_text.splitlines()
try:
application_name = next(line for line in lines if line.startswith("application-name:"))
except StopIteration:
raise AssertionError(f"no 'application-name:' found in config file: {textwrap.indent(config_text, " ")}")
assert "gotosocial-for-machine" in application_name, f"'gotosocial-for-machine' should be in {application_name=}"
try:
host = next(line for line in lines if line.startswith("host:"))
except StopIteration:
raise AssertionError(f"no 'host:' found in config file: {textwrap.indent(config_text, " ")}")
assert "gotosocial-machine" in host, f"'gotosocial-machine' should be in {host=}"
new_app_name = "yay!"
dynix_append("services.gotosocial.settings.application-name", f'{new_app_name}')
config_text = get_config_file()
lines = config_text.splitlines()
try:
application_name = next(line for line in lines if line.startswith("application-name:"))
except StopIteration:
raise AssertionError(f"no 'application-name:' found in config file: {textwrap.indent(config_text, " ")}")
assert new_app_name in application_name, f"'{new_app_name}' should be in {application_name=}"
machine.log("REBUILDING configuration inside VM")
machine.succeed("env PAGER= nixos-rebuild switch --log-format raw-with-logs --no-reexec --fallback")
machine.wait_for_unit("gotosocial.service")
config_text = get_config_file()
lines = config_text.splitlines()
try:
application_name = next(line for line in lines if line.startswith("application-name:"))
except StopIteration:
raise AssertionError(f"no 'application-name:' found in config file: {textwrap.indent(config_text, " ")}")
assert "gotosocial-for-machine" in application_name, f"'gotosocial-for-machine' should be in {application_name=}"
machine.start(allow_reboot=True)
machine.wait_for_unit("default.target") machine.wait_for_unit("default.target")
machine.wait_for_unit("install-dynix.service") machine.wait_for_unit("install-dynix.service")
dynix_out = machine.succeed("dynix --version")
assert "dynix" in dynix_out, f"dynix not in {dynix_out=}"
machine.log("REBUILDING configuration inside VM")
machine.succeed("env PAGER= nixos-rebuild switch --log-format raw-with-logs --no-reexec --fallback")
machine.wait_for_unit("gotosocial.service")
# Make sure the config before any dynamic changes is what we expect.
config_text = get_config_file()
lines = config_text.splitlines()
try: try:
application_name = next(line for line in lines if line.startswith("application-name:")) run_all_tests(machine, use_daemon=False)
except StopIteration: except Exception as e:
raise AssertionError(f"no 'application-name:' found in config file: {textwrap.indent(config_text, " ")}") machine.logger.error(f"{ANSI_RED}ERROR{ANSI_RESET} during {ANSI_BOLD}CLI{ANSI_RESET} tests: {e}")
assert "gotosocial-for-machine" in application_name, f"'gotosocial-for-machine' should be in {application_name=}" raise
machine.reboot()
machine.wait_for_unit("install-dynix.service")
try: try:
host = next(line for line in lines if line.startswith("host:")) run_all_tests(machine, use_daemon=True)
except StopIteration: except Exception as e:
raise AssertionError(f"no 'host:' found in config file: {textwrap.indent(config_text, " ")}") machine.logger.error(f"{ANSI_RED}ERROR{ANSI_RESET} during {ANSI_BOLD}daemon{ANSI_RESET} tests: {e}")
assert "gotosocial-machine" in host, f"'gotosocial-machine' should be in {host=}" raise
new_app_name = "yay!"
dynix_append("services.gotosocial.settings.application-name", f'"{new_app_name}"')
do_apply()
config_text = get_config_file()
lines = config_text.splitlines()
try:
application_name = next(line for line in lines if line.startswith("application-name:"))
except StopIteration:
raise AssertionError(f"no 'application-name:' found in config file: {textwrap.indent(config_text, " ")}")
assert new_app_name in application_name, f"'{new_app_name}' should be in {application_name=}"
machine.log("REBUILDING configuration inside VM")
machine.succeed("env PAGER= nixos-rebuild switch --log-format raw-with-logs --no-reexec --fallback")
machine.wait_for_unit("gotosocial.service")
config_text = get_config_file()
lines = config_text.splitlines()
try:
application_name = next(line for line in lines if line.startswith("application-name:"))
except StopIteration:
raise AssertionError(f"no 'application-name:' found in config file: {textwrap.indent(config_text, " ")}")
assert "gotosocial-for-machine" in application_name, f"'gotosocial-for-machine' should be in {application_name=}"

View file

@ -4,6 +4,7 @@
import functools import functools
from pathlib import Path from pathlib import Path
import json
import shlex import shlex
import textwrap import textwrap
import tomllib import tomllib
@ -20,6 +21,7 @@ if TYPE_CHECKING:
assert machine.shell is not None assert machine.shell is not None
ls = "eza -lah --color=always --group-directories-first" ls = "eza -lah --color=always --group-directories-first"
testing_client = "/root/.nix-profile/libexec/dynix-testing-client.py"
indent = functools.partial(textwrap.indent, prefix=' ') indent = functools.partial(textwrap.indent, prefix=' ')
@ -55,65 +57,114 @@ def get_config_file() -> dict[str, Any]:
with open(config_file_path, "rb") as f: with open(config_file_path, "rb") as f:
config_data = tomllib.load(f) config_data = tomllib.load(f)
try:
config_file_path.unlink() config_file_path.unlink()
except Exception as e:
machine.log(f"Couldn't unlike path {config_file_path}: {e}")
raise
return config_data return config_data
@beartype @beartype
def dynix_append(option: str, value: Any): def dynix_append_daemon(option: str, value: Any):
payload = json.dumps(dict(
action="append",
args=dict(
name=option,
value=value,
),
))
machine.succeed(f"echo '{payload}' | {testing_client} /run/user/0/dynix.sock")
@beartype
def dynix_append_cli(option: str, value: Any):
machine.succeed(f''' machine.succeed(f'''
dynix append {shlex.quote(option)} {shlex.quote(str(value))} dynix append {shlex.quote(option)} {shlex.quote(str(value))}
'''.strip()) '''.strip())
@beartype
def do_apply():
expr = textwrap.dedent(""" expr = textwrap.dedent("""
(import <nixpkgs/nixos> { }).config.dynamicism.applyDynamicConfiguration { } (import <nixpkgs/nixos> { }).config.dynamicism.applyDynamicConfiguration { }
""").strip() """).strip()
machine.succeed(rf""" machine.succeed(rf"""
nix run --show-trace --log-format raw-with-logs --impure -E {shlex.quote(expr)} nix run --show-trace --log-format raw-with-logs --impure -E {shlex.quote(expr)}
""".strip()) """.strip())
machine.wait_for_unit("default.target") @beartype
machine.wait_for_unit("install-dynix.service") def run_all_tests(machine: Machine, *, use_daemon: bool):
dynix_append = dynix_append_daemon if use_daemon else dynix_append_cli
dynix_out = machine.succeed("dynix --version") #
assert "dynix" in dynix_out, f"dynix not in {dynix_out=}" # Setup.
#
dynix_out = machine.succeed("dynix --version")
assert "dynix" in dynix_out, f"dynix not in {dynix_out=}"
# Config should have our initial values. machine.succeed("systemctl start user@0.service")
config_toml = get_config_file() machine.wait_for_unit("user@0.service")
assert int(config_toml['workers']) == 4, f"{config_toml['workers']=} != 4"
assert int(config_toml['max_connection_rate']) == 256, f"{config_toml['max_connection_rate']=} != 256"
with machine.nested("must succeed: initial nixos-rebuild switch"): run_log(machine, "systemctl start dynix-daemon.service")
machine.wait_for_unit("dynix-daemon.service")
machine.log("Checking initial harmonia.service conditions")
# Config should have our initial values.
config_toml = get_config_file()
assert int(config_toml['workers']) == 4, f"{config_toml['workers']=} != 4"
assert int(config_toml['max_connection_rate']) == 256, f"{config_toml['max_connection_rate']=} != 256"
with machine.nested("must succeed: initial nixos-rebuild switch"):
machine.succeed("env PAGER= nixos-rebuild switch --log-format raw-with-logs --no-reexec -v --fallback") machine.succeed("env PAGER= nixos-rebuild switch --log-format raw-with-logs --no-reexec -v --fallback")
# Config should not have changed. # Config should not have changed.
config_toml = get_config_file() config_toml = get_config_file()
assert int(config_toml['workers']) == 4, f"{config_toml['workers']=} != 4" assert int(config_toml['workers']) == 4, f"{config_toml['workers']=} != 4"
assert int(config_toml['max_connection_rate']) == 256, f"{config_toml['max_connection_rate']=} != 256" assert int(config_toml['max_connection_rate']) == 256, f"{config_toml['max_connection_rate']=} != 256"
new_workers = 20 machine.log("Testing dynamic workers=20")
dynix_append("services.harmonia.settings.workers", new_workers) new_workers = 20
do_apply() dynix_append("services.harmonia.settings.workers", new_workers)
# Workers, but not max connection rate, should have changed. machine.log("Testing that workers, but not max_connection_rate, changed")
config_toml = get_config_file() # Workers, but not max connection rate, should have changed.
assert int(config_toml['workers']) == new_workers, f"{config_toml['workers']=} != {new_workers}" config_toml = get_config_file()
assert int(config_toml['max_connection_rate']) == 256, f"{config_toml['max_connection_rate']=} != 256" from pprint import pformat
machine.log(pformat(config_toml))
assert int(config_toml['workers']) == new_workers, f"{config_toml['workers']=} != {new_workers}"
assert int(config_toml['max_connection_rate']) == 256, f"{config_toml['max_connection_rate']=} != 256"
new_max_connection_rate = 100 machine.log("Testing dynamic max_connection_rate=100")
dynix_append("services.harmonia.settings.max_connection_rate", new_max_connection_rate) new_max_connection_rate = 100
do_apply() dynix_append("services.harmonia.settings.max_connection_rate", new_max_connection_rate)
# Max connection rate should have changed, and workers should be the same as before. # Max connection rate should have changed, and workers should be the same as before.
config_toml = get_config_file() config_toml = get_config_file()
assert int(config_toml['max_connection_rate']) == new_max_connection_rate, f"{config_toml['max_connection_rate']=} != {new_max_connection_rate}" print(f"checking connection rate, {use_daemon=}")
assert int(config_toml['workers']) == new_workers, f"{config_toml['workers']=} != {new_workers}" assert int(config_toml['max_connection_rate']) == new_max_connection_rate, f"{config_toml['max_connection_rate']=} != {new_max_connection_rate}"
print(f"checking workers, {use_daemon=}")
assert int(config_toml['workers']) == new_workers, f"{config_toml['workers']=} != {new_workers}"
# And this should set everything back. machine.log("Done with tests; stopping dynix-daemon")
machine.succeed("env PAGER= nixos-rebuild switch --log-format raw-with-logs --no-reexec -v --fallback") machine.succeed("systemctl stop dynix-daemon.service")
machine.wait_for_unit("harmonia.service")
config_toml = get_config_file() # And this should set everything back.
assert int(config_toml['max_connection_rate']) == 256, f'{config_toml["max_connection_rate"]=} != 256' machine.succeed("env PAGER= nixos-rebuild switch --log-format raw-with-logs --no-reexec -v --fallback")
assert int(config_toml['workers']) == 4, f'{config_toml["workers"]=} != 4' machine.wait_for_unit("harmonia.service")
config_toml = get_config_file()
assert int(config_toml['max_connection_rate']) == 256, f'{config_toml["max_connection_rate"]=} != 256'
assert int(config_toml['workers']) == 4, f'{config_toml["workers"]=} != 4'
machine.start(allow_reboot=True)
machine.wait_for_unit("install-dynix.service")
try:
run_all_tests(machine, use_daemon=False)
except Exception as e:
machine.log(f"ERROR running CLI tests: {e}")
machine.reboot()
machine.wait_for_unit("install-dynix.service")
try:
run_all_tests(machine, use_daemon=True)
except Exception as e:
machine.log(f"ERROR running DAEMON tests: {e}")
raise