mirror of
https://gitlab.com/fabinfra/fabaccess/bffh.git
synced 2024-11-11 01:53:23 +01:00
208 lines
6.0 KiB
Rust
208 lines
6.0 KiB
Rust
#[macro_use]
|
|
extern crate slog;
|
|
|
|
#[macro_use]
|
|
extern crate capnp_rpc;
|
|
|
|
#[macro_use]
|
|
extern crate async_trait;
|
|
|
|
mod modules;
|
|
mod log;
|
|
mod api;
|
|
mod config;
|
|
mod error;
|
|
mod connection;
|
|
mod schema;
|
|
mod db;
|
|
mod machine;
|
|
mod builtin;
|
|
mod server;
|
|
mod network;
|
|
mod actor;
|
|
mod initiator;
|
|
|
|
use clap::{App, Arg};
|
|
|
|
use futures::prelude::*;
|
|
use futures::executor::{LocalPool, ThreadPool};
|
|
use futures::compat::Stream01CompatExt;
|
|
use futures::join;
|
|
use futures::task::LocalSpawn;
|
|
|
|
use std::io;
|
|
use std::io::Write;
|
|
use std::path::PathBuf;
|
|
use std::str::FromStr;
|
|
|
|
use std::sync::Arc;
|
|
|
|
use lmdb::Transaction;
|
|
use smol::net::TcpListener;
|
|
|
|
use smol::Executor;
|
|
|
|
use error::Error;
|
|
|
|
use slog::Logger;
|
|
|
|
use paho_mqtt::AsyncClient;
|
|
|
|
fn main() {
|
|
use clap::{crate_version, crate_description, crate_name};
|
|
|
|
// Argument parsing
|
|
// values for the name, description and version are pulled from `Cargo.toml`.
|
|
let matches = App::new(crate_name!())
|
|
.about(crate_description!())
|
|
.version(crate_version!())
|
|
.arg(Arg::with_name("config")
|
|
.help("Path to the config file to use")
|
|
.long("config")
|
|
.short("c")
|
|
.takes_value(true)
|
|
)
|
|
.arg(Arg::with_name("print default")
|
|
.help("Print a default config to stdout instead of running")
|
|
.long("print-default")
|
|
)
|
|
.arg(Arg::with_name("dump")
|
|
.help("Dump all databases into the given directory")
|
|
.long("dump")
|
|
.conflicts_with("load")
|
|
.takes_value(true)
|
|
)
|
|
.arg(Arg::with_name("load")
|
|
.help("Load databases from the given directory")
|
|
.long("load")
|
|
.conflicts_with("dump")
|
|
.takes_value(true)
|
|
)
|
|
.get_matches();
|
|
|
|
// Check for the --print-default option first because we don't need to do anything else in that
|
|
// case.
|
|
if matches.is_present("print default") {
|
|
let config = config::Config::default();
|
|
let encoded = serde_dhall::serialize(&config).to_string().unwrap();
|
|
|
|
// Direct writing to fd 1 is faster but also prevents any print-formatting that could
|
|
// invalidate the generated TOML
|
|
let stdout = io::stdout();
|
|
let mut handle = stdout.lock();
|
|
handle.write_all(&encoded.as_bytes()).unwrap();
|
|
|
|
// Early return to exit.
|
|
return;
|
|
}
|
|
|
|
let retval;
|
|
|
|
// Scope to drop everything before exiting.
|
|
{
|
|
// Initialize the logging subsystem first to be able to better document the progress from now
|
|
// on.
|
|
// TODO: Now would be a really good time to close stdin/out and move logging to syslog
|
|
// Log is in an Arc so we can do very cheap clones in closures.
|
|
let log = Arc::new(log::init());
|
|
info!(log, "Starting");
|
|
|
|
match maybe(matches, log.clone()) {
|
|
Ok(_) => retval = 0,
|
|
Err(e) => {
|
|
error!(log, "{}", e);
|
|
retval = -1;
|
|
}
|
|
}
|
|
}
|
|
|
|
std::process::exit(retval);
|
|
}
|
|
|
|
// Returning a `Result` from `main` allows us to use the `?` shorthand.
|
|
// In the case of an Err it will be printed using `fmt::Debug`
|
|
fn maybe(matches: clap::ArgMatches, log: Arc<Logger>) -> Result<(), Error> {
|
|
// If no `config` option is given use a preset default.
|
|
let configpath = matches.value_of("config").unwrap_or("/etc/diflouroborane.dhall");
|
|
let config = config::read(&PathBuf::from_str(configpath).unwrap())?;
|
|
debug!(log, "Loaded Config: {:?}", config);
|
|
|
|
if matches.is_present("dump") {
|
|
error!(log, "Dumping is currently not implemented");
|
|
Ok(())
|
|
} else if matches.is_present("load") {
|
|
let db = db::Databases::new(&log, &config)?;
|
|
let mut dir = PathBuf::from(matches.value_of_os("load").unwrap());
|
|
|
|
dir.push("users.toml");
|
|
let map = db::user::load_file(&dir)?;
|
|
for (uid,user) in map.iter() {
|
|
db.userdb.put_user(uid, user)?;
|
|
}
|
|
debug!(log, "Loaded users: {:?}", map);
|
|
dir.pop();
|
|
|
|
dir.push("roles.toml");
|
|
db.access.internal.load_roles(&dir)?;
|
|
dir.pop();
|
|
|
|
dir.push("pass.toml");
|
|
db.passdb.load_file(&dir);
|
|
dir.pop();
|
|
|
|
Ok(())
|
|
} else {
|
|
let ex = Executor::new();
|
|
|
|
let mqtt = AsyncClient::new(config.mqtt_url.clone())?;
|
|
let tok = mqtt.connect(paho_mqtt::ConnectOptions::new());
|
|
|
|
smol::block_on(tok)?;
|
|
|
|
let machines = machine::load(&config)?;
|
|
let (mut actor_map, actors) = actor::load(&log, &mqtt, &config)?;
|
|
let (mut init_map, initiators) = initiator::load(&log, &mqtt, &config)?;
|
|
|
|
// TODO: restore connections between initiators, machines, actors
|
|
let mut network = network::Network::new(machines, actor_map, init_map);
|
|
|
|
for (a,b) in config.actor_connections.iter() {
|
|
if let Err(e) = network.connect_actor(a,b) {
|
|
error!(log, "{}", e);
|
|
}
|
|
}
|
|
|
|
for (a,b) in config.init_connections.iter() {
|
|
if let Err(e) = network.connect_init(a,b) {
|
|
error!(log, "{}", e);
|
|
}
|
|
}
|
|
|
|
for actor in actors.into_iter() {
|
|
ex.spawn(actor).detach();
|
|
}
|
|
for init in initiators.into_iter() {
|
|
ex.spawn(init).detach();
|
|
}
|
|
|
|
let (signal, shutdown) = async_channel::bounded::<()>(1);
|
|
let (_, r) = easy_parallel::Parallel::new()
|
|
.each(0..4, |_| smol::block_on(ex.run(shutdown.recv())))
|
|
.finish(|| {
|
|
let db = db::Databases::new(&log, &config)?;
|
|
// TODO: Spawn api connections on their own (non-main) thread, use the main thread to
|
|
// handle signals (a cli if stdin is not closed?) and make it stop and clean up all threads
|
|
// when bffh should exit
|
|
let r = server::serve_api_connections(log.clone(), config, db, network);
|
|
|
|
// One of them would be enough really, but *shrug*
|
|
signal.try_send(());
|
|
std::mem::drop(signal);
|
|
|
|
return r;
|
|
});
|
|
|
|
return r;
|
|
}
|
|
}
|