mirror of
https://github.com/EFForg/rayhunter.git
synced 2026-04-27 07:59:59 -07:00
See https://github.com/EFForg/rayhunter/issues/334 Severity levels low, medium, high are now exposed to the UI in form of dotted, dashed and solid lines. The line on the UI represents the highest-so-far severity seen. Originally this was intended to be represented by Yellow/Orange/Red, but this would mean yet another divergence for colorblind mode. This is colorblind-friendly by default (I think...) As part of this, simplify EventType so that it becomes a flat "level" enum without nested variants. There is also a new debug endpoint that allows one to overwrite the display level directly for testing.
327 lines
11 KiB
Rust
327 lines
11 KiB
Rust
mod analysis;
|
|
mod config;
|
|
mod diag;
|
|
mod display;
|
|
mod error;
|
|
mod key_input;
|
|
mod notifications;
|
|
mod pcap;
|
|
mod qmdl_store;
|
|
mod server;
|
|
mod stats;
|
|
|
|
use std::net::SocketAddr;
|
|
use std::sync::Arc;
|
|
use std::sync::atomic::{AtomicBool, Ordering};
|
|
|
|
use crate::config::{parse_args, parse_config};
|
|
use crate::diag::run_diag_read_thread;
|
|
use crate::error::RayhunterError;
|
|
use crate::notifications::{NotificationService, run_notification_worker};
|
|
use crate::pcap::get_pcap;
|
|
use crate::qmdl_store::RecordingStore;
|
|
use crate::server::{
|
|
ServerState, debug_set_display_state, get_config, get_qmdl, get_zip, serve_static, set_config,
|
|
};
|
|
use crate::stats::{get_qmdl_manifest, get_system_stats};
|
|
|
|
use analysis::{
|
|
AnalysisCtrlMessage, AnalysisStatus, get_analysis_status, run_analysis_thread, start_analysis,
|
|
};
|
|
use axum::Router;
|
|
use axum::response::Redirect;
|
|
use axum::routing::{get, post};
|
|
use diag::{
|
|
DiagDeviceCtrlMessage, delete_all_recordings, delete_recording, get_analysis_report,
|
|
start_recording, stop_recording,
|
|
};
|
|
use log::{error, info};
|
|
use qmdl_store::RecordingStoreError;
|
|
use rayhunter::Device;
|
|
use rayhunter::diag_device::DiagDevice;
|
|
use tokio::net::TcpListener;
|
|
use tokio::select;
|
|
use tokio::sync::mpsc::{self, Sender};
|
|
use tokio::sync::{RwLock, oneshot};
|
|
use tokio::task::JoinHandle;
|
|
use tokio_util::task::TaskTracker;
|
|
|
|
type AppRouter = Router<Arc<ServerState>>;
|
|
|
|
fn get_router() -> AppRouter {
|
|
Router::new()
|
|
.route("/api/pcap/{name}", get(get_pcap))
|
|
.route("/api/qmdl/{name}", get(get_qmdl))
|
|
.route("/api/zip/{name}", get(get_zip))
|
|
.route("/api/system-stats", get(get_system_stats))
|
|
.route("/api/qmdl-manifest", get(get_qmdl_manifest))
|
|
.route("/api/start-recording", post(start_recording))
|
|
.route("/api/stop-recording", post(stop_recording))
|
|
.route("/api/delete-recording/{name}", post(delete_recording))
|
|
.route("/api/delete-all-recordings", post(delete_all_recordings))
|
|
.route("/api/analysis-report/{name}", get(get_analysis_report))
|
|
.route("/api/analysis", get(get_analysis_status))
|
|
.route("/api/analysis/{name}", post(start_analysis))
|
|
.route("/api/config", get(get_config))
|
|
.route("/api/config", post(set_config))
|
|
.route("/api/debug/display-state", post(debug_set_display_state))
|
|
.route("/", get(|| async { Redirect::permanent("/index.html") }))
|
|
.route("/{*path}", get(serve_static))
|
|
}
|
|
|
|
// Runs the axum server, taking all the elements needed to build up our
|
|
// ServerState and a oneshot Receiver that'll fire when it's time to shutdown
|
|
// (i.e. user hit ctrl+c)
|
|
async fn run_server(
|
|
task_tracker: &TaskTracker,
|
|
state: Arc<ServerState>,
|
|
server_shutdown_rx: oneshot::Receiver<()>,
|
|
) -> JoinHandle<()> {
|
|
info!("spinning up server");
|
|
let addr = SocketAddr::from(([0, 0, 0, 0], state.config.port));
|
|
let listener = TcpListener::bind(&addr).await.unwrap();
|
|
let app = get_router().with_state(state);
|
|
|
|
task_tracker.spawn(async move {
|
|
info!("The orca is hunting for stingrays...");
|
|
axum::serve(listener, app)
|
|
.with_graceful_shutdown(server_shutdown_signal(server_shutdown_rx))
|
|
.await
|
|
.unwrap();
|
|
})
|
|
}
|
|
|
|
async fn server_shutdown_signal(server_shutdown_rx: oneshot::Receiver<()>) {
|
|
server_shutdown_rx.await.unwrap();
|
|
info!("Server received shutdown signal, exiting...");
|
|
}
|
|
|
|
// Loads a RecordingStore if one exists, and if not, only create one if we're
|
|
// not in debug mode. If we fail to parse the manifest AND we're not in debug
|
|
// mode, try to recover the manifest from the existing QMDL files
|
|
async fn init_qmdl_store(config: &config::Config) -> Result<RecordingStore, RayhunterError> {
|
|
let store_exists = RecordingStore::exists(&config.qmdl_store_path).await?;
|
|
if config.debug_mode {
|
|
if store_exists {
|
|
Ok(RecordingStore::load(&config.qmdl_store_path).await?)
|
|
} else {
|
|
Err(RayhunterError::NoStoreDebugMode(
|
|
config.qmdl_store_path.clone(),
|
|
))
|
|
}
|
|
} else if store_exists {
|
|
match RecordingStore::load(&config.qmdl_store_path).await {
|
|
Ok(store) => Ok(store),
|
|
Err(RecordingStoreError::ParseManifestError(err)) => {
|
|
error!("failed to parse QMDL manifest: {err}");
|
|
info!("recovering manifest from existing QMDL files...");
|
|
Ok(RecordingStore::recover(&config.qmdl_store_path).await?)
|
|
}
|
|
Err(err) => Err(err.into()),
|
|
}
|
|
} else {
|
|
Ok(RecordingStore::create(&config.qmdl_store_path).await?)
|
|
}
|
|
}
|
|
|
|
// Start a thread that'll track when user hits ctrl+c. When that happens,
|
|
// trigger various cleanup tasks, including sending signals to other threads to
|
|
// shutdown
|
|
#[allow(clippy::too_many_arguments)]
|
|
fn run_shutdown_thread(
|
|
task_tracker: &TaskTracker,
|
|
diag_device_sender: Sender<DiagDeviceCtrlMessage>,
|
|
daemon_restart_rx: oneshot::Receiver<()>,
|
|
should_restart_flag: Arc<AtomicBool>,
|
|
server_shutdown_tx: oneshot::Sender<()>,
|
|
maybe_ui_shutdown_tx: Option<oneshot::Sender<()>>,
|
|
maybe_key_input_shutdown_tx: Option<oneshot::Sender<()>>,
|
|
qmdl_store_lock: Arc<RwLock<RecordingStore>>,
|
|
analysis_tx: Sender<AnalysisCtrlMessage>,
|
|
) -> JoinHandle<Result<(), RayhunterError>> {
|
|
info!("create shutdown thread");
|
|
|
|
task_tracker.spawn(async move {
|
|
select! {
|
|
res = tokio::signal::ctrl_c() => {
|
|
if let Err(err) = res {
|
|
error!("Unable to listen for shutdown signal: {err}");
|
|
}
|
|
|
|
should_restart_flag.store(false, Ordering::Relaxed);
|
|
}
|
|
res = daemon_restart_rx => {
|
|
if let Err(err) = res {
|
|
error!("Unable to listen for shutdown signal: {err}");
|
|
}
|
|
|
|
should_restart_flag.store(true, Ordering::Relaxed);
|
|
}
|
|
};
|
|
|
|
let mut qmdl_store = qmdl_store_lock.write().await;
|
|
if qmdl_store.current_entry.is_some() {
|
|
info!("Closing current QMDL entry...");
|
|
qmdl_store.close_current_entry().await?;
|
|
info!("Done!");
|
|
}
|
|
|
|
server_shutdown_tx
|
|
.send(())
|
|
.expect("couldn't send server shutdown signal");
|
|
if let Some(ui_shutdown_tx) = maybe_ui_shutdown_tx {
|
|
let _ = ui_shutdown_tx.send(());
|
|
}
|
|
if let Some(key_input_shutdown_tx) = maybe_key_input_shutdown_tx {
|
|
let _ = key_input_shutdown_tx.send(());
|
|
}
|
|
diag_device_sender
|
|
.send(DiagDeviceCtrlMessage::Exit)
|
|
.await
|
|
.expect("couldn't send Exit message to diag thread");
|
|
analysis_tx
|
|
.send(AnalysisCtrlMessage::Exit)
|
|
.await
|
|
.expect("couldn't send Exit message to analysis thread");
|
|
Ok(())
|
|
})
|
|
}
|
|
|
|
#[tokio::main(flavor = "current_thread")]
|
|
async fn main() -> Result<(), RayhunterError> {
|
|
env_logger::init();
|
|
|
|
rustls_rustcrypto::provider()
|
|
.install_default()
|
|
.expect("Couldn't install rustcrypto provider");
|
|
|
|
let args = parse_args();
|
|
|
|
loop {
|
|
let config = parse_config(&args.config_path).await?;
|
|
if !run_with_config(&args, config).await? {
|
|
return Ok(());
|
|
}
|
|
}
|
|
}
|
|
|
|
async fn run_with_config(
|
|
args: &config::Args,
|
|
config: config::Config,
|
|
) -> Result<bool, RayhunterError> {
|
|
// TaskTrackers give us an interface to spawn tokio threads, and then
|
|
// eventually await all of them ending
|
|
let task_tracker = TaskTracker::new();
|
|
println!("R A Y H U N T E R 🐳");
|
|
|
|
let store = init_qmdl_store(&config).await?;
|
|
let analysis_status = AnalysisStatus::new(&store);
|
|
let qmdl_store_lock = Arc::new(RwLock::new(store));
|
|
let (diag_tx, diag_rx) = mpsc::channel::<DiagDeviceCtrlMessage>(1);
|
|
let (ui_update_tx, ui_update_rx) = mpsc::channel::<display::DisplayState>(1);
|
|
let (analysis_tx, analysis_rx) = mpsc::channel::<AnalysisCtrlMessage>(5);
|
|
let mut maybe_ui_shutdown_tx = None;
|
|
let mut maybe_key_input_shutdown_tx = None;
|
|
|
|
let notification_service = NotificationService::new(config.ntfy_url.clone());
|
|
|
|
if !config.debug_mode {
|
|
let (ui_shutdown_tx, ui_shutdown_rx) = oneshot::channel();
|
|
maybe_ui_shutdown_tx = Some(ui_shutdown_tx);
|
|
info!("Using configuration for device: {0:?}", config.device);
|
|
let mut dev = DiagDevice::new(&config.device)
|
|
.await
|
|
.map_err(RayhunterError::DiagInitError)?;
|
|
dev.config_logs()
|
|
.await
|
|
.map_err(RayhunterError::DiagInitError)?;
|
|
|
|
info!("Starting Diag Thread");
|
|
run_diag_read_thread(
|
|
&task_tracker,
|
|
dev,
|
|
diag_rx,
|
|
diag_tx.clone(),
|
|
ui_update_tx.clone(),
|
|
qmdl_store_lock.clone(),
|
|
analysis_tx.clone(),
|
|
config.analyzers.clone(),
|
|
notification_service.new_handler(),
|
|
);
|
|
info!("Starting UI");
|
|
|
|
let update_ui = match &config.device {
|
|
Device::Orbic => display::orbic::update_ui,
|
|
Device::Tplink => display::tplink::update_ui,
|
|
Device::Tmobile => display::tmobile::update_ui,
|
|
Device::Wingtech => display::wingtech::update_ui,
|
|
Device::Pinephone => display::headless::update_ui,
|
|
Device::Uz801 => display::uz801::update_ui,
|
|
};
|
|
update_ui(&task_tracker, &config, ui_shutdown_rx, ui_update_rx);
|
|
|
|
info!("Starting Key Input service");
|
|
let (key_input_shutdown_tx, key_input_shutdown_rx) = oneshot::channel();
|
|
maybe_key_input_shutdown_tx = Some(key_input_shutdown_tx);
|
|
key_input::run_key_input_thread(
|
|
&task_tracker,
|
|
&config,
|
|
diag_tx.clone(),
|
|
key_input_shutdown_rx,
|
|
);
|
|
}
|
|
|
|
let (daemon_restart_tx, daemon_restart_rx) = oneshot::channel::<()>();
|
|
let (server_shutdown_tx, server_shutdown_rx) = oneshot::channel::<()>();
|
|
let analysis_status_lock = Arc::new(RwLock::new(analysis_status));
|
|
run_analysis_thread(
|
|
&task_tracker,
|
|
analysis_rx,
|
|
qmdl_store_lock.clone(),
|
|
analysis_status_lock.clone(),
|
|
config.analyzers.clone(),
|
|
);
|
|
let should_restart_flag = Arc::new(AtomicBool::new(false));
|
|
|
|
run_shutdown_thread(
|
|
&task_tracker,
|
|
diag_tx.clone(),
|
|
daemon_restart_rx,
|
|
should_restart_flag.clone(),
|
|
server_shutdown_tx,
|
|
maybe_ui_shutdown_tx,
|
|
maybe_key_input_shutdown_tx,
|
|
qmdl_store_lock.clone(),
|
|
analysis_tx.clone(),
|
|
);
|
|
run_notification_worker(&task_tracker, notification_service);
|
|
let state = Arc::new(ServerState {
|
|
config_path: args.config_path.clone(),
|
|
config,
|
|
qmdl_store_lock: qmdl_store_lock.clone(),
|
|
diag_device_ctrl_sender: diag_tx,
|
|
analysis_status_lock,
|
|
analysis_sender: analysis_tx,
|
|
daemon_restart_tx: Arc::new(RwLock::new(Some(daemon_restart_tx))),
|
|
ui_update_sender: Some(ui_update_tx),
|
|
});
|
|
run_server(&task_tracker, state, server_shutdown_rx).await;
|
|
|
|
task_tracker.close();
|
|
task_tracker.wait().await;
|
|
|
|
info!("see you space cowboy...");
|
|
Ok(should_restart_flag.load(Ordering::Relaxed))
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod test {
|
|
use super::*;
|
|
|
|
#[test]
|
|
fn test_get_router() {
|
|
// assert that creating the router does not panic from invalid route patterns.
|
|
let _ = get_router();
|
|
}
|
|
}
|