mirror of
https://github.com/edera-dev/krata.git
synced 2025-08-03 13:11:31 +00:00
* chore(code): simple code cleanup * chore(code): additional code cleanup * feature(krata): rework api and make ip assignment persistent to database * rework and cleanup * fix daemon config references
538 lines
17 KiB
Rust
538 lines
17 KiB
Rust
use std::{
|
|
collections::HashMap,
|
|
sync::atomic::{fence, Ordering},
|
|
time::Duration,
|
|
};
|
|
|
|
use anyhow::{anyhow, Result};
|
|
use log::{debug, error};
|
|
use tokio::{
|
|
select,
|
|
sync::mpsc::{channel, Receiver, Sender},
|
|
task::JoinHandle,
|
|
time::sleep,
|
|
};
|
|
use xenevtchn::EventChannelService;
|
|
use xengnt::{sys::GrantRef, GrantTab, MappedMemory};
|
|
use xenstore::{XsdClient, XsdInterface};
|
|
|
|
const SINGLE_CHANNEL_QUEUE_LEN: usize = 100;
|
|
const GROUPED_CHANNEL_QUEUE_LEN: usize = 1000;
|
|
|
|
#[repr(C)]
|
|
struct XenConsoleInterface {
|
|
input: [u8; XenConsoleInterface::INPUT_SIZE],
|
|
output: [u8; XenConsoleInterface::OUTPUT_SIZE],
|
|
in_cons: u32,
|
|
in_prod: u32,
|
|
out_cons: u32,
|
|
out_prod: u32,
|
|
}
|
|
|
|
unsafe impl Send for XenConsoleInterface {}
|
|
|
|
impl XenConsoleInterface {
|
|
const INPUT_SIZE: usize = 1024;
|
|
const OUTPUT_SIZE: usize = 2048;
|
|
}
|
|
|
|
pub struct ChannelService {
|
|
typ: String,
|
|
use_reserved_ref: Option<u64>,
|
|
backends: HashMap<u32, ChannelBackend>,
|
|
evtchn: EventChannelService,
|
|
store: XsdClient,
|
|
gnttab: GrantTab,
|
|
input_receiver: Receiver<(u32, Vec<u8>)>,
|
|
pub input_sender: Sender<(u32, Vec<u8>)>,
|
|
output_sender: Sender<(u32, Option<Vec<u8>>)>,
|
|
}
|
|
|
|
impl ChannelService {
|
|
pub async fn new(
|
|
typ: String,
|
|
use_reserved_ref: Option<u64>,
|
|
) -> Result<(
|
|
ChannelService,
|
|
Sender<(u32, Vec<u8>)>,
|
|
Receiver<(u32, Option<Vec<u8>>)>,
|
|
)> {
|
|
let (input_sender, input_receiver) = channel(GROUPED_CHANNEL_QUEUE_LEN);
|
|
let (output_sender, output_receiver) = channel(GROUPED_CHANNEL_QUEUE_LEN);
|
|
|
|
debug!("opening xenevtchn");
|
|
let evtchn = EventChannelService::open().await?;
|
|
debug!("opening xenstore");
|
|
let store = XsdClient::open().await?;
|
|
debug!("opening xengnt");
|
|
let gnttab = GrantTab::open()?;
|
|
|
|
Ok((
|
|
ChannelService {
|
|
typ,
|
|
use_reserved_ref,
|
|
backends: HashMap::new(),
|
|
evtchn,
|
|
store,
|
|
gnttab,
|
|
input_sender: input_sender.clone(),
|
|
input_receiver,
|
|
output_sender,
|
|
},
|
|
input_sender,
|
|
output_receiver,
|
|
))
|
|
}
|
|
|
|
pub async fn launch(mut self) -> Result<JoinHandle<()>> {
|
|
Ok(tokio::task::spawn(async move {
|
|
if let Err(error) = self.process().await {
|
|
error!("channel processor failed: {}", error);
|
|
}
|
|
}))
|
|
}
|
|
|
|
async fn process(&mut self) -> Result<()> {
|
|
self.scan_all_backends().await?;
|
|
let mut watch_handle = self
|
|
.store
|
|
.create_watch("/local/domain/0/backend/console")
|
|
.await?;
|
|
self.store.bind_watch(&watch_handle).await?;
|
|
loop {
|
|
select! {
|
|
x = watch_handle.receiver.recv() => match x {
|
|
Some(_) => {
|
|
self.scan_all_backends().await?;
|
|
}
|
|
|
|
None => {
|
|
break;
|
|
}
|
|
},
|
|
|
|
x = self.input_receiver.recv() => match x {
|
|
Some((domid, data)) => {
|
|
if let Some(backend) = self.backends.get_mut(&domid) {
|
|
let _ = backend.sender.try_send(data);
|
|
}
|
|
},
|
|
|
|
None => {
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
Ok(())
|
|
}
|
|
|
|
pub async fn send(&mut self, domid: u32, message: Vec<u8>) -> Result<()> {
|
|
if let Some(backend) = self.backends.get(&domid) {
|
|
backend.sender.send(message).await?;
|
|
}
|
|
Ok(())
|
|
}
|
|
|
|
async fn ensure_backend_exists(&mut self, domid: u32, id: u32, path: String) -> Result<()> {
|
|
if self.backends.contains_key(&domid) {
|
|
return Ok(());
|
|
}
|
|
let Some(frontend_path) = self.store.read_string(format!("{}/frontend", path)).await?
|
|
else {
|
|
return Ok(());
|
|
};
|
|
let Some(typ) = self
|
|
.store
|
|
.read_string(format!("{}/type", frontend_path))
|
|
.await?
|
|
else {
|
|
return Ok(());
|
|
};
|
|
|
|
if typ != self.typ {
|
|
return Ok(());
|
|
}
|
|
|
|
let backend = ChannelBackend::new(
|
|
path.clone(),
|
|
frontend_path.clone(),
|
|
domid,
|
|
id,
|
|
self.store.clone(),
|
|
self.evtchn.clone(),
|
|
self.gnttab.clone(),
|
|
self.output_sender.clone(),
|
|
self.use_reserved_ref,
|
|
)
|
|
.await?;
|
|
self.backends.insert(domid, backend);
|
|
Ok(())
|
|
}
|
|
|
|
async fn scan_all_backends(&mut self) -> Result<()> {
|
|
let domains = self.store.list("/local/domain/0/backend/console").await?;
|
|
let mut seen: Vec<u32> = Vec::new();
|
|
for domid_string in &domains {
|
|
let domid = domid_string.parse::<u32>()?;
|
|
let domid_path = format!("/local/domain/0/backend/console/{}", domid);
|
|
for id_string in self.store.list(&domid_path).await? {
|
|
let id = id_string.parse::<u32>()?;
|
|
let console_path = format!(
|
|
"/local/domain/0/backend/console/{}/{}",
|
|
domid_string, id_string
|
|
);
|
|
self.ensure_backend_exists(domid, id, console_path).await?;
|
|
seen.push(domid);
|
|
}
|
|
}
|
|
|
|
let mut gone: Vec<u32> = Vec::new();
|
|
for backend in self.backends.keys() {
|
|
if !seen.contains(backend) {
|
|
gone.push(*backend);
|
|
}
|
|
}
|
|
|
|
for item in gone {
|
|
if let Some(backend) = self.backends.remove(&item) {
|
|
drop(backend);
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
pub struct ChannelBackend {
|
|
pub domid: u32,
|
|
pub id: u32,
|
|
pub sender: Sender<Vec<u8>>,
|
|
raw_sender: Sender<(u32, Option<Vec<u8>>)>,
|
|
task: JoinHandle<()>,
|
|
}
|
|
|
|
impl Drop for ChannelBackend {
|
|
fn drop(&mut self) {
|
|
self.task.abort();
|
|
let _ = self.raw_sender.try_send((self.domid, None));
|
|
debug!(
|
|
"destroyed channel backend for domain {} channel {}",
|
|
self.domid, self.id
|
|
);
|
|
}
|
|
}
|
|
|
|
impl ChannelBackend {
|
|
#[allow(clippy::too_many_arguments)]
|
|
pub async fn new(
|
|
backend: String,
|
|
frontend: String,
|
|
domid: u32,
|
|
id: u32,
|
|
store: XsdClient,
|
|
evtchn: EventChannelService,
|
|
gnttab: GrantTab,
|
|
output_sender: Sender<(u32, Option<Vec<u8>>)>,
|
|
use_reserved_ref: Option<u64>,
|
|
) -> Result<ChannelBackend> {
|
|
let processor = KrataChannelBackendProcessor {
|
|
backend,
|
|
frontend,
|
|
domid,
|
|
id,
|
|
store,
|
|
evtchn,
|
|
gnttab,
|
|
use_reserved_ref,
|
|
};
|
|
|
|
let (input_sender, input_receiver) = channel(SINGLE_CHANNEL_QUEUE_LEN);
|
|
|
|
let task = processor
|
|
.launch(output_sender.clone(), input_receiver)
|
|
.await?;
|
|
Ok(ChannelBackend {
|
|
domid,
|
|
id,
|
|
task,
|
|
raw_sender: output_sender,
|
|
sender: input_sender,
|
|
})
|
|
}
|
|
}
|
|
|
|
#[derive(Clone)]
|
|
pub struct KrataChannelBackendProcessor {
|
|
use_reserved_ref: Option<u64>,
|
|
backend: String,
|
|
frontend: String,
|
|
id: u32,
|
|
domid: u32,
|
|
store: XsdClient,
|
|
evtchn: EventChannelService,
|
|
gnttab: GrantTab,
|
|
}
|
|
|
|
impl KrataChannelBackendProcessor {
|
|
async fn init(&self) -> Result<()> {
|
|
self.store
|
|
.write_string(format!("{}/state", self.backend), "3")
|
|
.await?;
|
|
debug!(
|
|
"created channel backend for domain {} channel {}",
|
|
self.domid, self.id
|
|
);
|
|
Ok(())
|
|
}
|
|
|
|
async fn on_frontend_state_change(&self) -> Result<bool> {
|
|
let state = self
|
|
.store
|
|
.read_string(format!("{}/state", self.backend))
|
|
.await?
|
|
.unwrap_or("0".to_string())
|
|
.parse::<u32>()?;
|
|
if state == 3 {
|
|
return Ok(true);
|
|
}
|
|
Ok(false)
|
|
}
|
|
|
|
async fn on_self_state_change(&self) -> Result<bool> {
|
|
let state = self
|
|
.store
|
|
.read_string(format!("{}/state", self.backend))
|
|
.await?
|
|
.unwrap_or("0".to_string())
|
|
.parse::<u32>()?;
|
|
if state == 5 {
|
|
return Ok(true);
|
|
}
|
|
Ok(false)
|
|
}
|
|
|
|
async fn launch(
|
|
&self,
|
|
output_sender: Sender<(u32, Option<Vec<u8>>)>,
|
|
input_receiver: Receiver<Vec<u8>>,
|
|
) -> Result<JoinHandle<()>> {
|
|
let owned = self.clone();
|
|
Ok(tokio::task::spawn(async move {
|
|
if let Err(error) = owned.processor(output_sender, input_receiver).await {
|
|
error!("failed to process krata channel: {}", error);
|
|
}
|
|
let _ = owned
|
|
.store
|
|
.write_string(format!("{}/state", owned.backend), "6")
|
|
.await;
|
|
}))
|
|
}
|
|
|
|
async fn processor(
|
|
&self,
|
|
sender: Sender<(u32, Option<Vec<u8>>)>,
|
|
mut receiver: Receiver<Vec<u8>>,
|
|
) -> Result<()> {
|
|
self.init().await?;
|
|
let mut frontend_state_change = self
|
|
.store
|
|
.create_watch(format!("{}/state", self.frontend))
|
|
.await?;
|
|
self.store.bind_watch(&frontend_state_change).await?;
|
|
|
|
let (ring_ref, port) = loop {
|
|
match frontend_state_change.receiver.recv().await {
|
|
Some(_) => {
|
|
if self.on_frontend_state_change().await? {
|
|
let mut tries = 0;
|
|
let (ring_ref, port) = loop {
|
|
let ring_ref = self
|
|
.store
|
|
.read_string(format!("{}/ring-ref", self.frontend))
|
|
.await?;
|
|
let port = self
|
|
.store
|
|
.read_string(format!("{}/port", self.frontend))
|
|
.await?;
|
|
|
|
if (ring_ref.is_none() || port.is_none()) && tries < 40 {
|
|
tries += 1;
|
|
self.store
|
|
.write_string(format!("{}/state", self.backend), "4")
|
|
.await?;
|
|
sleep(Duration::from_millis(250)).await;
|
|
continue;
|
|
}
|
|
break (ring_ref, port);
|
|
};
|
|
|
|
if ring_ref.is_none() || port.is_none() {
|
|
return Err(anyhow!("frontend did not give ring-ref and port"));
|
|
}
|
|
|
|
let Ok(mut ring_ref) = ring_ref.unwrap().parse::<u64>() else {
|
|
return Err(anyhow!("frontend gave invalid ring-ref"));
|
|
};
|
|
|
|
let Ok(port) = port.unwrap().parse::<u32>() else {
|
|
return Err(anyhow!("frontend gave invalid port"));
|
|
};
|
|
|
|
ring_ref = self.use_reserved_ref.unwrap_or(ring_ref);
|
|
debug!(
|
|
"channel backend for domain {} channel {}: ring-ref={} port={}",
|
|
self.domid, self.id, ring_ref, port,
|
|
);
|
|
break (ring_ref, port);
|
|
}
|
|
}
|
|
|
|
None => {
|
|
return Ok(());
|
|
}
|
|
}
|
|
};
|
|
|
|
self.store
|
|
.write_string(format!("{}/state", self.backend), "4")
|
|
.await?;
|
|
let memory = self
|
|
.gnttab
|
|
.map_grant_refs(
|
|
vec![GrantRef {
|
|
domid: self.domid,
|
|
reference: ring_ref as u32,
|
|
}],
|
|
true,
|
|
true,
|
|
)
|
|
.map_err(|e| {
|
|
anyhow!(
|
|
"failed to map grant ref {} for domid {}: {}",
|
|
ring_ref,
|
|
self.domid,
|
|
e
|
|
)
|
|
})?;
|
|
let mut channel = self.evtchn.bind(self.domid, port).await?;
|
|
unsafe {
|
|
let buffer = self.read_output_buffer(channel.local_port, &memory).await?;
|
|
if !buffer.is_empty() {
|
|
sender.send((self.domid, Some(buffer))).await?;
|
|
}
|
|
};
|
|
|
|
let mut self_state_change = self
|
|
.store
|
|
.create_watch(format!("{}/state", self.backend))
|
|
.await?;
|
|
self.store.bind_watch(&self_state_change).await?;
|
|
loop {
|
|
select! {
|
|
x = self_state_change.receiver.recv() => match x {
|
|
Some(_) => {
|
|
match self.on_self_state_change().await {
|
|
Err(error) => {
|
|
error!("failed to process state change for domain {} channel {}: {}", self.domid, self.id, error);
|
|
},
|
|
|
|
Ok(stop) => {
|
|
if stop {
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
},
|
|
|
|
None => {
|
|
break;
|
|
}
|
|
},
|
|
|
|
x = receiver.recv() => match x {
|
|
Some(data) => {
|
|
let mut index = 0;
|
|
loop {
|
|
if index >= data.len() {
|
|
break;
|
|
}
|
|
let interface = memory.ptr() as *mut XenConsoleInterface;
|
|
let cons = unsafe { (*interface).in_cons };
|
|
let mut prod = unsafe { (*interface).in_prod };
|
|
fence(Ordering::Release);
|
|
let space = (prod - cons) as usize;
|
|
if space > XenConsoleInterface::INPUT_SIZE {
|
|
error!("channel for domid {} has an invalid input space of {}", self.domid, space);
|
|
}
|
|
let free = XenConsoleInterface::INPUT_SIZE.wrapping_sub(space);
|
|
if free == 0 {
|
|
sleep(Duration::from_micros(100)).await;
|
|
continue;
|
|
}
|
|
let want = data.len().min(free);
|
|
let buffer = &data[index..want];
|
|
for b in buffer {
|
|
unsafe { (*interface).input[prod as usize & (XenConsoleInterface::INPUT_SIZE - 1)] = *b; };
|
|
prod = prod.wrapping_add(1);
|
|
}
|
|
fence(Ordering::Release);
|
|
unsafe { (*interface).in_prod = prod; };
|
|
self.evtchn.notify(channel.local_port).await?;
|
|
index += want;
|
|
}
|
|
},
|
|
|
|
None => {
|
|
break;
|
|
}
|
|
},
|
|
|
|
x = channel.receiver.recv() => match x {
|
|
Some(_) => {
|
|
unsafe {
|
|
let buffer = self.read_output_buffer(channel.local_port, &memory).await?;
|
|
if !buffer.is_empty() {
|
|
sender.send((self.domid, Some(buffer))).await?;
|
|
}
|
|
};
|
|
channel.unmask().await?;
|
|
},
|
|
|
|
None => {
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
Ok(())
|
|
}
|
|
|
|
async unsafe fn read_output_buffer<'a>(
|
|
&self,
|
|
local_port: u32,
|
|
memory: &MappedMemory<'a>,
|
|
) -> Result<Vec<u8>> {
|
|
let interface = memory.ptr() as *mut XenConsoleInterface;
|
|
let mut cons = (*interface).out_cons;
|
|
let prod = (*interface).out_prod;
|
|
fence(Ordering::Release);
|
|
let size = prod.wrapping_sub(cons);
|
|
let mut data: Vec<u8> = Vec::new();
|
|
if size == 0 || size as usize > XenConsoleInterface::OUTPUT_SIZE {
|
|
return Ok(data);
|
|
}
|
|
loop {
|
|
if cons == prod {
|
|
break;
|
|
}
|
|
data.push((*interface).output[cons as usize & (XenConsoleInterface::OUTPUT_SIZE - 1)]);
|
|
cons = cons.wrapping_add(1);
|
|
}
|
|
fence(Ordering::AcqRel);
|
|
(*interface).out_cons = cons;
|
|
self.evtchn.notify(local_port).await?;
|
|
Ok(data)
|
|
}
|
|
}
|