more hvm work

This commit is contained in:
Alex Zenla
2024-05-04 00:33:05 -07:00
parent a6d84d70f4
commit 2f69f339bc
6 changed files with 218 additions and 57 deletions

View File

@ -3,7 +3,7 @@ pub mod sys;
use crate::error::{Error, Result}; use crate::error::{Error, Result};
use crate::sys::{ use crate::sys::{
AddressSize, AssignDevice, CreateDomain, DomCtl, DomCtlValue, DomCtlVcpuContext, EvtChnAllocUnbound, GetDomainInfo, GetPageFrameInfo3, HvmParam, Hypercall, HypercallInit, IoMemPermission, IoPortPermission, IrqPermission, MaxMem, MaxVcpus, MemoryMap, MemoryReservation, MmapBatch, MmapResource, MmuExtOp, MultiCallEntry, PciAssignDevice, XenCapabilitiesInfo, DOMCTL_DEV_PCI, HYPERVISOR_DOMCTL, HYPERVISOR_EVENT_CHANNEL_OP, HYPERVISOR_HVM_OP, HYPERVISOR_MEMORY_OP, HYPERVISOR_MMUEXT_OP, HYPERVISOR_MULTICALL, HYPERVISOR_XEN_VERSION, XENVER_CAPABILITIES, XEN_DOMCTL_ASSIGN_DEVICE, XEN_DOMCTL_CREATEDOMAIN, XEN_DOMCTL_DESTROYDOMAIN, XEN_DOMCTL_GETDOMAININFO, XEN_DOMCTL_GETPAGEFRAMEINFO3, XEN_DOMCTL_HYPERCALL_INIT, XEN_DOMCTL_IOMEM_PERMISSION, XEN_DOMCTL_IOPORT_PERMISSION, XEN_DOMCTL_IRQ_PERMISSION, XEN_DOMCTL_MAX_MEM, XEN_DOMCTL_MAX_VCPUS, XEN_DOMCTL_PAUSEDOMAIN, XEN_DOMCTL_SETVCPUCONTEXT, XEN_DOMCTL_SET_ADDRESS_SIZE, XEN_DOMCTL_UNPAUSEDOMAIN, XEN_MEM_CLAIM_PAGES, XEN_MEM_MEMORY_MAP, XEN_MEM_POPULATE_PHYSMAP AddressSize, AssignDevice, CreateDomain, DomCtl, DomCtlValue, DomCtlVcpuContext, EvtChnAllocUnbound, GetDomainInfo, GetPageFrameInfo3, HvmContext, HvmParam, Hypercall, HypercallInit, IoMemPermission, IoPortPermission, IrqPermission, MaxMem, MaxVcpus, MemoryMap, MemoryReservation, MmapBatch, MmapResource, MmuExtOp, MultiCallEntry, PciAssignDevice, XenCapabilitiesInfo, DOMCTL_DEV_PCI, HYPERVISOR_DOMCTL, HYPERVISOR_EVENT_CHANNEL_OP, HYPERVISOR_HVM_OP, HYPERVISOR_MEMORY_OP, HYPERVISOR_MMUEXT_OP, HYPERVISOR_MULTICALL, HYPERVISOR_XEN_VERSION, XENVER_CAPABILITIES, XEN_DOMCTL_ASSIGN_DEVICE, XEN_DOMCTL_CREATEDOMAIN, XEN_DOMCTL_DESTROYDOMAIN, XEN_DOMCTL_GETDOMAININFO, XEN_DOMCTL_GETHVMCONTEXT, XEN_DOMCTL_GETPAGEFRAMEINFO3, XEN_DOMCTL_HYPERCALL_INIT, XEN_DOMCTL_IOMEM_PERMISSION, XEN_DOMCTL_IOPORT_PERMISSION, XEN_DOMCTL_IRQ_PERMISSION, XEN_DOMCTL_MAX_MEM, XEN_DOMCTL_MAX_VCPUS, XEN_DOMCTL_PAUSEDOMAIN, XEN_DOMCTL_SETHVMCONTEXT, XEN_DOMCTL_SETVCPUCONTEXT, XEN_DOMCTL_SET_ADDRESS_SIZE, XEN_DOMCTL_UNPAUSEDOMAIN, XEN_MEM_CLAIM_PAGES, XEN_MEM_MEMORY_MAP, XEN_MEM_POPULATE_PHYSMAP
}; };
use libc::{c_int, mmap, usleep, MAP_FAILED, MAP_SHARED, PROT_READ, PROT_WRITE}; use libc::{c_int, mmap, usleep, MAP_FAILED, MAP_SHARED, PROT_READ, PROT_WRITE};
use log::trace; use log::trace;
@ -17,7 +17,7 @@ use tokio::sync::Semaphore;
use std::fs::{File, OpenOptions}; use std::fs::{File, OpenOptions};
use std::os::fd::AsRawFd; use std::os::fd::AsRawFd;
use std::ptr::addr_of_mut; use std::ptr::{addr_of_mut, null_mut};
use std::slice; use std::slice;
#[derive(Clone)] #[derive(Clone)]
@ -819,4 +819,48 @@ impl XenCall {
.await?; .await?;
Ok(()) Ok(())
} }
pub async fn get_hvm_context(&self, domid: u32, buffer: Option<&mut [u8]>) -> Result<u32> {
trace!(
"domctl fd={} get_hvm_context domid={}",
self.handle.as_raw_fd(),
domid,
);
let mut domctl = DomCtl {
cmd: XEN_DOMCTL_GETHVMCONTEXT,
interface_version: self.domctl_interface_version,
domid,
value: DomCtlValue {
hvm_context: HvmContext {
size: buffer.as_ref().map(|x| x.len()).unwrap_or(0) as u32,
buffer: buffer.map(|x| x.as_mut_ptr()).unwrap_or(null_mut()) as u64,
}
},
};
self.hypercall1(HYPERVISOR_DOMCTL, addr_of_mut!(domctl) as c_ulong)
.await?;
Ok(unsafe { domctl.value.hvm_context.size })
}
pub async fn set_hvm_context(&self, domid: u32, buffer: &mut [u8]) -> Result<u32> {
trace!(
"domctl fd={} set_hvm_context domid={}",
self.handle.as_raw_fd(),
domid,
);
let mut domctl = DomCtl {
cmd: XEN_DOMCTL_SETHVMCONTEXT,
interface_version: self.domctl_interface_version,
domid,
value: DomCtlValue {
hvm_context: HvmContext {
size: buffer.len() as u32,
buffer: buffer.as_ptr() as u64,
}
},
};
self.hypercall1(HYPERVISOR_DOMCTL, addr_of_mut!(domctl) as c_ulong)
.await?;
Ok(unsafe { domctl.value.hvm_context.size })
}
} }

View File

@ -242,6 +242,7 @@ pub union DomCtlValue {
pub iomem_permission: IoMemPermission, pub iomem_permission: IoMemPermission,
pub irq_permission: IrqPermission, pub irq_permission: IrqPermission,
pub assign_device: AssignDevice, pub assign_device: AssignDevice,
pub hvm_context: HvmContext,
pub pad: [u8; 128], pub pad: [u8; 128],
} }
@ -491,6 +492,7 @@ pub struct TrapInfo {
#[repr(C)] #[repr(C)]
#[derive(Copy, Clone, Debug)] #[derive(Copy, Clone, Debug)]
#[allow(non_camel_case_types)]
pub struct x8664VcpuGuestContext { pub struct x8664VcpuGuestContext {
pub fpu_ctx: VcpuGuestContextFpuCtx, pub fpu_ctx: VcpuGuestContextFpuCtx,
pub flags: u64, pub flags: u64,
@ -689,3 +691,10 @@ pub struct HvmParam {
pub index: u32, pub index: u32,
pub value: u64, pub value: u64,
} }
#[repr(C)]
#[derive(Clone, Copy, Debug)]
pub struct HvmContext {
pub size: u32,
pub buffer: u64,
}

View File

@ -203,7 +203,7 @@ impl<I: BootImageLoader, P: BootSetupPlatform> BootSetup<I, P> {
} }
domain.initrd_segment = domain.initrd_segment =
initrd_segment.ok_or(Error::MemorySetupFailed("initd_segment missing"))?; initrd_segment.ok_or(Error::MemorySetupFailed("initrd_segment missing"))?;
domain.store_evtchn = self.call.evtchn_alloc_unbound(self.domid, 0).await?; domain.store_evtchn = self.call.evtchn_alloc_unbound(self.domid, 0).await?;
let _kernel_segment = let _kernel_segment =

View File

@ -326,6 +326,8 @@ impl XenClient {
.await?; .await?;
tx.write_string(format!("{}/domid", dom_path).as_str(), &domid.to_string()) tx.write_string(format!("{}/domid", dom_path).as_str(), &domid.to_string())
.await?; .await?;
tx.write_string(format!("{}/type", dom_path).as_str(), "PVH")
.await?;
tx.write_string( tx.write_string(
format!("{}/store/port", dom_path).as_str(), format!("{}/store/port", dom_path).as_str(),
&xenstore_evtchn.to_string(), &xenstore_evtchn.to_string(),

View File

@ -1,7 +1,7 @@
use crate::error::Result; use crate::error::Result;
use crate::sys::{XEN_PAGE_SHIFT, XEN_PAGE_SIZE}; use crate::sys::{XEN_PAGE_SHIFT, XEN_PAGE_SIZE};
use crate::Error; use crate::Error;
use libc::munmap; use libc::{memset, munmap};
use log::debug; use log::debug;
use nix::errno::Errno; use nix::errno::Errno;
use std::ffi::c_void; use std::ffi::c_void;
@ -125,7 +125,7 @@ impl PhysicalPages {
} }
pub async fn map_foreign_pages(&mut self, mfn: u64, size: u64) -> Result<PhysicalPage> { pub async fn map_foreign_pages(&mut self, mfn: u64, size: u64) -> Result<PhysicalPage> {
let num = ((size + XEN_PAGE_SIZE - 1) >> XEN_PAGE_SHIFT) as usize; let num = (size >> XEN_PAGE_SHIFT) as usize;
let mut pfns = vec![u64::MAX; num]; let mut pfns = vec![u64::MAX; num];
for (i, item) in pfns.iter_mut().enumerate().take(num) { for (i, item) in pfns.iter_mut().enumerate().take(num) {
*item = mfn + i as u64; *item = mfn + i as u64;
@ -165,9 +165,8 @@ impl PhysicalPages {
pfn pfn
}; };
let page = self.map_foreign_pages(mfn, count << XEN_PAGE_SHIFT).await?; let page = self.map_foreign_pages(mfn, count << XEN_PAGE_SHIFT).await?;
let _slice = unsafe { slice::from_raw_parts_mut(page.ptr as *mut u8, (count * XEN_PAGE_SIZE) as usize) }; let slice = unsafe { slice::from_raw_parts_mut(page.ptr as *mut u8, (count << XEN_PAGE_SHIFT) as usize) };
// slice.fill(0); slice.fill(0);
self.unmap(pfn)?;
Ok(()) Ok(())
} }

View File

@ -1,7 +1,5 @@
use std::{ use std::{
mem::size_of, mem::{size_of, MaybeUninit}, os::raw::{c_char, c_void}, ptr::addr_of_mut, slice
os::raw::{c_char, c_void},
slice,
}; };
use libc::munmap; use libc::munmap;
@ -115,23 +113,141 @@ pub struct HvmMemmapTableEntry {
pub reserved: u32, pub reserved: u32,
} }
const HVMLOADER_MODULE_MAX_COUNT: u32 = 2; #[repr(C)]
#[derive(Copy, Clone, Debug)]
struct HvmSaveDescriptor {
pub typecode: u16,
pub instance: u16,
pub length: u32,
}
#[repr(C)]
#[derive(Default, Copy, Clone, Debug)]
struct HvmSaveHeader {
magic: u32,
version: u32,
changeset: u64,
cpuid: u32,
gtsc_khz: u32,
}
#[repr(C)]
#[derive(Copy, Clone, Debug)]
struct HvmCpu {
pub fpu_regs: [u8; 512],
pub rax: u64,
pub rbx: u64,
pub rcx: u64,
pub rdx: u64,
pub rbp: u64,
pub rsi: u64,
pub rdi: u64,
pub rsp: u64,
pub r8: u64,
pub r9: u64,
pub r10: u64,
pub r11: u64,
pub r12: u64,
pub r13: u64,
pub r14: u64,
pub r15: u64,
pub rip: u64,
pub rflags: u64,
pub cr0: u64,
pub cr2: u64,
pub cr3: u64,
pub cr4: u64,
pub dr0: u64,
pub dr1: u64,
pub dr2: u64,
pub dr3: u64,
pub dr6: u64,
pub dr7: u64,
pub cs_sel: u32,
pub ds_sel: u32,
pub es_sel: u32,
pub fs_sel: u32,
pub gs_sel: u32,
pub ss_sel: u32,
pub tr_sel: u32,
pub ldtr_sel: u32,
pub cs_limit: u32,
pub ds_limit: u32,
pub es_limit: u32,
pub fs_limit: u32,
pub gs_limit: u32,
pub ss_limit: u32,
pub tr_limit: u32,
pub ldtr_limit: u32,
pub idtr_limit: u32,
pub gdtr_limit: u32,
pub cs_base: u64,
pub ds_base: u64,
pub es_base: u64,
pub fs_base: u64,
pub gs_base: u64,
pub ss_base: u64,
pub tr_base: u64,
pub ldtr_base: u64,
pub idtr_base: u64,
pub gdtr_base: u64,
pub cs_arbytes: u32,
pub ds_arbytes: u32,
pub es_arbytes: u32,
pub fs_arbytes: u32,
pub gs_arbytes: u32,
pub ss_arbytes: u32,
pub tr_arbytes: u32,
pub ldtr_arbytes: u32,
pub sysenter_cs: u64,
pub sysenter_esp: u64,
pub sysenter_eip: u64,
pub shadow_gs: u64,
pub msr_flags: u64,
pub msr_lstar: u64,
pub msr_star: u64,
pub msr_cstar: u64,
pub msr_syscall_mask: u64,
pub msr_efer: u64,
pub msr_tsc_aux: u64,
pub tsc: u64,
pub pending_event: u32,
pub error_code: u32,
pub flags: u32,
pub pad0: u32,
}
#[repr(C)]
#[derive(Copy, Clone, Debug)]
struct HvmEnd {}
#[repr(C)]
#[derive(Copy, Clone, Debug)]
struct BspCtx {
header_d: HvmSaveDescriptor,
header: HvmSaveHeader,
cpu_d: HvmSaveDescriptor,
cpu: HvmCpu,
end_d: HvmSaveDescriptor,
end: HvmEnd,
}
#[derive(Debug)] #[derive(Debug)]
struct VmemRange { struct VmemRange {
start: u64, start: u64,
end: u64, end: u64,
_flags: u32, _flags: u32,
nid: u32, _nid: u32,
} }
#[derive(Default)] #[derive(Default)]
pub struct X86PvhPlatform { pub struct X86PvhPlatform {
start_info_segment: Option<DomainSegment>, start_info_segment: Option<DomainSegment>,
boot_stack_segment: Option<DomainSegment>,
xenstore_segment: Option<DomainSegment>,
} }
const X86_CR0_PE: u64 = 0x01;
const X86_CR0_ET: u64 = 0x10;
impl X86PvhPlatform { impl X86PvhPlatform {
pub fn new() -> Self { pub fn new() -> Self {
Self { Self {
@ -188,7 +304,7 @@ impl BootSetupPlatform for X86PvhPlatform {
start: 0, start: 0,
end: domain.total_pages << self.page_shift(), end: domain.total_pages << self.page_shift(),
_flags: 0, _flags: 0,
nid: 0, _nid: 0,
}; };
vmemranges.push(stub); vmemranges.push(stub);
@ -257,7 +373,8 @@ impl BootSetupPlatform for X86PvhPlatform {
for i in 0..X86_HVM_NR_SPECIAL_PAGES { for i in 0..X86_HVM_NR_SPECIAL_PAGES {
special_array[i as usize] = special_pfn(i as u32); special_array[i as usize] = special_pfn(i as u32);
} }
let _pages = domain.call.populate_physmap(domain.domid, X86_HVM_NR_SPECIAL_PAGES, 0, 0, &special_array).await?; let pages = domain.call.populate_physmap(domain.domid, X86_HVM_NR_SPECIAL_PAGES, 0, 0, &special_array).await?;
println!("{:?}", pages);
domain.phys.clear_pages(special_pfn(0), X86_HVM_NR_SPECIAL_PAGES).await?; domain.phys.clear_pages(special_pfn(0), X86_HVM_NR_SPECIAL_PAGES).await?;
domain.call.set_hvm_param(domain.domid, HVM_PARAM_STORE_PFN, special_pfn(SPECIALPAGE_XENSTORE)).await?; domain.call.set_hvm_param(domain.domid, HVM_PARAM_STORE_PFN, special_pfn(SPECIALPAGE_XENSTORE)).await?;
domain.call.set_hvm_param(domain.domid, HVM_PARAM_BUFIOREQ_PFN, special_pfn(SPECIALPAGE_BUFIOREQ)).await?; domain.call.set_hvm_param(domain.domid, HVM_PARAM_BUFIOREQ_PFN, special_pfn(SPECIALPAGE_BUFIOREQ)).await?;
@ -296,54 +413,44 @@ impl BootSetupPlatform for X86PvhPlatform {
Ok(()) Ok(())
} }
async fn bootlate(&mut self, domain: &mut BootDomain) -> Result<()> { async fn bootlate(&mut self, _: &mut BootDomain) -> Result<()> {
Ok(()) Ok(())
} }
async fn vcpu(&mut self, domain: &mut BootDomain) -> Result<()> { async fn vcpu(&mut self, domain: &mut BootDomain) -> Result<()> {
let boot_stack_segment = self let size = domain.call.get_hvm_context(domain.domid, None).await?;
.boot_stack_segment let mut full_context = vec![0u8; size as usize];
.as_ref() domain.call.get_hvm_context(domain.domid, Some(&mut full_context)).await?;
.ok_or(Error::MemorySetupFailed("boot_stack_segment missing"))?; let mut ctx: BspCtx = unsafe { MaybeUninit::zeroed().assume_init() };
let start_info_segment = self unsafe { std::ptr::copy(full_context.as_ptr(), addr_of_mut!(ctx) as *mut u8, size_of::<HvmSaveDescriptor>() + size_of::<HvmSaveHeader>()) };
.start_info_segment ctx.cpu_d.instance = 0;
.as_ref() ctx.cpu.cs_base = 0;
.ok_or(Error::MemorySetupFailed("start_info_segment missing"))?; ctx.cpu.ds_base = 0;
let pg_pfn = 0; ctx.cpu.es_base = 0;
let pg_mfn = domain.phys.p2m[pg_pfn as usize]; ctx.cpu.ss_base = 0;
let mut vcpu = x8664VcpuGuestContext::default(); ctx.cpu.tr_base = 0;
vcpu.user_regs.rip = domain.image_info.virt_entry; ctx.cpu.cs_limit = !0;
vcpu.user_regs.rsp = ctx.cpu.ds_limit = !0;
domain.image_info.virt_base + (boot_stack_segment.pfn + 1) * self.page_size(); ctx.cpu.es_limit = !0;
vcpu.user_regs.rsi = ctx.cpu.ss_limit = !0;
domain.image_info.virt_base + (start_info_segment.pfn) * self.page_size(); ctx.cpu.tr_limit = 0x67;
vcpu.user_regs.rflags = 1 << 9; ctx.cpu.cs_arbytes = 0xc9b;
vcpu.debugreg[6] = 0xffff0ff0; ctx.cpu.ds_arbytes = 0xc93;
vcpu.debugreg[7] = 0x00000400; ctx.cpu.es_arbytes = 0xc93;
vcpu.flags = VGCF_IN_KERNEL | VGCF_ONLINE; ctx.cpu.ss_arbytes = 0xc93;
let cr3_pfn = pg_mfn; ctx.cpu.tr_arbytes = 0x8b;
vcpu.ctrlreg[3] = cr3_pfn << 12; ctx.cpu.cr0 = X86_CR0_PE | X86_CR0_ET;
vcpu.user_regs.ds = 0x0; ctx.cpu.rip = domain.image_info.virt_entry;
vcpu.user_regs.es = 0x0; ctx.cpu.dr6 = 0xffff0ff0;
vcpu.user_regs.fs = 0x0; ctx.cpu.dr7 = 0x00000400;
vcpu.user_regs.gs = 0x0; let addr = addr_of_mut!(ctx) as *mut u8;
vcpu.user_regs.ss = 0xe02b; let slice = unsafe { std::slice::from_raw_parts_mut(addr, size_of::<BspCtx>()) };
vcpu.user_regs.cs = 0xe033; domain.call.set_hvm_context(domain.domid, slice).await?;
vcpu.kernel_ss = vcpu.user_regs.ss as u64;
vcpu.kernel_sp = vcpu.user_regs.rsp;
trace!("vcpu context: {:?}", vcpu);
domain.call.set_vcpu_context(domain.domid, 0, xencall::sys::VcpuGuestContextAny { value: vcpu }).await?;
Ok(()) Ok(())
} }
async fn gnttab_seed(&mut self, domain: &mut BootDomain) -> Result<()> { async fn gnttab_seed(&mut self, domain: &mut BootDomain) -> Result<()> {
let xenstore_segment = self
.xenstore_segment
.as_ref()
.ok_or(Error::MemorySetupFailed("xenstore_segment missing"))?;
let console_gfn = domain.consoles.first().map(|x| x.1).unwrap_or(0) as usize; let console_gfn = domain.consoles.first().map(|x| x.1).unwrap_or(0) as usize;
let xenstore_gfn = domain.phys.p2m[xenstore_segment.pfn as usize];
let addr = domain let addr = domain
.call .call
.mmap(0, 1 << XEN_PAGE_SHIFT) .mmap(0, 1 << XEN_PAGE_SHIFT)
@ -359,7 +466,7 @@ impl BootSetupPlatform for X86PvhPlatform {
entries[0].frame = console_gfn as u32; entries[0].frame = console_gfn as u32;
entries[1].flags = 1 << 0; entries[1].flags = 1 << 0;
entries[1].domid = 0; entries[1].domid = 0;
entries[1].frame = xenstore_gfn as u32; entries[1].frame = domain.xenstore_mfn as u32;
unsafe { unsafe {
let result = munmap(addr as *mut c_void, 1 << XEN_PAGE_SHIFT); let result = munmap(addr as *mut c_void, 1 << XEN_PAGE_SHIFT);
if result != 0 { if result != 0 {