mirror of
https://github.com/edera-dev/krata.git
synced 2025-08-02 12:50:54 +00:00
debugging for fixing page table issues
This commit is contained in:
parent
900eba0bd9
commit
dfc3dc8e90
@ -233,16 +233,16 @@ impl BootSetup<'_> {
|
||||
debug!("BootSetup initialize image_info={:?}", image_info);
|
||||
self.virt_alloc_end = image_info.virt_base;
|
||||
let kernel_segment = self.load_kernel_segment(image_loader, &image_info)?;
|
||||
let mut page_table = PageTable::default();
|
||||
let p2m_segment = self.alloc_p2m_segment(&mut page_table, &image_info)?;
|
||||
let start_info_segment = self.alloc_page()?;
|
||||
let xenstore_segment = self.alloc_page()?;
|
||||
let console_segment = self.alloc_page()?;
|
||||
let mut page_table = PageTable::default();
|
||||
let page_table_segment = self.alloc_page_tables(&mut page_table, &image_info)?;
|
||||
let boot_stack_segment = self.alloc_page()?;
|
||||
if self.virt_pgtab_end > 0 {
|
||||
self.alloc_padding_pages(self.virt_pgtab_end)?;
|
||||
}
|
||||
let p2m_segment = self.alloc_p2m_segment(&mut page_table, &image_info)?;
|
||||
let state = BootState {
|
||||
kernel_segment,
|
||||
start_info_segment,
|
||||
@ -338,8 +338,15 @@ impl BootSetup<'_> {
|
||||
let prot = self.get_pg_prot(lvl_idx, pfn, &state.page_table);
|
||||
|
||||
let pfn_paddr = self.phys.p2m[pfn as usize] << X86_PAGE_SHIFT;
|
||||
let value = pfn_paddr | prot;
|
||||
if pfn == state.page_table_segment.pfn {
|
||||
debug!(
|
||||
"pgtable pfn: {:#x}, p: {:#x}, pfn_paddr: {:#x}, value: {:#x}",
|
||||
pfn, p, pfn_paddr, value
|
||||
);
|
||||
}
|
||||
unsafe {
|
||||
*pg.add(p as usize) = pfn_paddr | prot;
|
||||
*pg.add(p as usize) = value;
|
||||
}
|
||||
pfn += 1;
|
||||
}
|
||||
|
@ -74,25 +74,28 @@ impl PhysicalPages<'_> {
|
||||
|
||||
fn pfn_alloc(&mut self, pfn: u64, count: u64) -> Result<u64, XenClientError> {
|
||||
let mut entries = vec![MmapEntry::default(); count as usize];
|
||||
for (i, entry) in (0_u64..).zip(entries.iter_mut()) {
|
||||
entry.mfn = self.p2m[(pfn + i) as usize];
|
||||
for (i, entry) in entries.iter_mut().enumerate() {
|
||||
entry.mfn = self.p2m[pfn as usize + i];
|
||||
}
|
||||
let chunk_size = 1 << XEN_PAGE_SHIFT;
|
||||
let num_per_entry = chunk_size >> XEN_PAGE_SHIFT;
|
||||
let num = num_per_entry * entries.len();
|
||||
let mut pfns = vec![0u64; num];
|
||||
for i in 0..entries.len() {
|
||||
let num = num_per_entry * count as usize;
|
||||
let mut pfns = vec![u64::MAX; num];
|
||||
for i in 0..count as usize {
|
||||
for j in 0..num_per_entry {
|
||||
pfns[i * num_per_entry + j] = entries[i].mfn + j as u64;
|
||||
}
|
||||
}
|
||||
|
||||
let size = count << XEN_PAGE_SHIFT;
|
||||
let actual_mmap_len = (num as u64) << XEN_PAGE_SHIFT;
|
||||
let addr = self
|
||||
.call
|
||||
.mmap(0, size)
|
||||
.mmap(0, actual_mmap_len)
|
||||
.ok_or(XenClientError::new("failed to mmap address"))?;
|
||||
self.call.mmap_batch(self.domid, num as u64, addr, pfns)?;
|
||||
let result = self.call.mmap_batch(self.domid, num as u64, addr, pfns)?;
|
||||
if result != 0 {
|
||||
return Err(XenClientError::new("mmap_batch call failed"));
|
||||
}
|
||||
let page = PhysicalPage {
|
||||
pfn,
|
||||
ptr: addr,
|
||||
|
Loading…
Reference in New Issue
Block a user