1
0
mirror of https://github.com/rcore-os/rCore.git synced 2024-11-23 08:26:17 +04:00

Merge branch 'memory' into process

# Conflicts:
#	src/arch/x86_64/smp.rs
#	src/memory/mod.rs
This commit is contained in:
WangRunji 2018-04-28 12:59:58 +08:00
commit 5707dfd00a
31 changed files with 848 additions and 159 deletions

6
crate/memory/Cargo.toml Normal file
View File

@ -0,0 +1,6 @@
[package]
name = "memory"
version = "0.1.0"
authors = ["WangRunji <wangrunji0408@163.com>"]
[dependencies]

12
crate/memory/src/lib.rs Normal file
View File

@ -0,0 +1,12 @@
#![no_std]
#![feature(alloc)]
extern crate alloc;
pub mod physical;
pub mod paging;
pub mod memory_set;
pub mod swap;
type VirtAddr = usize;
const PAGE_SIZE: usize = 4096;

View File

@ -0,0 +1,89 @@
use alloc::vec::Vec;
type Addr = usize;
/// 一片连续内存空间,有相同的访问权限
/// 对应ucore中 `vma_struct`
#[derive(Debug, Eq, PartialEq)]
pub struct MemoryArea {
pub start_addr: Addr,
pub end_addr: Addr,
pub flags: u32,
pub name: &'static str,
}
impl MemoryArea {
pub fn contains(&self, addr: Addr) -> bool {
addr >= self.start_addr && addr < self.end_addr
}
fn is_overlap_with(&self, other: &MemoryArea) -> bool {
!(self.end_addr <= other.start_addr || self.start_addr >= other.end_addr)
}
}
/// 内存空间集合,包含若干段连续空间
/// 对应ucore中 `mm_struct`
#[derive(Debug)]
pub struct MemorySet {
areas: Vec<MemoryArea>,
}
impl MemorySet {
pub fn new() -> Self {
MemorySet { areas: Vec::<MemoryArea>::new() }
}
pub fn find_area(&self, addr: Addr) -> Option<&MemoryArea> {
self.areas.iter().find(|area| area.contains(addr))
}
pub fn push(&mut self, area: MemoryArea) {
assert!(area.start_addr <= area.end_addr, "invalid memory area");
if self.areas.iter()
.find(|other| area.is_overlap_with(other))
.is_some() {
panic!("memory area overlap");
}
self.areas.push(area);
}
}
#[cfg(test)]
mod test {
use super::*;
#[test]
fn push_and_find() {
let mut ms = MemorySet::new();
ms.push(MemoryArea {
start_addr: 0x0,
end_addr: 0x8,
flags: 0x0,
name: "code",
});
ms.push(MemoryArea {
start_addr: 0x8,
end_addr: 0x10,
flags: 0x1,
name: "data",
});
assert_eq!(ms.find_area(0x6).unwrap().name, "code");
assert_eq!(ms.find_area(0x11), None);
}
#[test]
#[should_panic]
fn push_overlap() {
let mut ms = MemorySet::new();
ms.push(MemoryArea {
start_addr: 0x0,
end_addr: 0x8,
flags: 0x0,
name: "code",
});
ms.push(MemoryArea {
start_addr: 0x4,
end_addr: 0x10,
flags: 0x1,
name: "data",
});
}
}

View File

@ -0,0 +1,4 @@
pub use self::page_table::*;
use super::*;
mod page_table;

View File

@ -0,0 +1,130 @@
use alloc::boxed::Box;
use super::*;
const PAGE_COUNT: usize = 16;
const PAGE_SIZE: usize = 4096;
pub struct MockPageTable {
mapped: [bool; PAGE_COUNT],
accessed: [bool; PAGE_COUNT],
dirty: [bool; PAGE_COUNT],
data: [u8; PAGE_SIZE * PAGE_COUNT],
page_fault_handler: Option<PageFaultHandler>,
capacity: usize,
}
type PageFaultHandler = Box<FnMut(&mut MockPageTable, VirtAddr)>;
impl PageTable for MockPageTable {
fn accessed(&self, addr: VirtAddr) -> bool {
self.accessed[addr / PAGE_SIZE]
}
fn dirty(&self, addr: VirtAddr) -> bool {
self.dirty[addr / PAGE_SIZE]
}
fn clear_accessed(&mut self, addr: usize) {
self.accessed[addr / PAGE_SIZE] = false;
}
fn clear_dirty(&mut self, addr: usize) {
self.dirty[addr / PAGE_SIZE] = false;
}
/// Map a page, return false if no more space
fn map(&mut self, addr: VirtAddr) -> bool {
if self.mapped.iter().filter(|&&b| b).count() == self.capacity {
return false;
}
self.mapped[addr / PAGE_SIZE] = true;
true
}
fn unmap(&mut self, addr: VirtAddr) {
self.mapped[addr / PAGE_SIZE] = false;
}
}
impl MockPageTable {
pub fn new(capacity: usize) -> Self {
use core::mem::uninitialized;
MockPageTable {
mapped: [false; PAGE_COUNT],
accessed: [false; PAGE_COUNT],
dirty: [false; PAGE_COUNT],
data: unsafe{ uninitialized() },
page_fault_handler: None,
capacity,
}
}
pub fn set_handler(&mut self, page_fault_handler: PageFaultHandler) {
self.page_fault_handler = Some(page_fault_handler);
}
fn trigger_page_fault_if_not_present(&mut self, addr: VirtAddr) {
let page_id = addr / PAGE_SIZE;
while !self.mapped[page_id] {
let self_mut = unsafe{ &mut *(self as *mut Self) };
(self.page_fault_handler.as_mut().unwrap())(self_mut, addr);
}
}
/// Read memory, mark accessed, trigger page fault if not present
pub fn read(&mut self, addr: VirtAddr) -> u8 {
let page_id = addr / PAGE_SIZE;
self.trigger_page_fault_if_not_present(addr);
self.accessed[page_id] = true;
self.data[addr]
}
/// Write memory, mark accessed and dirty, trigger page fault if not present
pub fn write(&mut self, addr: VirtAddr, data: u8) {
let page_id = addr / PAGE_SIZE;
self.trigger_page_fault_if_not_present(addr);
self.accessed[page_id] = true;
self.dirty[page_id] = true;
self.data[addr] = data;
}
}
#[cfg(test)]
mod test {
use super::*;
use alloc::arc::Arc;
use core::cell::RefCell;
#[test]
fn test() {
let page_fault_count = Arc::new(RefCell::new(0usize));
let mut pt = MockPageTable::new(2);
pt.set_handler(Box::new({
let page_fault_count1 = page_fault_count.clone();
move |pt: &mut MockPageTable, addr: VirtAddr| {
*page_fault_count1.borrow_mut() += 1;
pt.map(addr);
}
}));
pt.map(0);
pt.read(0);
assert_eq!(*page_fault_count.borrow(), 0);
assert!(pt.accessed(0));
assert!(!pt.dirty(0));
pt.clear_accessed(0);
assert!(!pt.accessed(0));
pt.read(1);
assert_eq!(*page_fault_count.borrow(), 0);
assert!(pt.accessed(0));
pt.write(0x1000, 0xff);
assert_eq!(*page_fault_count.borrow(), 1);
assert!(pt.accessed(0x1000));
assert!(pt.dirty(0x1000));
assert_eq!(pt.read(0x1000), 0xff);
pt.clear_dirty(0x1000);
assert!(!pt.dirty(0x1000));
assert_eq!(pt.map(0x2000), false);
pt.unmap(0);
pt.read(0);
assert_eq!(*page_fault_count.borrow(), 2);
}
}

View File

@ -0,0 +1,13 @@
use super::*;
pub use self::mock_page_table::MockPageTable;
mod mock_page_table;
pub trait PageTable {
fn accessed(&self, addr: VirtAddr) -> bool;
fn dirty(&self, addr: VirtAddr) -> bool;
fn clear_accessed(&mut self, addr: VirtAddr);
fn clear_dirty(&mut self, addr: VirtAddr);
fn map(&mut self, addr: VirtAddr) -> bool;
fn unmap(&mut self, addr: VirtAddr);
}

View File

@ -0,0 +1,33 @@
use super::*;
#[derive(Debug, PartialEq, Eq, PartialOrd, Ord)]
pub struct Frame {
number: usize,
}
impl Frame {
pub fn containing_address(address: PhysAddr) -> Frame {
Frame{ number: address.get() as usize / PAGE_SIZE }
}
//TODO: Set private
pub fn start_address(&self) -> PhysAddr {
PhysAddr::new((self.number * PAGE_SIZE) as u64)
}
pub fn clone(&self) -> Frame {
Frame { number: self.number }
}
//TODO: Set private
// pub fn range_inclusive(start: Frame, end: Frame) -> FrameIter {
// FrameIter {
// start: start,
// end: end,
// }
// }
}
impl Drop for Frame {
fn drop(&mut self) {
panic!("frame must be deallocate");
}
}

View File

@ -0,0 +1,11 @@
use super::*;
pub trait FrameAllocator {
fn allocate_frame(&mut self) -> Option<Frame>;
fn deallocate_frame(&mut self, frame: Frame);
}
pub trait MemoryArea {
fn begin(&self) -> PhysAddr;
fn end(&self) -> PhysAddr;
}

View File

@ -0,0 +1,9 @@
pub use self::physaddr::PhysAddr;
pub use self::frame::Frame;
pub use self::frame_allocator::FrameAllocator;
use super::*;
mod frame;
mod physaddr;
mod frame_allocator;

View File

@ -0,0 +1,50 @@
use core::fmt;
/// Represents a physical memory address
#[derive(Copy, Clone, Eq, Ord, PartialEq, PartialOrd)]
pub struct PhysAddr(u64);
impl PhysAddr {
pub fn new(addr: u64) -> PhysAddr {
PhysAddr(addr)
}
pub fn get(&self) -> u64 {
self.0
}
}
impl fmt::Debug for PhysAddr {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
write!(f, "{:#x}", self.0)
}
}
impl fmt::Binary for PhysAddr {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
self.0.fmt(f)
}
}
impl fmt::Display for PhysAddr {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
self.0.fmt(f)
}
}
impl fmt::LowerHex for PhysAddr {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
self.0.fmt(f)
}
}
impl fmt::Octal for PhysAddr {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
self.0.fmt(f)
}
}
impl fmt::UpperHex for PhysAddr {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
self.0.fmt(f)
}
}

View File

@ -0,0 +1,135 @@
use alloc::vec_deque::VecDeque;
use super::*;
pub struct EnhancedClockSwapManager<T: 'static + SwappablePageTable> {
page_table: &'static mut T,
clock_ptr: usize,
deque: VecDeque<VirtAddr>,
}
impl<T: 'static + SwappablePageTable> SwapManager for EnhancedClockSwapManager<T> {
fn tick(&mut self) {
}
fn push(&mut self, addr: usize) {
let pos = if self.clock_ptr == 0 {self.deque.len()} else {self.clock_ptr};
self.deque.insert(pos, addr);
}
fn remove(&mut self, addr: usize) {
let id = self.deque.iter()
.position(|&x| x == addr)
.expect("address not found");
if id < self.clock_ptr {
self.clock_ptr -= 1;
}
self.deque.remove(id);
}
fn pop(&mut self) -> Option<usize> {
loop {
let addr = self.deque[self.clock_ptr];
let accessed = self.page_table.accessed(addr);
let dirty = self.page_table.dirty(addr);
match (accessed, dirty) {
(true, _) => {
self.page_table.clear_accessed(addr);
},
(false, true) => {
if self.page_table.swap_out(addr).is_ok() {
self.page_table.clear_dirty(addr);
}
},
_ => {
return self.remove_current();
}
}
self.move_next();
}
}
}
impl<T: 'static + SwappablePageTable> EnhancedClockSwapManager<T> {
pub fn new(page_table: &'static mut T) -> Self {
EnhancedClockSwapManager {
page_table,
clock_ptr: 0,
deque: VecDeque::<VirtAddr>::new()
}
}
fn remove_current(&mut self) -> Option<VirtAddr> {
let addr = self.deque.remove(self.clock_ptr);
if self.clock_ptr == self.deque.len() {
self.clock_ptr = 0;
}
return addr;
}
fn move_next(&mut self) {
self.clock_ptr += 1;
if self.clock_ptr == self.deque.len() {
self.clock_ptr = 0;
}
}
}
#[cfg(test)]
mod test {
use super::*;
use alloc::{arc::Arc, boxed::Box};
use core::mem::uninitialized;
use core::cell::RefCell;
use paging::MockPageTable;
impl SwappablePageTable for MockPageTable {
fn swap_out(&mut self, addr: usize) -> Result<(), ()> {
Ok(())
}
}
enum MemOp {
R(usize), W(usize)
}
#[test]
fn test() {
use self::MemOp::{R, W};
let page_fault_count = Arc::new(RefCell::new(0usize));
let mut pt = Box::new(MockPageTable::new(4));
let static_pt = unsafe{ &mut *(pt.as_mut() as *mut MockPageTable) };
pt.set_handler(Box::new({
let page_fault_count1 = page_fault_count.clone();
let mut clock = EnhancedClockSwapManager::new(static_pt);
move |pt: &mut MockPageTable, addr: VirtAddr| {
*page_fault_count1.borrow_mut() += 1;
if !pt.map(addr) { // is full?
pt.unmap(clock.pop().unwrap());
pt.map(addr);
}
clock.push(addr);
}
}));
let op_seq = [
R(0x1000), R(0x2000), R(0x3000), R(0x4000),
R(0x3000), W(0x1000), R(0x4000), W(0x2000), R(0x5000),
R(0x2000), W(0x1000), R(0x2000), R(0x3000), R(0x4000)];
let pgfault_count = [
1, 2, 3, 4,
4, 4, 4, 4, 5,
5, 5, 5, 6, 7];
for (op, &count) in op_seq.iter().zip(pgfault_count.iter()) {
match op {
R(addr) => {pt.read(*addr);},
W(addr) => pt.write(*addr, 0),
}
assert_eq!(*(*page_fault_count).borrow(), count);
}
}
}

View File

@ -0,0 +1,87 @@
use alloc::vec_deque::VecDeque;
use super::*;
pub struct FifoSwapManager {
deque: VecDeque<VirtAddr>,
}
impl SwapManager for FifoSwapManager {
fn tick(&mut self) {
}
fn push(&mut self, addr: usize) {
self.deque.push_back(addr);
}
fn remove(&mut self, addr: usize) {
let id = self.deque.iter()
.position(|&x| x == addr)
.expect("address not found");
self.deque.remove(id);
}
fn pop(&mut self) -> Option<VirtAddr> {
self.deque.pop_front()
}
}
impl FifoSwapManager {
pub fn new() -> Self {
FifoSwapManager {
deque: VecDeque::<VirtAddr>::new()
}
}
}
#[cfg(test)]
mod test {
use super::*;
use alloc::{arc::Arc, boxed::Box};
use core::cell::RefCell;
use paging::MockPageTable;
enum MemOp {
R(usize), W(usize)
}
#[test]
fn test() {
use self::MemOp::{R, W};
let page_fault_count = Arc::new(RefCell::new(0usize));
let mut pt = MockPageTable::new(4);
pt.set_handler(Box::new({
let page_fault_count1 = page_fault_count.clone();
let mut fifo = FifoSwapManager::new();
move |pt: &mut MockPageTable, addr: VirtAddr| {
*page_fault_count1.borrow_mut() += 1;
if !pt.map(addr) { // is full?
pt.unmap(fifo.pop().unwrap());
pt.map(addr);
}
fifo.push(addr);
}
}));
let op_seq = [
R(0x1000), R(0x2000), R(0x3000), R(0x4000),
W(0x3000), W(0x1000), W(0x4000), W(0x2000), W(0x5000),
W(0x2000), W(0x1000), W(0x2000), W(0x3000), W(0x4000),
W(0x5000), R(0x1000), W(0x1000)];
let pgfault_count = [
1, 2, 3, 4,
4, 4, 4, 4, 5,
5, 6, 7, 8, 9,
10, 11, 11];
for (op, &count) in op_seq.iter().zip(pgfault_count.iter()) {
match op {
R(addr) => {pt.read(*addr);},
W(addr) => pt.write(*addr, 0),
}
assert_eq!(*(*page_fault_count).borrow(), count);
}
}
}

View File

@ -0,0 +1,79 @@
use super::Swapper;
use alloc::btree_map::BTreeMap;
pub struct MockSwapper {
map: BTreeMap<usize, [u8; 4096]>,
}
impl Swapper for MockSwapper {
fn swap_out(&mut self, data: &[u8; 4096]) -> Result<usize, ()> {
let id = self.alloc_id();
self.map.insert(id, data.clone());
Ok(id)
}
fn swap_update(&mut self, token: usize, data: &[u8; 4096]) -> Result<(), ()> {
if !self.map.contains_key(&token) {
return Err(());
}
self.map.insert(token, data.clone());
Ok(())
}
fn swap_in(&mut self, token: usize, data: &mut [u8; 4096]) -> Result<(), ()> {
match self.map.remove(&token) {
Some(d) => *data = d,
None => return Err(()),
}
Ok(())
}
}
impl MockSwapper {
pub fn new() -> Self {
MockSwapper {map: BTreeMap::new()}
}
fn alloc_id(&self) -> usize {
(0 .. 100usize).find(|i| !self.map.contains_key(i)).unwrap()
}
}
#[cfg(test)]
mod test {
use super::*;
use core::mem::uninitialized;
fn assert_data_eq(data1: &[u8; 4096], data2: &[u8; 4096]) {
for (&a, &b) in data2.iter().zip(data1.iter()) {
assert_eq!(a, b);
}
}
#[test]
fn swap_out_in() {
let mut swapper = MockSwapper::new();
let mut data: [u8; 4096] = unsafe{ uninitialized() };
let data1: [u8; 4096] = unsafe{ uninitialized() };
let token = swapper.swap_out(&data1).unwrap();
swapper.swap_in(token, &mut data).unwrap();
assert_data_eq(&data, &data1);
}
#[test]
fn swap_update() {
let mut swapper = MockSwapper::new();
let mut data: [u8; 4096] = unsafe{ uninitialized() };
let data1: [u8; 4096] = unsafe{ uninitialized() };
let data2: [u8; 4096] = unsafe{ uninitialized() };
let token = swapper.swap_out(&data1).unwrap();
swapper.swap_update(token, &data2).unwrap();
swapper.swap_in(token, &mut data).unwrap();
assert_data_eq(&data, &data2);
}
#[test]
fn invalid_token() {
let mut swapper = MockSwapper::new();
let mut data: [u8; 4096] = unsafe{ uninitialized() };
assert_eq!(swapper.swap_in(0, &mut data), Err(()));
}
}

View File

@ -0,0 +1,30 @@
pub use self::fifo::FifoSwapManager;
pub use self::enhanced_clock::EnhancedClockSwapManager;
use super::*;
use super::paging::PageTable;
mod fifo;
mod enhanced_clock;
mod mock_swapper;
pub trait SwapManager {
/// Called when tick interrupt occured
fn tick(&mut self);
/// Called when map a swappable page into the memory
fn push(&mut self, addr: VirtAddr);
/// Called to delete the addr entry from the swap manager
fn remove(&mut self, addr: VirtAddr);
/// Try to swap out a page, return then victim
fn pop(&mut self) -> Option<VirtAddr>;
}
pub trait Swapper {
fn swap_out(&mut self, data: &[u8; 4096]) -> Result<usize, ()>;
fn swap_update(&mut self, token: usize, data: &[u8; 4096]) -> Result<(), ()>;
fn swap_in(&mut self, token: usize, data: &mut [u8; 4096]) -> Result<(), ()>;
}
pub trait SwappablePageTable: PageTable {
fn swap_out(&mut self, addr: VirtAddr) -> Result<(), ()>;
}

View File

@ -4,23 +4,23 @@ mod structs;
use self::structs::*;
use consts::*;
pub fn init() -> Result<ACPI_Result, ACPI_Error> {
pub fn init() -> Result<AcpiResult, AcpiError> {
let rsdp = find_rsdp().expect("acpi: rsdp not found.");
if rsdp.RsdtPhysicalAddress > PHYSICAL_MEMORY_LIMIT {
return Err(ACPI_Error::NotMapped);
if rsdp.rsdt_physical_address > PHYSICAL_MEMORY_LIMIT {
return Err(AcpiError::NotMapped);
}
debug!("RSDT at {:#x}", rsdp.RsdtPhysicalAddress);
let rsdt = unsafe{ &*(rsdp.RsdtPhysicalAddress as *const rsdt) };
let mut madt: Option<&'static madt> = None;
debug!("RSDT at {:#x}", rsdp.rsdt_physical_address);
let rsdt = unsafe{ &*(rsdp.rsdt_physical_address as *const Rsdt) };
let mut madt: Option<&'static Madt> = None;
for i in 0 .. rsdt.entry_count() {
let entry = rsdt.entry_at(i);
if entry > PHYSICAL_MEMORY_LIMIT {
return Err(ACPI_Error::NotMapped);
return Err(AcpiError::NotMapped);
}
let header = unsafe{ &*(entry as *const header) };
let header = unsafe{ &*(entry as *const Header) };
// debug!("{:?}", header);
if &header.Signature == b"APIC" {
madt = Some(unsafe{ &*(entry as *const madt) });
if &header.signature == b"APIC" {
madt = Some(unsafe{ &*(entry as *const Madt) });
}
}
debug!("{:?}", madt);
@ -33,7 +33,7 @@ const PHYSICAL_MEMORY_LIMIT: u32 = 0x0E000000;
const PHYSICAL_MEMORY_LIMIT: u32 = 0x80000000;
#[derive(Debug)]
pub struct ACPI_Result {
pub struct AcpiResult {
pub cpu_num: u8,
pub cpu_acpi_ids: [u8; MAX_CPU_NUM],
pub ioapic_id: u8,
@ -41,13 +41,13 @@ pub struct ACPI_Result {
}
#[derive(Debug)]
pub enum ACPI_Error {
pub enum AcpiError {
NotMapped,
IOACPI_NotFound,
IoacpiNotFound,
}
fn config_smp(madt: &'static madt) -> Result<ACPI_Result, ACPI_Error> {
let lapic_addr = madt.LapicAddress as *const ();
fn config_smp(madt: &'static Madt) -> Result<AcpiResult, AcpiError> {
let lapic_addr = madt.lapic_address as *const ();
let mut cpu_num = 0u8;
let mut cpu_acpi_ids: [u8; MAX_CPU_NUM] = [0; MAX_CPU_NUM];
@ -56,33 +56,33 @@ fn config_smp(madt: &'static madt) -> Result<ACPI_Result, ACPI_Error> {
debug!("{:?}", entry);
match &entry {
&MadtEntry::LocalApic(ref lapic) => {
cpu_acpi_ids[cpu_num as usize] = lapic.Id;
cpu_acpi_ids[cpu_num as usize] = lapic.id;
cpu_num += 1;
},
&MadtEntry::IoApic(ref ioapic) => {
ioapic_id = Some(ioapic.Id);
ioapic_id = Some(ioapic.id);
},
_ => {},
}
}
if ioapic_id.is_none() {
return Err(ACPI_Error::IOACPI_NotFound);
return Err(AcpiError::IoacpiNotFound);
}
let ioapic_id = ioapic_id.unwrap();
Ok(ACPI_Result { cpu_num, cpu_acpi_ids, ioapic_id, lapic_addr })
Ok(AcpiResult { cpu_num, cpu_acpi_ids, ioapic_id, lapic_addr })
}
/// See https://wiki.osdev.org/RSDP -- Detecting the RSDP
fn find_rsdp() -> Option<&'static rsdp> {
fn find_rsdp() -> Option<&'static Rsdp> {
use util::{Checkable, find_in_memory};
let ebda = unsafe { *(0x40E as *const u16) as usize } << 4;
debug!("EBDA at {:#x}", ebda);
macro_rules! return_if_find_in {
($begin:expr, $end:expr) => (
if let Some(addr) = unsafe{ find_in_memory::<rsdp>($begin, $end, 4) } {
return Some(unsafe{ &*(addr as *const rsdp) });
if let Some(addr) = unsafe{ find_in_memory::<Rsdp>($begin, $end, 4) } {
return Some(unsafe{ &*(addr as *const Rsdp) });
}
)
}

View File

@ -6,48 +6,48 @@ use core::mem::size_of;
#[repr(C)]
#[derive(Debug)]
pub struct rsdp {
pub Signature: [u8; 8],
pub Checksum: u8,
pub OemId: [i8; 6],
pub Revision: u8,
pub RsdtPhysicalAddress: u32,
pub Length: u32,
pub XsdtPhysicalAddress: u64,
pub ExtendedChecksum: u8,
pub Reserved: [u8; 3],
pub struct Rsdp {
pub signature: [u8; 8],
pub checksum: u8,
pub oem_id: [i8; 6],
pub revision: u8,
pub rsdt_physical_address: u32,
pub length: u32,
pub xsdt_physical_address: u64,
pub extended_checksum: u8,
pub reserved: [u8; 3],
}
impl Checkable for rsdp {
impl Checkable for Rsdp {
fn check(&self) -> bool {
&self.Signature == b"RSD PTR " && bytes_sum(self) == 0
&self.signature == b"RSD PTR " && bytes_sum(self) == 0
}
}
#[repr(C)]
#[derive(Debug)]
pub struct header {
pub Signature: [u8; 4],
pub Length: u32,
pub Revision: u8,
pub Checksum: u8,
pub OemId: [i8; 6],
pub OemTableId: [i8; 8],
pub OemRevision: u32,
pub AslCompilerId: [i8; 4],
pub AslCompilerRevision: u32,
pub struct Header {
pub signature: [u8; 4],
pub length: u32,
pub revision: u8,
pub checksum: u8,
pub oem_id: [i8; 6],
pub oem_table_id: [i8; 8],
pub oem_revision: u32,
pub asl_compiler_id: [i8; 4],
pub asl_compiler_revision: u32,
}
#[repr(C)]
#[derive(Debug)]
pub struct rsdt {
pub Header: header,
TableOffsetEntry: [u32; 0],
pub struct Rsdt {
pub header: Header,
table_offset_entry: [u32; 0],
}
impl rsdt {
impl Rsdt {
pub fn entry_count(&self) -> usize {
(self.Header.Length as usize - size_of::<Self>()) / 4
(self.header.length as usize - size_of::<Self>()) / 4
}
pub fn entry_at(&self, id: usize) -> u32 {
assert!(id < self.entry_count());
@ -60,62 +60,62 @@ impl rsdt {
#[repr(C)]
#[derive(Debug)]
pub struct madt {
pub Header: header,
pub LapicAddress: u32,
pub Flags: u32,
Table: [u32; 0],
pub struct Madt {
pub header: Header,
pub lapic_address: u32,
pub flags: u32,
table: [u32; 0],
}
impl Checkable for madt {
impl Checkable for Madt {
fn check(&self) -> bool {
&self.Header.Signature == b"APIC" && self.Header.Length >= size_of::<Self>() as u32
&self.header.signature == b"APIC" && self.header.length >= size_of::<Self>() as u32
}
}
#[derive(Debug)]
pub enum MadtEntry {
Unknown(MadtEntry_Unknown),
LocalApic(MadtEntry_LocalApic),
IoApic(MadtEntry_IoApic),
Unknown(MadtEntryUnknown),
LocalApic(MadtEntryLocalApic),
IoApic(MadtEntryIoApic),
}
#[repr(C)]
#[derive(Debug, Clone)]
pub struct MadtEntry_Unknown {
pub Type: u8,
pub Length: u8,
pub struct MadtEntryUnknown {
pub type_: u8,
pub length: u8,
}
#[repr(C)]
#[derive(Debug, Clone)]
pub struct MadtEntry_LocalApic {
pub Type: u8, // 0
pub Length: u8,
pub ProcessorId: u8,
pub Id: u8,
pub LapicFlags: u32,
pub struct MadtEntryLocalApic {
pub type_: u8, // 0
pub length: u8,
pub processor_id: u8,
pub id: u8,
pub lapic_flags: u32,
}
#[repr(C)]
#[derive(Debug, Clone)]
pub struct MadtEntry_IoApic {
pub Type: u8, // 1
pub Length: u8,
pub Id: u8,
pub Reserved: u8,
pub Address: u32,
pub GlobalIrqBase: u32,
pub struct MadtEntryIoApic {
pub type_: u8, // 1
pub length: u8,
pub id: u8,
pub reserved: u8,
pub address: u32,
pub global_irq_base: u32,
}
#[derive(Debug)]
pub struct MadtEntryIter<'a> {
madt: &'a madt,
madt: &'a Madt,
ptr: *const u8,
end_ptr: *const u8,
}
impl madt {
impl Madt {
pub fn entry_iter(&self) -> MadtEntryIter {
let ptr = unsafe{ (self as *const Self).offset(1) } as *const u8;
let end_ptr = unsafe{ ptr.offset(self.Header.Length as isize) };
let end_ptr = unsafe{ ptr.offset(self.header.length as isize) };
MadtEntryIter { madt: self, ptr, end_ptr }
}
}
@ -127,12 +127,12 @@ impl<'a> Iterator for MadtEntryIter<'a> {
return None;
}
unsafe {
let typeId = *self.ptr.offset(0);
let type_id = *self.ptr.offset(0);
let len = *self.ptr.offset(1);
let ret = Some(match typeId {
0 => MadtEntry::LocalApic( (&*(self.ptr as *const MadtEntry_LocalApic)).clone() ),
1 => MadtEntry::IoApic( (&*(self.ptr as *const MadtEntry_IoApic)).clone() ),
_ => MadtEntry::Unknown( (&*(self.ptr as *const MadtEntry_Unknown)).clone() ),
let ret = Some(match type_id {
0 => MadtEntry::LocalApic( (&*(self.ptr as *const MadtEntryLocalApic)).clone() ),
1 => MadtEntry::IoApic( (&*(self.ptr as *const MadtEntryIoApic)).clone() ),
_ => MadtEntry::Unknown( (&*(self.ptr as *const MadtEntryUnknown)).clone() ),
});
self.ptr = self.ptr.offset(len as isize);
ret

View File

@ -1,4 +1,5 @@
extern {
//noinspection RsStaticConstNaming
static mut lapic: *const ();
fn lapicinit(); // must set `lapic` first
fn lapiceoi(); // ack

View File

@ -3,7 +3,7 @@ use x86::cpuid::CpuId;
use x86::msr::*;
use memory::Frame;
use paging::{ActivePageTable, PhysicalAddress, Page, VirtualAddress};
use paging::{ActivePageTable, PhysAddr, Page, VirtualAddress};
use paging::entry::EntryFlags;
pub static mut LOCAL_APIC: LocalApic = LocalApic {
@ -32,7 +32,7 @@ impl LocalApic {
if ! self.x2 {
let page = Page::containing_address(VirtualAddress::new(self.address));
let frame = Frame::containing_address(PhysicalAddress::new(self.address - ::KERNEL_OFFSET));
let frame = Frame::containing_address(PhysAddr::new(self.address - ::KERNEL_OFFSET));
let result = active_table.map_to(page, frame, EntryFlags::PRESENT | EntryFlags::WRITABLE | EntryFlags::NO_EXECUTE);
result.flush(active_table);
}

View File

@ -7,7 +7,7 @@ pub mod pic;
pub mod keyboard;
pub mod pit;
pub fn init<F>(mut page_map: F) -> acpi::ACPI_Result
pub fn init<F>(mut page_map: F) -> acpi::AcpiResult
where F: FnMut(usize) {
assert_has_not_been_called!();

View File

@ -18,7 +18,7 @@ impl Entry {
pub fn pointed_frame(&self) -> Option<Frame> {
if self.flags().contains(EntryFlags::PRESENT) {
Some(Frame::containing_address(
Some(Frame::of_addr(
self.0 as usize & 0x000fffff_fffff000
))
} else {
@ -27,7 +27,7 @@ impl Entry {
}
pub fn set(&mut self, frame: Frame, flags: EntryFlags) {
assert!(frame.start_address().0 & !0x000fffff_fffff000 == 0);
assert_eq!(frame.start_address().0 & !0x000fffff_fffff000, 0);
self.0 = (frame.start_address().0) | flags.bits();
}
}

View File

@ -22,10 +22,10 @@ impl Mapper {
unsafe { self.p4.as_mut() }
}
pub fn translate(&self, virtual_address: VirtualAddress) -> Option<PhysicalAddress> {
pub fn translate(&self, virtual_address: VirtAddr) -> Option<PhysAddr> {
let offset = virtual_address % PAGE_SIZE;
self.translate_page(Page::containing_address(virtual_address))
.map(|frame| PhysicalAddress((frame.start_address().get() + offset) as u64))
self.translate_page(Page::of_addr(virtual_address))
.map(|frame| PhysAddr((frame.start_address().get() + offset) as u64))
}
pub fn translate_page(&self, page: Page) -> Option<Frame> {
@ -38,8 +38,8 @@ impl Mapper {
if let Some(start_frame) = p3_entry.pointed_frame() {
if p3_entry.flags().contains(EntryFlags::HUGE_PAGE) {
// address must be 1GiB aligned
assert!(start_frame.start_address().get() % (ENTRY_COUNT * ENTRY_COUNT * PAGE_SIZE) == 0);
return Some(Frame::containing_address(
assert_eq!(start_frame.start_address().get() % (ENTRY_COUNT * ENTRY_COUNT * PAGE_SIZE), 0);
return Some(Frame::of_addr(
start_frame.start_address().get() +
(page.p2_index() * ENTRY_COUNT + page.p1_index()) * PAGE_SIZE
));
@ -51,8 +51,8 @@ impl Mapper {
if let Some(start_frame) = p2_entry.pointed_frame() {
if p2_entry.flags().contains(EntryFlags::HUGE_PAGE) {
// address must be 2MiB aligned
assert!(start_frame.start_address().get() % ENTRY_COUNT == 0);
return Some(Frame::containing_address(
assert_eq!(start_frame.start_address().get() % ENTRY_COUNT, 0);
return Some(Frame::of_addr(
start_frame.start_address().get() + page.p1_index() * PAGE_SIZE
));
}
@ -92,7 +92,7 @@ impl Mapper {
pub fn identity_map<A>(&mut self, frame: Frame, flags: EntryFlags, allocator: &mut A)
where A: FrameAllocator
{
let page = Page::containing_address(frame.start_address().to_identity_virtual());
let page = Page::of_addr(frame.start_address().to_identity_virtual());
self.map_to(page, frame, flags, allocator)
}

View File

@ -17,7 +17,7 @@ pub struct Page {
}
impl Page {
pub fn containing_address(address: VirtualAddress) -> Page {
pub fn of_addr(address: VirtAddr) -> Page {
assert!(address < 0x0000_8000_0000_0000 ||
address >= 0xffff_8000_0000_0000,
"invalid address: 0x{:x}", address);
@ -43,8 +43,8 @@ impl Page {
pub fn range_inclusive(start: Page, end: Page) -> PageIter {
PageIter {
start: start,
end: end,
start,
end,
}
}
}
@ -113,7 +113,7 @@ impl ActivePageTable {
use x86_64::registers::control_regs;
{
let backup = Frame::containing_address(
let backup = Frame::of_addr(
control_regs::cr3().0 as usize);
// map temporary_page to current p4 table
@ -139,7 +139,7 @@ impl ActivePageTable {
use x86_64::registers::control_regs;
let old_table = InactivePageTable {
p4_frame: Frame::containing_address(
p4_frame: Frame::of_addr(
control_regs::cr3().0 as usize
),
};

View File

@ -1,6 +1,6 @@
use super::{Page, ActivePageTable};
use super::table::{Table, Level1};
use memory::{Frame, FrameAllocator, VirtualAddress};
use memory::{Frame, FrameAllocator, VirtAddr};
pub struct TemporaryPage {
page: Page,
@ -12,7 +12,7 @@ impl TemporaryPage {
where A: FrameAllocator
{
TemporaryPage {
page: page,
page,
allocator: TinyAllocator::new(allocator),
}
}
@ -20,7 +20,7 @@ impl TemporaryPage {
/// Maps the temporary page to the given frame in the active table.
/// Returns the start address of the temporary page.
pub fn map(&mut self, frame: Frame, active_table: &mut ActivePageTable)
-> VirtualAddress
-> VirtAddr
{
use super::entry::EntryFlags;

View File

@ -1,5 +1,5 @@
use arch::driver::{acpi::ACPI_Result, apic::start_ap};
use memory::{MemoryController, PhysicalAddress};
use arch::driver::{acpi::AcpiResult, apic::start_ap};
use memory::{MemoryController, PhysAddr};
extern {
fn entryother_start(); // physical addr of entryother
@ -8,11 +8,11 @@ extern {
const ENTRYOTHER_ADDR: u32 = 0x7000;
pub fn start_other_cores(acpi: &ACPI_Result, mc: &mut MemoryController) {
pub fn start_other_cores(acpi: &AcpiResult, mc: &mut MemoryController) {
mc.map_page_identity(ENTRYOTHER_ADDR as usize - 1);
mc.map_page_identity(ENTRYOTHER_ADDR as usize);
mc.map_page_identity(entryother_start as usize);
mc.map_page_p2v(PhysicalAddress(0));
mc.map_page_p2v(PhysAddr(0));
copy_entryother();
let args = unsafe{ &mut *(ENTRYOTHER_ADDR as *mut EntryArgs).offset(-1) };

View File

@ -1,3 +1,4 @@
#![allow(dead_code)]
pub const MAX_CPU_NUM: usize = 8;
// Copy from Redox consts.rs:

View File

@ -21,7 +21,7 @@ impl VgaWriter {
VgaWriter {
column_position: 0,
color: Color::LightGray,
buffer: buffer,
buffer,
}
}

View File

@ -1,27 +1,27 @@
use consts::{KERNEL_OFFSET, KERNEL_SIZE};
pub use x86_64::{PhysicalAddress};
pub type VirtualAddress = usize;
pub use x86_64::PhysicalAddress as PhysAddr;
pub type VirtAddr = usize;
pub trait FromToVirtualAddress {
fn get(&self) -> usize;
fn to_identity_virtual(&self) -> VirtualAddress;
fn to_kernel_virtual(&self) -> VirtualAddress;
fn from_kernel_virtual(addr: VirtualAddress) -> Self;
fn to_identity_virtual(&self) -> VirtAddr;
fn to_kernel_virtual(&self) -> VirtAddr;
fn from_kernel_virtual(addr: VirtAddr) -> Self;
}
impl FromToVirtualAddress for PhysicalAddress {
impl FromToVirtualAddress for PhysAddr {
fn get(&self) -> usize {
self.0 as usize
}
fn to_identity_virtual(&self) -> VirtualAddress {
fn to_identity_virtual(&self) -> VirtAddr {
self.0 as usize
}
fn to_kernel_virtual(&self) -> VirtualAddress {
fn to_kernel_virtual(&self) -> VirtAddr {
assert!((self.0 as usize) < KERNEL_SIZE);
self.0 as usize + KERNEL_OFFSET
}
fn from_kernel_virtual(addr: VirtualAddress) -> Self {
fn from_kernel_virtual(addr: VirtAddr) -> Self {
assert!(addr >= KERNEL_OFFSET && addr < KERNEL_OFFSET + KERNEL_SIZE);
PhysicalAddress((addr - KERNEL_OFFSET) as u64)
PhysAddr((addr - KERNEL_OFFSET) as u64)
}
}

View File

@ -1,4 +1,4 @@
use memory::{Frame, FrameAllocator, PhysicalAddress};
use memory::{Frame, FrameAllocator, PhysAddr};
use multiboot2::{MemoryAreaIter, MemoryArea};
pub struct AreaFrameAllocator {
@ -21,7 +21,7 @@ impl FrameAllocator for AreaFrameAllocator {
// the last frame of the current area
let current_area_last_frame = {
let address = area.base_addr + area.length - 1;
Frame::containing_address(address as usize)
Frame::of_addr(address as usize)
};
if frame > current_area_last_frame {
@ -55,18 +55,18 @@ impl FrameAllocator for AreaFrameAllocator {
}
impl AreaFrameAllocator {
pub fn new(kernel_start: PhysicalAddress, kernel_end: PhysicalAddress,
multiboot_start: PhysicalAddress, multiboot_end: PhysicalAddress,
pub fn new(kernel_start: PhysAddr, kernel_end: PhysAddr,
multiboot_start: PhysAddr, multiboot_end: PhysAddr,
memory_areas: MemoryAreaIter) -> AreaFrameAllocator
{
let mut allocator = AreaFrameAllocator {
next_free_frame: Frame::containing_address(0),
next_free_frame: Frame::of_addr(0),
current_area: None,
areas: memory_areas,
kernel_start: Frame::containing_address(kernel_start.0 as usize),
kernel_end: Frame::containing_address(kernel_end.0 as usize),
multiboot_start: Frame::containing_address(multiboot_start.0 as usize),
multiboot_end: Frame::containing_address(multiboot_end.0 as usize),
kernel_start: Frame::of_addr(kernel_start.0 as usize),
kernel_end: Frame::of_addr(kernel_end.0 as usize),
multiboot_start: Frame::of_addr(multiboot_start.0 as usize),
multiboot_end: Frame::of_addr(multiboot_end.0 as usize),
};
allocator.choose_next_area();
allocator
@ -75,11 +75,11 @@ impl AreaFrameAllocator {
fn choose_next_area(&mut self) {
self.current_area = self.areas.clone().filter(|area| {
let address = area.base_addr + area.length - 1;
Frame::containing_address(address as usize) >= self.next_free_frame
Frame::of_addr(address as usize) >= self.next_free_frame
}).min_by_key(|area| area.base_addr);
if let Some(area) = self.current_area {
let start_frame = Frame::containing_address(area.base_addr as usize);
let start_frame = Frame::of_addr(area.base_addr as usize);
if self.next_free_frame < start_frame {
self.next_free_frame = start_frame;
}

View File

@ -1,4 +1,4 @@
use super::address::PhysicalAddress;
use super::address::PhysAddr;
pub const PAGE_SIZE: usize = 4096;
@ -8,12 +8,12 @@ pub struct Frame {
}
impl Frame {
pub fn containing_address(address: usize) -> Frame {
pub fn of_addr(address: usize) -> Frame {
Frame{ number: address / PAGE_SIZE }
}
//TODO: Set private
pub fn start_address(&self) -> PhysicalAddress {
PhysicalAddress((self.number * PAGE_SIZE) as u64)
pub fn start_address(&self) -> PhysAddr {
PhysAddr((self.number * PAGE_SIZE) as u64)
}
pub fn clone(&self) -> Frame {
@ -22,8 +22,8 @@ impl Frame {
//TODO: Set private
pub fn range_inclusive(start: Frame, end: Frame) -> FrameIter {
FrameIter {
start: start,
end: end,
start,
end,
}
}
}

View File

@ -22,13 +22,13 @@ pub fn init(boot_info: &BootInformation) -> MemoryController {
let elf_sections_tag = boot_info.elf_sections_tag().expect(
"Elf sections tag required");
let kernel_start = PhysicalAddress(elf_sections_tag.sections()
let kernel_start = PhysAddr(elf_sections_tag.sections()
.filter(|s| s.is_allocated()).map(|s| s.start_address()).min().unwrap() as u64);
let kernel_end = PhysicalAddress::from_kernel_virtual(elf_sections_tag.sections()
let kernel_end = PhysAddr::from_kernel_virtual(elf_sections_tag.sections()
.filter(|s| s.is_allocated()).map(|s| s.end_address()).max().unwrap());
let boot_info_start = PhysicalAddress(boot_info.start_address() as u64);
let boot_info_end = PhysicalAddress(boot_info.end_address() as u64);
let boot_info_start = PhysAddr(boot_info.start_address() as u64);
let boot_info_end = PhysAddr(boot_info.end_address() as u64);
println!("kernel start: {:#x}, kernel end: {:#x}",
kernel_start,
@ -51,8 +51,8 @@ pub fn init(boot_info: &BootInformation) -> MemoryController {
use self::paging::Page;
use consts::{KERNEL_HEAP_OFFSET, KERNEL_HEAP_SIZE};
let heap_start_page = Page::containing_address(KERNEL_HEAP_OFFSET);
let heap_end_page = Page::containing_address(KERNEL_HEAP_OFFSET + KERNEL_HEAP_SIZE-1);
let heap_start_page = Page::of_addr(KERNEL_HEAP_OFFSET);
let heap_end_page = Page::of_addr(KERNEL_HEAP_OFFSET + KERNEL_HEAP_SIZE-1);
for page in Page::range_inclusive(heap_start_page, heap_end_page) {
active_table.map(page, EntryFlags::WRITABLE, &mut frame_allocator);
@ -78,7 +78,7 @@ pub fn remap_the_kernel<A>(allocator: &mut A, boot_info: &BootInformation)
-> (ActivePageTable, Stack)
where A: FrameAllocator
{
let mut temporary_page = TemporaryPage::new(Page::containing_address(0xcafebabe), allocator);
let mut temporary_page = TemporaryPage::new(Page::of_addr(0xcafebabe), allocator);
let mut active_table = unsafe { ActivePageTable::new() };
let mut new_table = {
@ -95,8 +95,7 @@ pub fn remap_the_kernel<A>(allocator: &mut A, boot_info: &BootInformation)
// section is not loaded to memory
continue;
}
assert!(section.start_address() % PAGE_SIZE == 0,
"sections need to be page aligned");
assert_eq!(section.start_address() % PAGE_SIZE, 0, "sections need to be page aligned");
println!("mapping section at addr: {:#x}, size: {:#x}",
section.addr, section.size);
@ -104,7 +103,7 @@ pub fn remap_the_kernel<A>(allocator: &mut A, boot_info: &BootInformation)
let flags = EntryFlags::from_elf_section_flags(section);
fn to_physical_frame(addr: usize) -> Frame {
Frame::containing_address(
Frame::of_addr(
if addr < KERNEL_OFFSET { addr }
else { addr - KERNEL_OFFSET })
}
@ -113,18 +112,18 @@ pub fn remap_the_kernel<A>(allocator: &mut A, boot_info: &BootInformation)
let end_frame = to_physical_frame(section.end_address() - 1);
for frame in Frame::range_inclusive(start_frame, end_frame) {
let page = Page::containing_address(frame.start_address().to_kernel_virtual());
let page = Page::of_addr(frame.start_address().to_kernel_virtual());
mapper.map_to(page, frame, flags, allocator);
}
}
// identity map the VGA text buffer
let vga_buffer_frame = Frame::containing_address(0xb8000);
let vga_buffer_frame = Frame::of_addr(0xb8000);
mapper.identity_map(vga_buffer_frame, EntryFlags::WRITABLE, allocator);
// identity map the multiboot info structure
let multiboot_start = Frame::containing_address(boot_info.start_address());
let multiboot_end = Frame::containing_address(boot_info.end_address() - 1);
let multiboot_start = Frame::of_addr(boot_info.start_address());
let multiboot_end = Frame::of_addr(boot_info.end_address() - 1);
for frame in Frame::range_inclusive(multiboot_start, multiboot_end) {
mapper.identity_map(frame, EntryFlags::PRESENT, allocator);
}
@ -135,8 +134,8 @@ pub fn remap_the_kernel<A>(allocator: &mut A, boot_info: &BootInformation)
// turn the stack bottom into a guard page
extern { fn stack_bottom(); }
let stack_bottom = PhysicalAddress(stack_bottom as u64).to_kernel_virtual();
let stack_bottom_page = Page::containing_address(stack_bottom);
let stack_bottom = PhysAddr(stack_bottom as u64).to_kernel_virtual();
let stack_bottom_page = Page::of_addr(stack_bottom);
active_table.unmap(stack_bottom_page, allocator);
let kernel_stack = Stack::new(stack_bottom + 8 * PAGE_SIZE, stack_bottom + 1 * PAGE_SIZE);
println!("guard page at {:#x}", stack_bottom_page.start_address());
@ -161,13 +160,13 @@ impl MemoryController {
size_in_pages)
}
pub fn map_page_identity(&mut self, addr: usize) {
let frame = Frame::containing_address(addr);
let frame = Frame::of_addr(addr);
let flags = EntryFlags::WRITABLE;
self.active_table.identity_map(frame, flags, &mut self.frame_allocator);
}
pub fn map_page_p2v(&mut self, addr: PhysicalAddress) {
let page = Page::containing_address(addr.to_kernel_virtual());
let frame = Frame::containing_address(addr.get());
pub fn map_page_p2v(&mut self, addr: PhysAddr) {
let page = Page::of_addr(addr.to_kernel_virtual());
let frame = Frame::of_addr(addr.get());
let flags = EntryFlags::WRITABLE;
self.active_table.map_to(page, frame, flags, &mut self.frame_allocator);
}

View File

@ -64,8 +64,8 @@ impl Stack {
pub(super) fn new(top: usize, bottom: usize) -> Stack {
assert!(top > bottom);
Stack {
top: top,
bottom: bottom,
top,
bottom,
}
}