2020-03-16 04:18:46 +04:00
|
|
|
// block.rs
|
|
|
|
// Block device using VirtIO protocol
|
|
|
|
// Stephen Marz
|
|
|
|
// 10 March 2020
|
|
|
|
|
2020-04-25 02:37:48 +04:00
|
|
|
use crate::{kmem::{kfree, kmalloc, talloc, tfree},
|
2020-03-16 04:18:46 +04:00
|
|
|
page::{zalloc, PAGE_SIZE},
|
2020-04-26 03:51:27 +04:00
|
|
|
process::{add_kernel_process_args,
|
|
|
|
get_by_pid,
|
|
|
|
set_running,
|
|
|
|
set_waiting},
|
2020-03-16 04:18:46 +04:00
|
|
|
virtio,
|
2020-04-26 03:51:27 +04:00
|
|
|
virtio::{Descriptor,
|
|
|
|
MmioOffsets,
|
|
|
|
Queue,
|
|
|
|
StatusField,
|
|
|
|
VIRTIO_RING_SIZE}};
|
2020-03-16 04:18:46 +04:00
|
|
|
use core::mem::size_of;
|
|
|
|
|
|
|
|
#[repr(C)]
|
|
|
|
pub struct Geometry {
|
|
|
|
cylinders: u16,
|
|
|
|
heads: u8,
|
|
|
|
sectors: u8,
|
|
|
|
}
|
|
|
|
|
|
|
|
#[repr(C)]
|
|
|
|
pub struct Topology {
|
|
|
|
physical_block_exp: u8,
|
|
|
|
alignment_offset: u8,
|
|
|
|
min_io_size: u16,
|
|
|
|
opt_io_size: u32,
|
|
|
|
}
|
|
|
|
|
|
|
|
// There is a configuration space for VirtIO that begins
|
|
|
|
// at offset 0x100 and continues to the size of the configuration.
|
|
|
|
// The structure below represents the configuration for a
|
|
|
|
// block device. Really, all that this OS cares about is the
|
|
|
|
// capacity.
|
|
|
|
#[repr(C)]
|
|
|
|
pub struct Config {
|
|
|
|
capacity: u64,
|
|
|
|
size_max: u32,
|
|
|
|
seg_max: u32,
|
|
|
|
geometry: Geometry,
|
|
|
|
blk_size: u32,
|
|
|
|
topology: Topology,
|
|
|
|
writeback: u8,
|
|
|
|
unused0: [u8; 3],
|
|
|
|
max_discard_sector: u32,
|
|
|
|
max_discard_seg: u32,
|
|
|
|
discard_sector_alignment: u32,
|
|
|
|
max_write_zeroes_sectors: u32,
|
|
|
|
max_write_zeroes_seg: u32,
|
|
|
|
write_zeroes_may_unmap: u8,
|
|
|
|
unused1: [u8; 3],
|
|
|
|
}
|
|
|
|
|
|
|
|
// The header/data/status is a block request
|
|
|
|
// packet. We send the header to tell the direction
|
|
|
|
// (blktype: IN/OUT) and then the starting sector
|
|
|
|
// we want to read. Then, we put the data buffer
|
|
|
|
// as the Data structure and finally an 8-bit
|
|
|
|
// status. The device will write one of three values
|
|
|
|
// in here: 0 = success, 1 = io error, 2 = unsupported
|
|
|
|
// operation.
|
|
|
|
#[repr(C)]
|
|
|
|
pub struct Header {
|
|
|
|
blktype: u32,
|
|
|
|
reserved: u32,
|
|
|
|
sector: u64,
|
|
|
|
}
|
|
|
|
|
|
|
|
#[repr(C)]
|
|
|
|
pub struct Data {
|
|
|
|
data: *mut u8,
|
|
|
|
}
|
|
|
|
|
|
|
|
#[repr(C)]
|
|
|
|
pub struct Status {
|
|
|
|
status: u8,
|
|
|
|
}
|
|
|
|
|
|
|
|
#[repr(C)]
|
|
|
|
pub struct Request {
|
|
|
|
header: Header,
|
|
|
|
data: Data,
|
|
|
|
status: Status,
|
|
|
|
head: u16,
|
2020-04-24 22:39:56 +04:00
|
|
|
|
|
|
|
// Do not change anything above this line.
|
|
|
|
// This is the PID of watcher. We store the PID
|
|
|
|
// because it is possible that the process DIES
|
|
|
|
// before we get here. If we used a pointer, we
|
|
|
|
// may dereference invalid memory.
|
|
|
|
watcher: u16,
|
2020-03-16 04:18:46 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
// Internal block device structure
|
|
|
|
// We keep our own used_idx and idx for
|
|
|
|
// descriptors. There is a shared index, but that
|
|
|
|
// tells us or the device if we've kept up with where
|
|
|
|
// we are for the available (us) or used (device) ring.
|
|
|
|
pub struct BlockDevice {
|
|
|
|
queue: *mut Queue,
|
|
|
|
dev: *mut u32,
|
|
|
|
idx: u16,
|
|
|
|
ack_used_idx: u16,
|
|
|
|
read_only: bool,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Type values
|
|
|
|
pub const VIRTIO_BLK_T_IN: u32 = 0;
|
|
|
|
pub const VIRTIO_BLK_T_OUT: u32 = 1;
|
|
|
|
pub const VIRTIO_BLK_T_FLUSH: u32 = 4;
|
|
|
|
pub const VIRTIO_BLK_T_DISCARD: u32 = 11;
|
|
|
|
pub const VIRTIO_BLK_T_WRITE_ZEROES: u32 = 13;
|
|
|
|
|
|
|
|
// Status values
|
|
|
|
pub const VIRTIO_BLK_S_OK: u8 = 0;
|
|
|
|
pub const VIRTIO_BLK_S_IOERR: u8 = 1;
|
|
|
|
pub const VIRTIO_BLK_S_UNSUPP: u8 = 2;
|
|
|
|
|
|
|
|
// Feature bits
|
|
|
|
pub const VIRTIO_BLK_F_SIZE_MAX: u32 = 1;
|
|
|
|
pub const VIRTIO_BLK_F_SEG_MAX: u32 = 2;
|
|
|
|
pub const VIRTIO_BLK_F_GEOMETRY: u32 = 4;
|
|
|
|
pub const VIRTIO_BLK_F_RO: u32 = 5;
|
|
|
|
pub const VIRTIO_BLK_F_BLK_SIZE: u32 = 6;
|
|
|
|
pub const VIRTIO_BLK_F_FLUSH: u32 = 9;
|
|
|
|
pub const VIRTIO_BLK_F_TOPOLOGY: u32 = 10;
|
|
|
|
pub const VIRTIO_BLK_F_CONFIG_WCE: u32 = 11;
|
|
|
|
pub const VIRTIO_BLK_F_DISCARD: u32 = 13;
|
|
|
|
pub const VIRTIO_BLK_F_WRITE_ZEROES: u32 = 14;
|
|
|
|
|
2020-04-24 22:39:56 +04:00
|
|
|
// We might get several types of errors, but they can be enumerated here.
|
|
|
|
pub enum BlockErrors {
|
|
|
|
Success = 0,
|
|
|
|
BlockDeviceNotFound,
|
|
|
|
InvalidArgument,
|
2020-04-25 02:46:31 +04:00
|
|
|
ReadOnly,
|
2020-04-24 22:39:56 +04:00
|
|
|
}
|
|
|
|
|
2020-03-16 04:18:46 +04:00
|
|
|
// Much like with processes, Rust requires some initialization
|
|
|
|
// when we declare a static. In this case, we use the Option
|
|
|
|
// value type to signal that the variable exists, but not the
|
|
|
|
// queue itself. We will replace this with an actual queue when
|
|
|
|
// we initialize the block system.
|
2020-04-26 03:51:27 +04:00
|
|
|
static mut BLOCK_DEVICES: [Option<BlockDevice>; 8] =
|
|
|
|
[None, None, None, None, None, None, None, None];
|
2020-03-16 04:18:46 +04:00
|
|
|
|
|
|
|
pub fn setup_block_device(ptr: *mut u32) -> bool {
|
|
|
|
unsafe {
|
|
|
|
// We can get the index of the device based on its address.
|
|
|
|
// 0x1000_1000 is index 0
|
|
|
|
// 0x1000_2000 is index 1
|
|
|
|
// ...
|
|
|
|
// 0x1000_8000 is index 7
|
2020-04-26 03:51:27 +04:00
|
|
|
// To get the number that changes over, we shift right 12 places
|
|
|
|
// (3 hex digits)
|
2020-03-16 04:18:46 +04:00
|
|
|
let idx = (ptr as usize - virtio::MMIO_VIRTIO_START) >> 12;
|
|
|
|
// [Driver] Device Initialization
|
|
|
|
// 1. Reset the device (write 0 into status)
|
|
|
|
ptr.add(MmioOffsets::Status.scale32()).write_volatile(0);
|
|
|
|
let mut status_bits = StatusField::Acknowledge.val32();
|
|
|
|
// 2. Set ACKNOWLEDGE status bit
|
2020-04-26 03:51:27 +04:00
|
|
|
ptr.add(MmioOffsets::Status.scale32())
|
|
|
|
.write_volatile(status_bits);
|
2020-03-16 04:18:46 +04:00
|
|
|
// 3. Set the DRIVER status bit
|
|
|
|
status_bits |= StatusField::DriverOk.val32();
|
2020-04-26 03:51:27 +04:00
|
|
|
ptr.add(MmioOffsets::Status.scale32())
|
|
|
|
.write_volatile(status_bits);
|
2020-03-16 04:18:46 +04:00
|
|
|
// 4. Read device feature bits, write subset of feature
|
|
|
|
// bits understood by OS and driver to the device.
|
2020-04-26 03:51:27 +04:00
|
|
|
let host_features =
|
|
|
|
ptr.add(MmioOffsets::HostFeatures.scale32())
|
|
|
|
.read_volatile();
|
2020-03-16 04:18:46 +04:00
|
|
|
let guest_features = host_features & !(1 << VIRTIO_BLK_F_RO);
|
|
|
|
let ro = host_features & (1 << VIRTIO_BLK_F_RO) != 0;
|
2020-04-26 03:51:27 +04:00
|
|
|
ptr.add(MmioOffsets::GuestFeatures.scale32())
|
|
|
|
.write_volatile(guest_features);
|
2020-03-16 04:18:46 +04:00
|
|
|
// 5. Set the FEATURES_OK status bit
|
|
|
|
status_bits |= StatusField::FeaturesOk.val32();
|
2020-04-26 03:51:27 +04:00
|
|
|
ptr.add(MmioOffsets::Status.scale32())
|
|
|
|
.write_volatile(status_bits);
|
2020-03-16 04:18:46 +04:00
|
|
|
// 6. Re-read status to ensure FEATURES_OK is still set.
|
|
|
|
// Otherwise, it doesn't support our features.
|
2020-04-26 03:51:27 +04:00
|
|
|
let status_ok =
|
|
|
|
ptr.add(MmioOffsets::Status.scale32()).read_volatile();
|
2020-03-16 04:18:46 +04:00
|
|
|
// If the status field no longer has features_ok set,
|
|
|
|
// that means that the device couldn't accept
|
|
|
|
// the features that we request. Therefore, this is
|
|
|
|
// considered a "failed" state.
|
|
|
|
if false == StatusField::features_ok(status_ok) {
|
|
|
|
print!("features fail...");
|
2020-04-26 03:51:27 +04:00
|
|
|
ptr.add(MmioOffsets::Status.scale32())
|
|
|
|
.write_volatile(StatusField::Failed.val32());
|
2020-03-16 04:18:46 +04:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
// 7. Perform device-specific setup.
|
|
|
|
// Set the queue num. We have to make sure that the
|
|
|
|
// queue size is valid because the device can only take
|
|
|
|
// a certain size.
|
2020-04-26 03:51:27 +04:00
|
|
|
let qnmax = ptr.add(MmioOffsets::QueueNumMax.scale32())
|
|
|
|
.read_volatile();
|
|
|
|
ptr.add(MmioOffsets::QueueNum.scale32())
|
|
|
|
.write_volatile(VIRTIO_RING_SIZE as u32);
|
2020-03-16 04:18:46 +04:00
|
|
|
if VIRTIO_RING_SIZE as u32 > qnmax {
|
|
|
|
print!("queue size fail...");
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
// First, if the block device array is empty, create it!
|
|
|
|
// We add 4095 to round this up and then do an integer
|
|
|
|
// divide to truncate the decimal. We don't add 4096,
|
|
|
|
// because if it is exactly 4096 bytes, we would get two
|
|
|
|
// pages, not one.
|
2020-04-26 03:51:27 +04:00
|
|
|
let num_pages =
|
|
|
|
(size_of::<Queue>() + PAGE_SIZE - 1) / PAGE_SIZE;
|
2020-03-16 04:18:46 +04:00
|
|
|
// println!("np = {}", num_pages);
|
|
|
|
// We allocate a page for each device. This will the the
|
|
|
|
// descriptor where we can communicate with the block
|
|
|
|
// device. We will still use an MMIO register (in
|
|
|
|
// particular, QueueNotify) to actually tell the device
|
|
|
|
// we put something in memory. We also have to be
|
|
|
|
// careful with memory ordering. We don't want to
|
|
|
|
// issue a notify before all memory writes have
|
|
|
|
// finished. We will look at that later, but we need
|
|
|
|
// what is called a memory "fence" or barrier.
|
|
|
|
ptr.add(MmioOffsets::QueueSel.scale32()).write_volatile(0);
|
|
|
|
// Alignment is very important here. This is the memory address
|
2020-04-26 03:51:27 +04:00
|
|
|
// alignment between the available and used rings. If this is
|
|
|
|
// wrong, then we and the device will refer to different memory
|
|
|
|
// addresses and hence get the wrong data in the used ring.
|
2020-03-16 04:18:46 +04:00
|
|
|
// ptr.add(MmioOffsets::QueueAlign.scale32()).write_volatile(2);
|
|
|
|
let queue_ptr = zalloc(num_pages) as *mut Queue;
|
|
|
|
let queue_pfn = queue_ptr as u32;
|
2020-04-26 03:51:27 +04:00
|
|
|
ptr.add(MmioOffsets::GuestPageSize.scale32())
|
|
|
|
.write_volatile(PAGE_SIZE as u32);
|
2020-03-16 04:18:46 +04:00
|
|
|
// QueuePFN is a physical page number, however it
|
|
|
|
// appears for QEMU we have to write the entire memory
|
|
|
|
// address. This is a physical memory address where we
|
|
|
|
// (the OS) and the block device have in common for
|
|
|
|
// making and receiving requests.
|
2020-04-26 03:51:27 +04:00
|
|
|
ptr.add(MmioOffsets::QueuePfn.scale32())
|
|
|
|
.write_volatile(queue_pfn / PAGE_SIZE as u32);
|
2020-03-16 04:18:46 +04:00
|
|
|
// We need to store all of this data as a "BlockDevice"
|
|
|
|
// structure We will be referring to this structure when
|
|
|
|
// making block requests AND when handling responses.
|
|
|
|
let bd = BlockDevice { queue: queue_ptr,
|
|
|
|
dev: ptr,
|
|
|
|
idx: 0,
|
|
|
|
ack_used_idx: 0,
|
|
|
|
read_only: ro, };
|
|
|
|
BLOCK_DEVICES[idx] = Some(bd);
|
|
|
|
|
|
|
|
// 8. Set the DRIVER_OK status bit. Device is now "live"
|
|
|
|
status_bits |= StatusField::DriverOk.val32();
|
2020-04-26 03:51:27 +04:00
|
|
|
ptr.add(MmioOffsets::Status.scale32())
|
|
|
|
.write_volatile(status_bits);
|
2020-03-16 04:18:46 +04:00
|
|
|
|
|
|
|
true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
pub fn fill_next_descriptor(bd: &mut BlockDevice, desc: Descriptor) -> u16 {
|
|
|
|
unsafe {
|
2020-04-26 03:51:27 +04:00
|
|
|
// The ring structure increments here first. This allows us to
|
|
|
|
// skip index 0, which then in the used ring will show that .id
|
|
|
|
// > 0. This is one way to error check. We will eventually get
|
|
|
|
// back to 0 as this index is cyclical. However, it shows if the
|
|
|
|
// first read/write actually works.
|
2020-03-16 04:18:46 +04:00
|
|
|
bd.idx = (bd.idx + 1) % VIRTIO_RING_SIZE as u16;
|
|
|
|
(*bd.queue).desc[bd.idx as usize] = desc;
|
2020-04-26 03:51:27 +04:00
|
|
|
if (*bd.queue).desc[bd.idx as usize].flags
|
|
|
|
& virtio::VIRTIO_DESC_F_NEXT
|
|
|
|
!= 0
|
|
|
|
{
|
2020-03-16 04:18:46 +04:00
|
|
|
// If the next flag is set, we need another descriptor.
|
2020-04-26 03:51:27 +04:00
|
|
|
(*bd.queue).desc[bd.idx as usize].next =
|
|
|
|
(bd.idx + 1) % VIRTIO_RING_SIZE as u16;
|
2020-03-16 04:18:46 +04:00
|
|
|
}
|
|
|
|
bd.idx
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/// This is now a common block operation for both reads and writes. Therefore,
|
|
|
|
/// when one thing needs to change, we can change it for both reads and writes.
|
|
|
|
/// There is a lot of error checking that I haven't done. The block device reads
|
2020-04-26 03:51:27 +04:00
|
|
|
/// sectors at a time, which are 512 bytes. Therefore, our buffer must be
|
|
|
|
/// capable of storing multiples of 512 bytes depending on the size. The size is
|
|
|
|
/// also a multiple of 512, but we don't really check that.
|
2020-03-16 04:18:46 +04:00
|
|
|
/// We DO however, check that we aren't writing to an R/O device. This would
|
|
|
|
/// cause a I/O error if we tried to write to a R/O device.
|
2020-04-25 02:46:31 +04:00
|
|
|
pub fn block_op(dev: usize,
|
|
|
|
buffer: *mut u8,
|
|
|
|
size: u32,
|
|
|
|
offset: u64,
|
|
|
|
write: bool,
|
|
|
|
watcher: u16)
|
|
|
|
-> Result<u32, BlockErrors>
|
|
|
|
{
|
2020-03-16 04:18:46 +04:00
|
|
|
unsafe {
|
|
|
|
if let Some(bdev) = BLOCK_DEVICES[dev - 1].as_mut() {
|
2020-04-26 03:51:27 +04:00
|
|
|
// Check to see if we are trying to write to a read only
|
|
|
|
// device.
|
2020-04-24 22:39:56 +04:00
|
|
|
if bdev.read_only && write {
|
2020-03-16 04:18:46 +04:00
|
|
|
println!("Trying to write to read/only!");
|
2020-04-24 22:39:56 +04:00
|
|
|
return Err(BlockErrors::ReadOnly);
|
|
|
|
}
|
|
|
|
if size % 512 != 0 {
|
|
|
|
return Err(BlockErrors::InvalidArgument);
|
2020-03-16 04:18:46 +04:00
|
|
|
}
|
|
|
|
let sector = offset / 512;
|
2020-04-26 03:51:27 +04:00
|
|
|
// TODO: Before we get here, we are NOT allowed to
|
|
|
|
// schedule a read or write OUTSIDE of the disk's size.
|
|
|
|
// So, we can read capacity from the configuration space
|
|
|
|
// to ensure we stay within bounds.
|
2020-03-16 04:18:46 +04:00
|
|
|
let blk_request_size = size_of::<Request>();
|
2020-04-26 03:51:27 +04:00
|
|
|
let blk_request =
|
|
|
|
kmalloc(blk_request_size) as *mut Request;
|
|
|
|
let desc =
|
|
|
|
Descriptor { addr: &(*blk_request).header
|
|
|
|
as *const Header
|
|
|
|
as u64,
|
|
|
|
len: size_of::<Header>() as u32,
|
|
|
|
flags: virtio::VIRTIO_DESC_F_NEXT,
|
|
|
|
next: 0, };
|
2020-03-16 04:18:46 +04:00
|
|
|
let head_idx = fill_next_descriptor(bdev, desc);
|
|
|
|
(*blk_request).header.sector = sector;
|
2020-04-26 03:51:27 +04:00
|
|
|
// A write is an "out" direction, whereas a read is an
|
|
|
|
// "in" direction.
|
2020-04-24 22:39:56 +04:00
|
|
|
(*blk_request).header.blktype = if write {
|
2020-03-16 04:18:46 +04:00
|
|
|
VIRTIO_BLK_T_OUT
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
VIRTIO_BLK_T_IN
|
|
|
|
};
|
2020-04-26 03:51:27 +04:00
|
|
|
// We put 111 in the status. Whenever the device
|
|
|
|
// finishes, it will write into status. If we read
|
|
|
|
// status and it is 111, we know that it wasn't written
|
|
|
|
// to by the device.
|
2020-03-16 04:18:46 +04:00
|
|
|
(*blk_request).data.data = buffer;
|
|
|
|
(*blk_request).header.reserved = 0;
|
|
|
|
(*blk_request).status.status = 111;
|
2020-04-24 22:39:56 +04:00
|
|
|
(*blk_request).watcher = watcher;
|
2020-04-26 03:51:27 +04:00
|
|
|
let desc =
|
|
|
|
Descriptor { addr: buffer as u64,
|
|
|
|
len: size,
|
|
|
|
flags: virtio::VIRTIO_DESC_F_NEXT
|
|
|
|
| if !write {
|
|
|
|
virtio::VIRTIO_DESC_F_WRITE
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
0
|
|
|
|
},
|
|
|
|
next: 0, };
|
2020-03-16 04:18:46 +04:00
|
|
|
let _data_idx = fill_next_descriptor(bdev, desc);
|
2020-04-26 03:51:27 +04:00
|
|
|
let desc =
|
|
|
|
Descriptor { addr: &(*blk_request).status
|
|
|
|
as *const Status
|
|
|
|
as u64,
|
|
|
|
len: size_of::<Status>() as u32,
|
|
|
|
flags: virtio::VIRTIO_DESC_F_WRITE,
|
|
|
|
next: 0, };
|
2020-03-16 04:18:46 +04:00
|
|
|
let _status_idx = fill_next_descriptor(bdev, desc);
|
2020-04-26 03:51:27 +04:00
|
|
|
(*bdev.queue).avail.ring[(*bdev.queue).avail.idx
|
|
|
|
as usize
|
|
|
|
% virtio::VIRTIO_RING_SIZE] = head_idx;
|
|
|
|
(*bdev.queue).avail.idx =
|
|
|
|
(*bdev.queue).avail.idx.wrapping_add(1);
|
|
|
|
// The only queue a block device has is 0, which is the
|
|
|
|
// request queue.
|
|
|
|
bdev.dev
|
|
|
|
.add(MmioOffsets::QueueNotify.scale32())
|
|
|
|
.write_volatile(0);
|
2020-04-24 22:39:56 +04:00
|
|
|
Ok(size)
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
Err(BlockErrors::BlockDeviceNotFound)
|
2020-03-16 04:18:46 +04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-04-26 03:51:27 +04:00
|
|
|
pub fn read(dev: usize,
|
|
|
|
buffer: *mut u8,
|
|
|
|
size: u32,
|
|
|
|
offset: u64)
|
|
|
|
-> Result<u32, BlockErrors>
|
|
|
|
{
|
2020-04-24 22:39:56 +04:00
|
|
|
block_op(dev, buffer, size, offset, false, 0)
|
2020-03-16 04:18:46 +04:00
|
|
|
}
|
|
|
|
|
2020-04-26 03:51:27 +04:00
|
|
|
pub fn write(dev: usize,
|
|
|
|
buffer: *mut u8,
|
|
|
|
size: u32,
|
|
|
|
offset: u64)
|
|
|
|
-> Result<u32, BlockErrors>
|
|
|
|
{
|
2020-04-24 22:39:56 +04:00
|
|
|
block_op(dev, buffer, size, offset, true, 0)
|
2020-03-16 04:18:46 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
/// Here we handle block specific interrupts. Here, we need to check
|
|
|
|
/// the used ring and wind it up until we've handled everything.
|
|
|
|
/// This is how the device tells us that it's finished a request.
|
|
|
|
pub fn pending(bd: &mut BlockDevice) {
|
|
|
|
// Here we need to check the used ring and then free the resources
|
|
|
|
// given by the descriptor id.
|
|
|
|
unsafe {
|
|
|
|
let ref queue = *bd.queue;
|
|
|
|
while bd.ack_used_idx != queue.used.idx {
|
2020-04-26 03:51:27 +04:00
|
|
|
let ref elem = queue.used.ring
|
|
|
|
[bd.ack_used_idx as usize % VIRTIO_RING_SIZE];
|
2020-04-13 05:37:06 +04:00
|
|
|
bd.ack_used_idx = bd.ack_used_idx.wrapping_add(1);
|
2020-04-26 03:51:27 +04:00
|
|
|
// Requests stay resident on the heap until this
|
|
|
|
// function, so we can recapture the address here
|
|
|
|
let rq = queue.desc[elem.id as usize].addr
|
|
|
|
as *const Request;
|
2020-04-24 22:39:56 +04:00
|
|
|
|
2020-04-26 03:51:27 +04:00
|
|
|
// A process might be waiting for this interrupt. Awaken
|
|
|
|
// the process attached here.
|
2020-04-24 22:39:56 +04:00
|
|
|
let pid_of_watcher = (*rq).watcher;
|
|
|
|
// A PID of 0 means that we don't have a watcher.
|
|
|
|
if pid_of_watcher > 0 {
|
|
|
|
set_running(pid_of_watcher);
|
2020-04-25 23:15:25 +04:00
|
|
|
let proc = get_by_pid(pid_of_watcher);
|
2020-04-26 03:51:27 +04:00
|
|
|
(*(*proc).get_frame_mut()).regs[10] =
|
|
|
|
(*rq).status.status as usize;
|
|
|
|
// TODO: Set GpA0 to the value of the return
|
|
|
|
// status.
|
2020-04-24 22:39:56 +04:00
|
|
|
}
|
2020-03-16 04:18:46 +04:00
|
|
|
kfree(rq as *mut u8);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/// The trap code will route PLIC interrupts 1..=8 for virtio devices. When
|
|
|
|
/// virtio determines that this is a block device, it sends it here.
|
|
|
|
pub fn handle_interrupt(idx: usize) {
|
|
|
|
unsafe {
|
|
|
|
if let Some(bdev) = BLOCK_DEVICES[idx].as_mut() {
|
|
|
|
pending(bdev);
|
|
|
|
}
|
|
|
|
else {
|
2020-04-26 03:51:27 +04:00
|
|
|
println!(
|
|
|
|
"Invalid block device for interrupt {}",
|
|
|
|
idx + 1
|
|
|
|
);
|
2020-03-16 04:18:46 +04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2020-04-24 22:39:56 +04:00
|
|
|
|
2020-04-25 02:37:48 +04:00
|
|
|
// ///////////////////////////////////////////////
|
|
|
|
// // BLOCK PROCESSES (KERNEL PROCESSES)
|
|
|
|
// ///////////////////////////////////////////////
|
|
|
|
struct ProcArgs {
|
2020-04-25 02:46:31 +04:00
|
|
|
pub pid: u16,
|
|
|
|
pub dev: usize,
|
2020-04-25 02:37:48 +04:00
|
|
|
pub buffer: *mut u8,
|
2020-04-25 02:46:31 +04:00
|
|
|
pub size: u32,
|
2020-04-25 02:37:48 +04:00
|
|
|
pub offset: u64,
|
|
|
|
}
|
|
|
|
|
2020-04-25 02:46:31 +04:00
|
|
|
/// This will be a
|
2020-04-25 02:37:48 +04:00
|
|
|
fn read_proc(args_addr: usize) {
|
|
|
|
let args_ptr = args_addr as *mut ProcArgs;
|
|
|
|
let args = unsafe { args_ptr.as_ref().unwrap() };
|
2020-04-26 03:51:27 +04:00
|
|
|
let _ = block_op(
|
|
|
|
args.dev,
|
|
|
|
args.buffer,
|
|
|
|
args.size,
|
|
|
|
args.offset,
|
|
|
|
false,
|
|
|
|
args.pid,
|
|
|
|
);
|
2020-04-25 02:37:48 +04:00
|
|
|
tfree(args_ptr);
|
2020-04-25 03:16:45 +04:00
|
|
|
// This should be handled by the RA now.
|
|
|
|
// syscall_exit();
|
2020-04-25 02:37:48 +04:00
|
|
|
}
|
|
|
|
|
2020-04-26 03:51:27 +04:00
|
|
|
pub fn process_read(pid: u16,
|
|
|
|
dev: usize,
|
|
|
|
buffer: *mut u8,
|
|
|
|
size: u32,
|
|
|
|
offset: u64)
|
|
|
|
{
|
|
|
|
// println!("Block read {}, {}, 0x{:x}, {}, {}", pid, dev, buffer as
|
|
|
|
// usize, size, offset);
|
2020-04-25 02:37:48 +04:00
|
|
|
let args = talloc::<ProcArgs>().unwrap();
|
|
|
|
args.pid = pid;
|
|
|
|
args.dev = dev;
|
|
|
|
args.buffer = buffer;
|
|
|
|
args.size = size;
|
|
|
|
args.offset = offset;
|
2020-04-24 22:39:56 +04:00
|
|
|
set_waiting(pid);
|
2020-04-26 03:51:27 +04:00
|
|
|
let _ = add_kernel_process_args(
|
|
|
|
read_proc,
|
|
|
|
args as *mut ProcArgs as usize,
|
|
|
|
);
|
2020-04-25 02:37:48 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
fn write_proc(args_addr: usize) {
|
|
|
|
let args_ptr = args_addr as *mut ProcArgs;
|
|
|
|
let args = unsafe { args_ptr.as_ref().unwrap() };
|
|
|
|
|
2020-04-26 03:51:27 +04:00
|
|
|
let _ = block_op(
|
|
|
|
args.dev,
|
|
|
|
args.buffer,
|
|
|
|
args.size,
|
|
|
|
args.offset,
|
|
|
|
true,
|
|
|
|
args.pid,
|
|
|
|
);
|
2020-04-25 02:37:48 +04:00
|
|
|
tfree(args_ptr);
|
2020-04-25 03:16:45 +04:00
|
|
|
// syscall_exit();
|
2020-04-24 22:39:56 +04:00
|
|
|
}
|
|
|
|
|
2020-04-26 03:51:27 +04:00
|
|
|
pub fn process_write(pid: u16,
|
|
|
|
dev: usize,
|
|
|
|
buffer: *mut u8,
|
|
|
|
size: u32,
|
|
|
|
offset: u64)
|
|
|
|
{
|
2020-04-25 02:37:48 +04:00
|
|
|
let args = talloc::<ProcArgs>().unwrap();
|
|
|
|
args.pid = pid;
|
|
|
|
args.dev = dev;
|
|
|
|
args.buffer = buffer;
|
|
|
|
args.size = size;
|
|
|
|
args.offset = offset;
|
2020-04-24 22:39:56 +04:00
|
|
|
set_waiting(pid);
|
2020-04-26 03:51:27 +04:00
|
|
|
let _ = add_kernel_process_args(
|
|
|
|
write_proc,
|
|
|
|
args as *mut ProcArgs as usize,
|
|
|
|
);
|
2020-04-25 02:46:31 +04:00
|
|
|
}
|