1
0
mirror of https://github.com/rcore-os/rCore.git synced 2024-11-22 16:16:16 +04:00

merge interrupt & context switch for rv32 & rv64

This commit is contained in:
WangRunji 2018-12-27 19:09:33 +08:00
parent de24f6673c
commit b3a8e95d78
3 changed files with 68 additions and 128 deletions

View File

@ -15,50 +15,26 @@ pub unsafe fn restore(flags: usize) {
}
#[inline(always)]
#[cfg(target_arch = "riscv32")]
#[cfg(any(target_arch = "riscv32", target_arch = "riscv64"))]
pub unsafe fn disable_and_store() -> usize {
if option_env!("m_mode").is_some() {
let mstatus: usize;
asm!("csrrci $0, 0x300, 1 << 3" : "=r"(mstatus));
asm!("csrci mstatus, 1 << 3" : "=r"(mstatus));
mstatus & (1 << 3)
} else {
let sstatus: usize;
asm!("csrrci $0, 0x100, 1 << 1" : "=r"(sstatus));
asm!("csrci sstatus, 1 << 1" : "=r"(sstatus));
sstatus & (1 << 1)
}
}
#[inline(always)]
#[cfg(target_arch = "riscv64")]
pub unsafe fn disable_and_store() -> usize {
if option_env!("m_mode").is_some() {
let mstatus: usize;
asm!("csrrci $0, 0x300, 1 << 3" : "=r"(mstatus));
mstatus & (1 << 3)
} else {
let sstatus: usize;
asm!("csrrci $0, 0x100, 1 << 1" : "=r"(sstatus));
sstatus & (1 << 1)
}
}
#[inline(always)]
#[cfg(target_arch = "riscv32")]
#[cfg(any(target_arch = "riscv32", target_arch = "riscv64"))]
pub unsafe fn restore(flags: usize) {
if option_env!("m_mode").is_some() {
asm!("csrs 0x300, $0" :: "r"(flags));
asm!("csrs mstatus, $0" :: "r"(flags));
} else {
asm!("csrs 0x100, $0" :: "r"(flags));
}
}
#[inline(always)]
#[cfg(target_arch = "riscv64")]
pub unsafe fn restore(flags: usize) {
if option_env!("m_mode").is_some() {
asm!("csrs 0x300, $0" :: "r"(flags));
} else {
asm!("csrs 0x100, $0" :: "r"(flags));
asm!("csrs sstatus, $0" :: "r"(flags));
}
}

View File

@ -148,105 +148,70 @@ pub struct Context(usize);
impl Context {
/// Switch to another kernel thread.
///
/// Defined in `trap.asm`.
///
/// Push all callee-saved registers at the current kernel stack.
/// Store current sp, switch to target.
/// Pop all callee-saved registers, then return to the target.
#[cfg(target_arch = "riscv32")]
#[naked]
#[inline(never)]
pub unsafe extern fn switch(&mut self, target: &mut Self) {
asm!(
"
#[cfg(target_arch = "riscv32")]
asm!(r"
.equ XLENB, 4
.macro Load reg, mem
lw \reg, \mem
.endm
.macro Store reg, mem
sw \reg, \mem
.endm");
#[cfg(target_arch = "riscv64")]
asm!(r"
.equ XLENB, 8
.macro Load reg, mem
ld \reg, \mem
.endm
.macro Store reg, mem
sd \reg, \mem
.endm");
asm!("
// save from's registers
addi sp, sp, -4*14
sw sp, 0(a0)
sw ra, 0*4(sp)
sw s0, 2*4(sp)
sw s1, 3*4(sp)
sw s2, 4*4(sp)
sw s3, 5*4(sp)
sw s4, 6*4(sp)
sw s5, 7*4(sp)
sw s6, 8*4(sp)
sw s7, 9*4(sp)
sw s8, 10*4(sp)
sw s9, 11*4(sp)
sw s10, 12*4(sp)
sw s11, 13*4(sp)
csrrs s11, 0x180, x0 // satp
sw s11, 1*4(sp)
addi sp, sp, (-XLENB*14)
Store sp, 0(a0)
Store ra, 0*XLENB(sp)
Store s0, 2*XLENB(sp)
Store s1, 3*XLENB(sp)
Store s2, 4*XLENB(sp)
Store s3, 5*XLENB(sp)
Store s4, 6*XLENB(sp)
Store s5, 7*XLENB(sp)
Store s6, 8*XLENB(sp)
Store s7, 9*XLENB(sp)
Store s8, 10*XLENB(sp)
Store s9, 11*XLENB(sp)
Store s10, 12*XLENB(sp)
Store s11, 13*XLENB(sp)
csrr s11, satp
Store s11, 1*XLENB(sp)
// restore to's registers
lw sp, 0(a1)
lw s11, 1*4(sp)
csrrw x0, 0x180, s11 // satp
lw ra, 0*4(sp)
lw s0, 2*4(sp)
lw s1, 3*4(sp)
lw s2, 4*4(sp)
lw s3, 5*4(sp)
lw s4, 6*4(sp)
lw s5, 7*4(sp)
lw s6, 8*4(sp)
lw s7, 9*4(sp)
lw s8, 10*4(sp)
lw s9, 11*4(sp)
lw s10, 12*4(sp)
lw s11, 13*4(sp)
addi sp, sp, 4*14
Load sp, 0(a1)
Load s11, 1*XLENB(sp)
csrw satp, s11
Load ra, 0*XLENB(sp)
Load s0, 2*XLENB(sp)
Load s1, 3*XLENB(sp)
Load s2, 4*XLENB(sp)
Load s3, 5*XLENB(sp)
Load s4, 6*XLENB(sp)
Load s5, 7*XLENB(sp)
Load s6, 8*XLENB(sp)
Load s7, 9*XLENB(sp)
Load s8, 10*XLENB(sp)
Load s9, 11*XLENB(sp)
Load s10, 12*XLENB(sp)
Load s11, 13*XLENB(sp)
addi sp, sp, (XLENB*14)
sw zero, 0(a1)
ret"
: : : : "volatile" )
}
#[cfg(target_arch = "riscv64")]
#[naked]
#[inline(never)]
pub unsafe extern fn switch(&mut self, target: &mut Self) {
asm!(
"
// save from's registers
addi sp, sp, -8*14
sd sp, 0(a0)
sd ra, 0*8(sp)
sd s0, 2*8(sp)
sd s1, 3*8(sp)
sd s2, 4*8(sp)
sd s3, 5*8(sp)
sd s4, 6*8(sp)
sd s5, 7*8(sp)
sd s6, 8*8(sp)
sd s7, 9*8(sp)
sd s8, 10*8(sp)
sd s9, 11*8(sp)
sd s10, 12*8(sp)
sd s11, 13*8(sp)
csrrs s11, 0x180, x0 // satp
sd s11, 1*8(sp)
// restore to's registers
ld sp, 0(a1)
ld s11, 1*8(sp)
csrrw x0, 0x180, s11 // satp
ld ra, 0*8(sp)
ld s0, 2*8(sp)
ld s1, 3*8(sp)
ld s2, 4*8(sp)
ld s3, 5*8(sp)
ld s4, 6*8(sp)
ld s5, 7*8(sp)
ld s6, 8*8(sp)
ld s7, 9*8(sp)
ld s8, 10*8(sp)
ld s9, 11*8(sp)
ld s10, 12*8(sp)
ld s11, 13*8(sp)
addi sp, sp, 8*14
sd zero, 0(a1)
Store zero, 0(a1)
ret"
: : : : "volatile" )
}

View File

@ -74,29 +74,28 @@ global_asm!("
.macro TEST_BACK_TO_KERNEL
andi s0, s1, 1 << 8 // sstatus.SPP = 1
.endm
");
#[cfg(target_pointer_width = "32")]
global_asm!("
#[cfg(target_arch = "riscv32")]
global_asm!(r"
.equ XLENB, 4
.equ XLENb, 32
.macro LOAD a1, a2
lw \\a1, \\a2*XLENB(sp)
lw \a1, \a2*XLENB(sp)
.endm
.macro STORE a1, a2
sw \\a1, \\a2*XLENB(sp)
sw \a1, \a2*XLENB(sp)
.endm
");
#[cfg(target_pointer_width = "64")]
global_asm!("
#[cfg(target_arch = "riscv64")]
global_asm!(r"
.equ XLENB, 8
.equ XLENb, 64
.macro LOAD a1, a2
ld \\a1, \\a2*XLENB(sp)
ld \a1, \a2*XLENB(sp)
.endm
.macro STORE a1, a2
sd \\a1, \\a2*XLENB(sp)
sd \a1, \a2*XLENB(sp)
.endm
");