Merge recent updates from ch8

This commit is contained in:
Yifan Wu 2022-01-26 01:53:39 -08:00
commit e1ae9b6236
99 changed files with 1959 additions and 1437 deletions

1
.gitignore vendored
View file

@ -2,6 +2,7 @@
os/target/* os/target/*
os/.idea/* os/.idea/*
os/src/link_app.S os/src/link_app.S
os/src/linker.ld
os/last-* os/last-*
os/Cargo.lock os/Cargo.lock
os/last-* os/last-*

View file

@ -239,7 +239,6 @@ Here are the updates since 3.5.0:
* [ ] rewrite practice doc and remove some inproper questions * [ ] rewrite practice doc and remove some inproper questions
* [ ] provide smooth debug experience at a Rust source code level * [ ] provide smooth debug experience at a Rust source code level
* [ ] format the code using official tools * [ ] format the code using official tools
* [ ] support other platforms
### Crates ### Crates

View file

@ -1,12 +1,9 @@
use easy_fs::{ use clap::{App, Arg};
BlockDevice, use easy_fs::{BlockDevice, EasyFileSystem};
EasyFileSystem, use std::fs::{read_dir, File, OpenOptions};
}; use std::io::{Read, Seek, SeekFrom, Write};
use std::fs::{File, OpenOptions, read_dir};
use std::io::{Read, Write, Seek, SeekFrom};
use std::sync::Mutex;
use std::sync::Arc; use std::sync::Arc;
use clap::{Arg, App}; use std::sync::Mutex;
const BLOCK_SZ: usize = 512; const BLOCK_SZ: usize = 512;
@ -34,17 +31,19 @@ fn main() {
fn easy_fs_pack() -> std::io::Result<()> { fn easy_fs_pack() -> std::io::Result<()> {
let matches = App::new("EasyFileSystem packer") let matches = App::new("EasyFileSystem packer")
.arg(Arg::with_name("source") .arg(
.short("s") Arg::with_name("source")
.long("source") .short("s")
.takes_value(true) .long("source")
.help("Executable source dir(with backslash)") .takes_value(true)
.help("Executable source dir(with backslash)"),
) )
.arg(Arg::with_name("target") .arg(
.short("t") Arg::with_name("target")
.long("target") .short("t")
.takes_value(true) .long("target")
.help("Executable target dir(with backslash)") .takes_value(true)
.help("Executable target dir(with backslash)"),
) )
.get_matches(); .get_matches();
let src_path = matches.value_of("source").unwrap(); let src_path = matches.value_of("source").unwrap();
@ -60,11 +59,7 @@ fn easy_fs_pack() -> std::io::Result<()> {
f f
}))); })));
// 16MiB, at most 4095 files // 16MiB, at most 4095 files
let efs = EasyFileSystem::create( let efs = EasyFileSystem::create(block_file, 16 * 2048, 1);
block_file.clone(),
16 * 2048,
1,
);
let root_inode = Arc::new(EasyFileSystem::root_inode(&efs)); let root_inode = Arc::new(EasyFileSystem::root_inode(&efs));
let apps: Vec<_> = read_dir(src_path) let apps: Vec<_> = read_dir(src_path)
.unwrap() .unwrap()
@ -103,11 +98,7 @@ fn efs_test() -> std::io::Result<()> {
f.set_len(8192 * 512).unwrap(); f.set_len(8192 * 512).unwrap();
f f
}))); })));
EasyFileSystem::create( EasyFileSystem::create(block_file.clone(), 4096, 1);
block_file.clone(),
4096,
1,
);
let efs = EasyFileSystem::open(block_file.clone()); let efs = EasyFileSystem::open(block_file.clone());
let root_inode = EasyFileSystem::root_inode(&efs); let root_inode = EasyFileSystem::root_inode(&efs);
root_inode.create("filea"); root_inode.create("filea");
@ -121,17 +112,11 @@ fn efs_test() -> std::io::Result<()> {
//let mut buffer = [0u8; 512]; //let mut buffer = [0u8; 512];
let mut buffer = [0u8; 233]; let mut buffer = [0u8; 233];
let len = filea.read_at(0, &mut buffer); let len = filea.read_at(0, &mut buffer);
assert_eq!( assert_eq!(greet_str, core::str::from_utf8(&buffer[..len]).unwrap(),);
greet_str,
core::str::from_utf8(&buffer[..len]).unwrap(),
);
let mut random_str_test = |len: usize| { let mut random_str_test = |len: usize| {
filea.clear(); filea.clear();
assert_eq!( assert_eq!(filea.read_at(0, &mut buffer), 0,);
filea.read_at(0, &mut buffer),
0,
);
let mut str = String::new(); let mut str = String::new();
use rand; use rand;
// random digit // random digit
@ -148,9 +133,7 @@ fn efs_test() -> std::io::Result<()> {
break; break;
} }
offset += len; offset += len;
read_str.push_str( read_str.push_str(core::str::from_utf8(&read_buffer[..len]).unwrap());
core::str::from_utf8(&read_buffer[..len]).unwrap()
);
} }
assert_eq!(str, read_str); assert_eq!(str, read_str);
}; };

View file

@ -1,9 +1,5 @@
use super::{get_block_cache, BlockDevice, BLOCK_SZ};
use alloc::sync::Arc; use alloc::sync::Arc;
use super::{
BlockDevice,
BLOCK_SZ,
get_block_cache,
};
type BitmapBlock = [u64; 64]; type BitmapBlock = [u64; 64];
@ -17,7 +13,7 @@ pub struct Bitmap {
/// Return (block_pos, bits64_pos, inner_pos) /// Return (block_pos, bits64_pos, inner_pos)
fn decomposition(mut bit: usize) -> (usize, usize, usize) { fn decomposition(mut bit: usize) -> (usize, usize, usize) {
let block_pos = bit / BLOCK_BITS; let block_pos = bit / BLOCK_BITS;
bit = bit % BLOCK_BITS; bit %= BLOCK_BITS;
(block_pos, bit / 64, bit % 64) (block_pos, bit / 64, bit % 64)
} }
@ -34,14 +30,15 @@ impl Bitmap {
let pos = get_block_cache( let pos = get_block_cache(
block_id + self.start_block_id as usize, block_id + self.start_block_id as usize,
Arc::clone(block_device), Arc::clone(block_device),
).lock().modify(0, |bitmap_block: &mut BitmapBlock| { )
.lock()
.modify(0, |bitmap_block: &mut BitmapBlock| {
if let Some((bits64_pos, inner_pos)) = bitmap_block if let Some((bits64_pos, inner_pos)) = bitmap_block
.iter() .iter()
.enumerate() .enumerate()
.find(|(_, bits64)| **bits64 != u64::MAX) .find(|(_, bits64)| **bits64 != u64::MAX)
.map(|(bits64_pos, bits64)| { .map(|(bits64_pos, bits64)| (bits64_pos, bits64.trailing_ones() as usize))
(bits64_pos, bits64.trailing_ones() as usize) {
}) {
// modify cache // modify cache
bitmap_block[bits64_pos] |= 1u64 << inner_pos; bitmap_block[bits64_pos] |= 1u64 << inner_pos;
Some(block_id * BLOCK_BITS + bits64_pos * 64 + inner_pos as usize) Some(block_id * BLOCK_BITS + bits64_pos * 64 + inner_pos as usize)
@ -58,13 +55,12 @@ impl Bitmap {
pub fn dealloc(&self, block_device: &Arc<dyn BlockDevice>, bit: usize) { pub fn dealloc(&self, block_device: &Arc<dyn BlockDevice>, bit: usize) {
let (block_pos, bits64_pos, inner_pos) = decomposition(bit); let (block_pos, bits64_pos, inner_pos) = decomposition(bit);
get_block_cache( get_block_cache(block_pos + self.start_block_id, Arc::clone(block_device))
block_pos + self.start_block_id, .lock()
Arc::clone(block_device) .modify(0, |bitmap_block: &mut BitmapBlock| {
).lock().modify(0, |bitmap_block: &mut BitmapBlock| { assert!(bitmap_block[bits64_pos] & (1u64 << inner_pos) > 0);
assert!(bitmap_block[bits64_pos] & (1u64 << inner_pos) > 0); bitmap_block[bits64_pos] -= 1u64 << inner_pos;
bitmap_block[bits64_pos] -= 1u64 << inner_pos; });
});
} }
pub fn maximum(&self) -> usize { pub fn maximum(&self) -> usize {

View file

@ -1,7 +1,4 @@
use super::{ use super::{BlockDevice, BLOCK_SZ};
BLOCK_SZ,
BlockDevice,
};
use alloc::collections::VecDeque; use alloc::collections::VecDeque;
use alloc::sync::Arc; use alloc::sync::Arc;
use lazy_static::*; use lazy_static::*;
@ -16,10 +13,7 @@ pub struct BlockCache {
impl BlockCache { impl BlockCache {
/// Load a new BlockCache from disk. /// Load a new BlockCache from disk.
pub fn new( pub fn new(block_id: usize, block_device: Arc<dyn BlockDevice>) -> Self {
block_id: usize,
block_device: Arc<dyn BlockDevice>
) -> Self {
let mut cache = [0u8; BLOCK_SZ]; let mut cache = [0u8; BLOCK_SZ];
block_device.read_block(block_id, &mut cache); block_device.read_block(block_id, &mut cache);
Self { Self {
@ -34,14 +28,20 @@ impl BlockCache {
&self.cache[offset] as *const _ as usize &self.cache[offset] as *const _ as usize
} }
pub fn get_ref<T>(&self, offset: usize) -> &T where T: Sized { pub fn get_ref<T>(&self, offset: usize) -> &T
where
T: Sized,
{
let type_size = core::mem::size_of::<T>(); let type_size = core::mem::size_of::<T>();
assert!(offset + type_size <= BLOCK_SZ); assert!(offset + type_size <= BLOCK_SZ);
let addr = self.addr_of_offset(offset); let addr = self.addr_of_offset(offset);
unsafe { &*(addr as *const T) } unsafe { &*(addr as *const T) }
} }
pub fn get_mut<T>(&mut self, offset: usize) -> &mut T where T: Sized { pub fn get_mut<T>(&mut self, offset: usize) -> &mut T
where
T: Sized,
{
let type_size = core::mem::size_of::<T>(); let type_size = core::mem::size_of::<T>();
assert!(offset + type_size <= BLOCK_SZ); assert!(offset + type_size <= BLOCK_SZ);
self.modified = true; self.modified = true;
@ -53,7 +53,7 @@ impl BlockCache {
f(self.get_ref(offset)) f(self.get_ref(offset))
} }
pub fn modify<T, V>(&mut self, offset:usize, f: impl FnOnce(&mut T) -> V) -> V { pub fn modify<T, V>(&mut self, offset: usize, f: impl FnOnce(&mut T) -> V) -> V {
f(self.get_mut(offset)) f(self.get_mut(offset))
} }
@ -79,7 +79,9 @@ pub struct BlockCacheManager {
impl BlockCacheManager { impl BlockCacheManager {
pub fn new() -> Self { pub fn new() -> Self {
Self { queue: VecDeque::new() } Self {
queue: VecDeque::new(),
}
} }
pub fn get_block_cache( pub fn get_block_cache(
@ -87,27 +89,28 @@ impl BlockCacheManager {
block_id: usize, block_id: usize,
block_device: Arc<dyn BlockDevice>, block_device: Arc<dyn BlockDevice>,
) -> Arc<Mutex<BlockCache>> { ) -> Arc<Mutex<BlockCache>> {
if let Some(pair) = self.queue if let Some(pair) = self.queue.iter().find(|pair| pair.0 == block_id) {
.iter() Arc::clone(&pair.1)
.find(|pair| pair.0 == block_id) {
Arc::clone(&pair.1)
} else { } else {
// substitute // substitute
if self.queue.len() == BLOCK_CACHE_SIZE { if self.queue.len() == BLOCK_CACHE_SIZE {
// from front to tail // from front to tail
if let Some((idx, _)) = self.queue if let Some((idx, _)) = self
.queue
.iter() .iter()
.enumerate() .enumerate()
.find(|(_, pair)| Arc::strong_count(&pair.1) == 1) { .find(|(_, pair)| Arc::strong_count(&pair.1) == 1)
{
self.queue.drain(idx..=idx); self.queue.drain(idx..=idx);
} else { } else {
panic!("Run out of BlockCache!"); panic!("Run out of BlockCache!");
} }
} }
// load block into mem and push back // load block into mem and push back
let block_cache = Arc::new(Mutex::new( let block_cache = Arc::new(Mutex::new(BlockCache::new(
BlockCache::new(block_id, Arc::clone(&block_device)) block_id,
)); Arc::clone(&block_device),
)));
self.queue.push_back((block_id, Arc::clone(&block_cache))); self.queue.push_back((block_id, Arc::clone(&block_cache)));
block_cache block_cache
} }
@ -115,16 +118,17 @@ impl BlockCacheManager {
} }
lazy_static! { lazy_static! {
pub static ref BLOCK_CACHE_MANAGER: Mutex<BlockCacheManager> = Mutex::new( pub static ref BLOCK_CACHE_MANAGER: Mutex<BlockCacheManager> =
BlockCacheManager::new() Mutex::new(BlockCacheManager::new());
);
} }
pub fn get_block_cache( pub fn get_block_cache(
block_id: usize, block_id: usize,
block_device: Arc<dyn BlockDevice> block_device: Arc<dyn BlockDevice>,
) -> Arc<Mutex<BlockCache>> { ) -> Arc<Mutex<BlockCache>> {
BLOCK_CACHE_MANAGER.lock().get_block_cache(block_id, block_device) BLOCK_CACHE_MANAGER
.lock()
.get_block_cache(block_id, block_device)
} }
pub fn block_cache_sync_all() { pub fn block_cache_sync_all() {

View file

@ -1,6 +1,6 @@
use core::any::Any; use core::any::Any;
pub trait BlockDevice : Send + Sync + Any { pub trait BlockDevice: Send + Sync + Any {
fn read_block(&self, block_id: usize, buf: &mut [u8]); fn read_block(&self, block_id: usize, buf: &mut [u8]);
fn write_block(&self, block_id: usize, buf: &[u8]); fn write_block(&self, block_id: usize, buf: &[u8]);
} }

View file

@ -1,16 +1,10 @@
use alloc::sync::Arc;
use spin::Mutex;
use super::{ use super::{
BlockDevice, block_cache_sync_all, get_block_cache, Bitmap, BlockDevice, DiskInode, DiskInodeType, Inode,
Bitmap,
SuperBlock, SuperBlock,
DiskInode,
DiskInodeType,
Inode,
get_block_cache,
block_cache_sync_all,
}; };
use crate::BLOCK_SZ; use crate::BLOCK_SZ;
use alloc::sync::Arc;
use spin::Mutex;
pub struct EasyFileSystem { pub struct EasyFileSystem {
pub block_device: Arc<dyn BlockDevice>, pub block_device: Arc<dyn BlockDevice>,
@ -50,39 +44,36 @@ impl EasyFileSystem {
}; };
// clear all blocks // clear all blocks
for i in 0..total_blocks { for i in 0..total_blocks {
get_block_cache( get_block_cache(i as usize, Arc::clone(&block_device))
i as usize, .lock()
Arc::clone(&block_device) .modify(0, |data_block: &mut DataBlock| {
) for byte in data_block.iter_mut() {
.lock() *byte = 0;
.modify(0, |data_block: &mut DataBlock| { }
for byte in data_block.iter_mut() { *byte = 0; } });
});
} }
// initialize SuperBlock // initialize SuperBlock
get_block_cache(0, Arc::clone(&block_device)) get_block_cache(0, Arc::clone(&block_device)).lock().modify(
.lock() 0,
.modify(0, |super_block: &mut SuperBlock| { |super_block: &mut SuperBlock| {
super_block.initialize( super_block.initialize(
total_blocks, total_blocks,
inode_bitmap_blocks, inode_bitmap_blocks,
inode_area_blocks, inode_area_blocks,
data_bitmap_blocks, data_bitmap_blocks,
data_area_blocks, data_area_blocks,
); );
}); },
);
// write back immediately // write back immediately
// create a inode for root node "/" // create a inode for root node "/"
assert_eq!(efs.alloc_inode(), 0); assert_eq!(efs.alloc_inode(), 0);
let (root_inode_block_id, root_inode_offset) = efs.get_disk_inode_pos(0); let (root_inode_block_id, root_inode_offset) = efs.get_disk_inode_pos(0);
get_block_cache( get_block_cache(root_inode_block_id as usize, Arc::clone(&block_device))
root_inode_block_id as usize, .lock()
Arc::clone(&block_device) .modify(root_inode_offset, |disk_inode: &mut DiskInode| {
) disk_inode.initialize(DiskInodeType::Directory);
.lock() });
.modify(root_inode_offset, |disk_inode: &mut DiskInode| {
disk_inode.initialize(DiskInodeType::Directory);
});
block_cache_sync_all(); block_cache_sync_all();
Arc::new(Mutex::new(efs)) Arc::new(Mutex::new(efs))
} }
@ -97,10 +88,7 @@ impl EasyFileSystem {
super_block.inode_bitmap_blocks + super_block.inode_area_blocks; super_block.inode_bitmap_blocks + super_block.inode_area_blocks;
let efs = Self { let efs = Self {
block_device, block_device,
inode_bitmap: Bitmap::new( inode_bitmap: Bitmap::new(1, super_block.inode_bitmap_blocks as usize),
1,
super_block.inode_bitmap_blocks as usize
),
data_bitmap: Bitmap::new( data_bitmap: Bitmap::new(
(1 + inode_total_blocks) as usize, (1 + inode_total_blocks) as usize,
super_block.data_bitmap_blocks as usize, super_block.data_bitmap_blocks as usize,
@ -117,19 +105,17 @@ impl EasyFileSystem {
// acquire efs lock temporarily // acquire efs lock temporarily
let (block_id, block_offset) = efs.lock().get_disk_inode_pos(0); let (block_id, block_offset) = efs.lock().get_disk_inode_pos(0);
// release efs lock // release efs lock
Inode::new( Inode::new(block_id, block_offset, Arc::clone(efs), block_device)
block_id,
block_offset,
Arc::clone(efs),
block_device,
)
} }
pub fn get_disk_inode_pos(&self, inode_id: u32) -> (u32, usize) { pub fn get_disk_inode_pos(&self, inode_id: u32) -> (u32, usize) {
let inode_size = core::mem::size_of::<DiskInode>(); let inode_size = core::mem::size_of::<DiskInode>();
let inodes_per_block = (BLOCK_SZ / inode_size) as u32; let inodes_per_block = (BLOCK_SZ / inode_size) as u32;
let block_id = self.inode_area_start_block + inode_id / inodes_per_block; let block_id = self.inode_area_start_block + inode_id / inodes_per_block;
(block_id, (inode_id % inodes_per_block) as usize * inode_size) (
block_id,
(inode_id % inodes_per_block) as usize * inode_size,
)
} }
pub fn get_data_block_id(&self, data_block_id: u32) -> u32 { pub fn get_data_block_id(&self, data_block_id: u32) -> u32 {
@ -146,18 +132,16 @@ impl EasyFileSystem {
} }
pub fn dealloc_data(&mut self, block_id: u32) { pub fn dealloc_data(&mut self, block_id: u32) {
get_block_cache( get_block_cache(block_id as usize, Arc::clone(&self.block_device))
block_id as usize, .lock()
Arc::clone(&self.block_device) .modify(0, |data_block: &mut DataBlock| {
) data_block.iter_mut().for_each(|p| {
.lock() *p = 0;
.modify(0, |data_block: &mut DataBlock| { })
data_block.iter_mut().for_each(|p| { *p = 0; }) });
});
self.data_bitmap.dealloc( self.data_bitmap.dealloc(
&self.block_device, &self.block_device,
(block_id - self.data_area_start_block) as usize (block_id - self.data_area_start_block) as usize,
) )
} }
} }

View file

@ -1,11 +1,7 @@
use core::fmt::{Debug, Formatter, Result}; use super::{get_block_cache, BlockDevice, BLOCK_SZ};
use super::{
BLOCK_SZ,
BlockDevice,
get_block_cache,
};
use alloc::sync::Arc; use alloc::sync::Arc;
use alloc::vec::Vec; use alloc::vec::Vec;
use core::fmt::{Debug, Formatter, Result};
const EFS_MAGIC: u32 = 0x3b800001; const EFS_MAGIC: u32 = 0x3b800001;
const INODE_DIRECT_COUNT: usize = 28; const INODE_DIRECT_COUNT: usize = 28;
@ -115,7 +111,8 @@ impl DiskInode {
if data_blocks > INDIRECT1_BOUND { if data_blocks > INDIRECT1_BOUND {
total += 1; total += 1;
// sub indirect1 // sub indirect1
total += (data_blocks - INDIRECT1_BOUND + INODE_INDIRECT1_COUNT - 1) / INODE_INDIRECT1_COUNT; total +=
(data_blocks - INDIRECT1_BOUND + INODE_INDIRECT1_COUNT - 1) / INODE_INDIRECT1_COUNT;
} }
total as u32 total as u32
} }
@ -135,22 +132,16 @@ impl DiskInode {
}) })
} else { } else {
let last = inner_id - INDIRECT1_BOUND; let last = inner_id - INDIRECT1_BOUND;
let indirect1 = get_block_cache( let indirect1 = get_block_cache(self.indirect2 as usize, Arc::clone(block_device))
self.indirect2 as usize, .lock()
Arc::clone(block_device) .read(0, |indirect2: &IndirectBlock| {
) indirect2[last / INODE_INDIRECT1_COUNT]
.lock() });
.read(0, |indirect2: &IndirectBlock| { get_block_cache(indirect1 as usize, Arc::clone(block_device))
indirect2[last / INODE_INDIRECT1_COUNT] .lock()
}); .read(0, |indirect1: &IndirectBlock| {
get_block_cache( indirect1[last % INODE_INDIRECT1_COUNT]
indirect1 as usize, })
Arc::clone(block_device)
)
.lock()
.read(0, |indirect1: &IndirectBlock| {
indirect1[last % INODE_INDIRECT1_COUNT]
})
} }
} }
pub fn increase_size( pub fn increase_size(
@ -169,7 +160,7 @@ impl DiskInode {
current_blocks += 1; current_blocks += 1;
} }
// alloc indirect1 // alloc indirect1
if total_blocks > INODE_DIRECT_COUNT as u32{ if total_blocks > INODE_DIRECT_COUNT as u32 {
if current_blocks == INODE_DIRECT_COUNT as u32 { if current_blocks == INODE_DIRECT_COUNT as u32 {
self.indirect1 = new_blocks.next().unwrap(); self.indirect1 = new_blocks.next().unwrap();
} }
@ -179,17 +170,14 @@ impl DiskInode {
return; return;
} }
// fill indirect1 // fill indirect1
get_block_cache( get_block_cache(self.indirect1 as usize, Arc::clone(block_device))
self.indirect1 as usize, .lock()
Arc::clone(block_device) .modify(0, |indirect1: &mut IndirectBlock| {
) while current_blocks < total_blocks.min(INODE_INDIRECT1_COUNT as u32) {
.lock() indirect1[current_blocks as usize] = new_blocks.next().unwrap();
.modify(0, |indirect1: &mut IndirectBlock| { current_blocks += 1;
while current_blocks < total_blocks.min(INODE_INDIRECT1_COUNT as u32) { }
indirect1[current_blocks as usize] = new_blocks.next().unwrap(); });
current_blocks += 1;
}
});
// alloc indirect2 // alloc indirect2
if total_blocks > INODE_INDIRECT1_COUNT as u32 { if total_blocks > INODE_INDIRECT1_COUNT as u32 {
if current_blocks == INODE_INDIRECT1_COUNT as u32 { if current_blocks == INODE_INDIRECT1_COUNT as u32 {
@ -206,76 +194,27 @@ impl DiskInode {
let a1 = total_blocks as usize / INODE_INDIRECT1_COUNT; let a1 = total_blocks as usize / INODE_INDIRECT1_COUNT;
let b1 = total_blocks as usize % INODE_INDIRECT1_COUNT; let b1 = total_blocks as usize % INODE_INDIRECT1_COUNT;
// alloc low-level indirect1 // alloc low-level indirect1
get_block_cache( get_block_cache(self.indirect2 as usize, Arc::clone(block_device))
self.indirect2 as usize, .lock()
Arc::clone(block_device) .modify(0, |indirect2: &mut IndirectBlock| {
) while (a0 < a1) || (a0 == a1 && b0 < b1) {
.lock() if b0 == 0 {
.modify(0, |indirect2: &mut IndirectBlock| { indirect2[a0] = new_blocks.next().unwrap();
while (a0 < a1) || (a0 == a1 && b0 < b1) { }
if b0 == 0 { // fill current
indirect2[a0] = new_blocks.next().unwrap(); get_block_cache(indirect2[a0] as usize, Arc::clone(block_device))
.lock()
.modify(0, |indirect1: &mut IndirectBlock| {
indirect1[b0] = new_blocks.next().unwrap();
});
// move to next
b0 += 1;
if b0 == INODE_INDIRECT1_COUNT {
b0 = 0;
a0 += 1;
}
} }
// fill current });
get_block_cache(
indirect2[a0] as usize,
Arc::clone(block_device)
)
.lock()
.modify(0, |indirect1: &mut IndirectBlock| {
indirect1[b0] = new_blocks.next().unwrap();
});
// move to next
b0 += 1;
if b0 == INODE_INDIRECT1_COUNT {
b0 = 0;
a0 += 1;
}
}
});
// alloc indirect2
if total_blocks > INODE_INDIRECT1_COUNT as u32 {
if current_blocks == INODE_INDIRECT1_COUNT as u32 {
self.indirect2 = new_blocks.next().unwrap();
}
current_blocks -= INODE_INDIRECT1_COUNT as u32;
total_blocks -= INODE_INDIRECT1_COUNT as u32;
} else {
return;
}
// fill indirect2 from (a0, b0) -> (a1, b1)
let mut a0 = current_blocks as usize / INODE_INDIRECT1_COUNT;
let mut b0 = current_blocks as usize % INODE_INDIRECT1_COUNT;
let a1 = total_blocks as usize / INODE_INDIRECT1_COUNT;
let b1 = total_blocks as usize % INODE_INDIRECT1_COUNT;
// alloc low-level indirect1
get_block_cache(
self.indirect2 as usize,
Arc::clone(block_device)
)
.lock()
.modify(0, |indirect2: &mut IndirectBlock| {
while (a0 < a1) || (a0 == a1 && b0 < b1) {
if b0 == 0 {
indirect2[a0] = new_blocks.next().unwrap();
}
// fill current
get_block_cache(
indirect2[a0] as usize,
Arc::clone(block_device)
)
.lock()
.modify(0, |indirect1: &mut IndirectBlock| {
indirect1[b0] = new_blocks.next().unwrap();
});
// move to next
b0 += 1;
if b0 == INODE_INDIRECT1_COUNT {
b0 = 0;
a0 += 1;
}
}
});
} }
/// Clear size to zero and return blocks that should be deallocated. /// Clear size to zero and return blocks that should be deallocated.
@ -301,18 +240,15 @@ impl DiskInode {
return v; return v;
} }
// indirect1 // indirect1
get_block_cache( get_block_cache(self.indirect1 as usize, Arc::clone(block_device))
self.indirect1 as usize, .lock()
Arc::clone(block_device), .modify(0, |indirect1: &mut IndirectBlock| {
) while current_blocks < data_blocks.min(INODE_INDIRECT1_COUNT) {
.lock() v.push(indirect1[current_blocks]);
.modify(0, |indirect1: &mut IndirectBlock| { //indirect1[current_blocks] = 0;
while current_blocks < data_blocks.min(INODE_INDIRECT1_COUNT) { current_blocks += 1;
v.push(indirect1[current_blocks]); }
//indirect1[current_blocks] = 0; });
current_blocks += 1;
}
});
self.indirect1 = 0; self.indirect1 = 0;
// indirect2 block // indirect2 block
if data_blocks > INODE_INDIRECT1_COUNT { if data_blocks > INODE_INDIRECT1_COUNT {
@ -325,45 +261,33 @@ impl DiskInode {
assert!(data_blocks <= INODE_INDIRECT2_COUNT); assert!(data_blocks <= INODE_INDIRECT2_COUNT);
let a1 = data_blocks / INODE_INDIRECT1_COUNT; let a1 = data_blocks / INODE_INDIRECT1_COUNT;
let b1 = data_blocks % INODE_INDIRECT1_COUNT; let b1 = data_blocks % INODE_INDIRECT1_COUNT;
get_block_cache( get_block_cache(self.indirect2 as usize, Arc::clone(block_device))
self.indirect2 as usize, .lock()
Arc::clone(block_device), .modify(0, |indirect2: &mut IndirectBlock| {
) // full indirect1 blocks
.lock() for entry in indirect2.iter_mut().take(a1) {
.modify(0, |indirect2: &mut IndirectBlock| { v.push(*entry);
// full indirect1 blocks get_block_cache(*entry as usize, Arc::clone(block_device))
for i in 0..a1 { .lock()
v.push(indirect2[i]); .modify(0, |indirect1: &mut IndirectBlock| {
get_block_cache( for entry in indirect1.iter() {
indirect2[i] as usize, v.push(*entry);
Arc::clone(block_device), }
) });
.lock() }
.modify(0, |indirect1: &mut IndirectBlock| { // last indirect1 block
for j in 0..INODE_INDIRECT1_COUNT { if b1 > 0 {
v.push(indirect1[j]); v.push(indirect2[a1]);
//indirect1[j] = 0; get_block_cache(indirect2[a1] as usize, Arc::clone(block_device))
} .lock()
}); .modify(0, |indirect1: &mut IndirectBlock| {
//indirect2[i] = 0; for entry in indirect1.iter().take(b1) {
} v.push(*entry);
// last indirect1 block }
if b1 > 0 { });
v.push(indirect2[a1]); //indirect2[a1] = 0;
get_block_cache( }
indirect2[a1] as usize, });
Arc::clone(block_device),
)
.lock()
.modify(0, |indirect1: &mut IndirectBlock| {
for j in 0..b1 {
v.push(indirect1[j]);
//indirect1[j] = 0;
}
});
//indirect2[a1] = 0;
}
});
self.indirect2 = 0; self.indirect2 = 0;
v v
} }
@ -398,7 +322,9 @@ impl DiskInode {
}); });
read_size += block_read_size; read_size += block_read_size;
// move to next block // move to next block
if end_current_block == end { break; } if end_current_block == end {
break;
}
start_block += 1; start_block += 1;
start = end_current_block; start = end_current_block;
} }
@ -424,7 +350,7 @@ impl DiskInode {
let block_write_size = end_current_block - start; let block_write_size = end_current_block - start;
get_block_cache( get_block_cache(
self.get_block_id(start_block as u32, block_device) as usize, self.get_block_id(start_block as u32, block_device) as usize,
Arc::clone(block_device) Arc::clone(block_device),
) )
.lock() .lock()
.modify(0, |data_block: &mut DataBlock| { .modify(0, |data_block: &mut DataBlock| {
@ -434,7 +360,9 @@ impl DiskInode {
}); });
write_size += block_write_size; write_size += block_write_size;
// move to next block // move to next block
if end_current_block == end { break; } if end_current_block == end {
break;
}
start_block += 1; start_block += 1;
start = end_current_block; start = end_current_block;
} }
@ -466,20 +394,10 @@ impl DirEntry {
} }
} }
pub fn as_bytes(&self) -> &[u8] { pub fn as_bytes(&self) -> &[u8] {
unsafe { unsafe { core::slice::from_raw_parts(self as *const _ as usize as *const u8, DIRENT_SZ) }
core::slice::from_raw_parts(
self as *const _ as usize as *const u8,
DIRENT_SZ,
)
}
} }
pub fn as_bytes_mut(&mut self) -> &mut [u8] { pub fn as_bytes_mut(&mut self) -> &mut [u8] {
unsafe { unsafe { core::slice::from_raw_parts_mut(self as *mut _ as usize as *mut u8, DIRENT_SZ) }
core::slice::from_raw_parts_mut(
self as *mut _ as usize as *mut u8,
DIRENT_SZ,
)
}
} }
pub fn name(&self) -> &str { pub fn name(&self) -> &str {
let len = (0usize..).find(|i| self.name[*i] == 0).unwrap(); let len = (0usize..).find(|i| self.name[*i] == 0).unwrap();

View file

@ -2,17 +2,17 @@
extern crate alloc; extern crate alloc;
mod block_dev;
mod layout;
mod efs;
mod bitmap; mod bitmap;
mod vfs;
mod block_cache; mod block_cache;
mod block_dev;
mod efs;
mod layout;
mod vfs;
pub const BLOCK_SZ: usize = 512; pub const BLOCK_SZ: usize = 512;
use bitmap::Bitmap;
use block_cache::{block_cache_sync_all, get_block_cache};
pub use block_dev::BlockDevice; pub use block_dev::BlockDevice;
pub use efs::EasyFileSystem; pub use efs::EasyFileSystem;
pub use vfs::Inode;
use layout::*; use layout::*;
use bitmap::Bitmap; pub use vfs::Inode;
use block_cache::{get_block_cache, block_cache_sync_all};

View file

@ -1,15 +1,9 @@
use super::{ use super::{
BlockDevice, block_cache_sync_all, get_block_cache, BlockDevice, DirEntry, DiskInode, DiskInodeType,
DiskInode, EasyFileSystem, DIRENT_SZ,
DiskInodeType,
DirEntry,
EasyFileSystem,
DIRENT_SZ,
get_block_cache,
block_cache_sync_all,
}; };
use alloc::sync::Arc;
use alloc::string::String; use alloc::string::String;
use alloc::sync::Arc;
use alloc::vec::Vec; use alloc::vec::Vec;
use spin::{Mutex, MutexGuard}; use spin::{Mutex, MutexGuard};
@ -37,35 +31,25 @@ impl Inode {
} }
fn read_disk_inode<V>(&self, f: impl FnOnce(&DiskInode) -> V) -> V { fn read_disk_inode<V>(&self, f: impl FnOnce(&DiskInode) -> V) -> V {
get_block_cache( get_block_cache(self.block_id, Arc::clone(&self.block_device))
self.block_id, .lock()
Arc::clone(&self.block_device) .read(self.block_offset, f)
).lock().read(self.block_offset, f)
} }
fn modify_disk_inode<V>(&self, f: impl FnOnce(&mut DiskInode) -> V) -> V { fn modify_disk_inode<V>(&self, f: impl FnOnce(&mut DiskInode) -> V) -> V {
get_block_cache( get_block_cache(self.block_id, Arc::clone(&self.block_device))
self.block_id, .lock()
Arc::clone(&self.block_device) .modify(self.block_offset, f)
).lock().modify(self.block_offset, f)
} }
fn find_inode_id( fn find_inode_id(&self, name: &str, disk_inode: &DiskInode) -> Option<u32> {
&self,
name: &str,
disk_inode: &DiskInode,
) -> Option<u32> {
// assert it is a directory // assert it is a directory
assert!(disk_inode.is_dir()); assert!(disk_inode.is_dir());
let file_count = (disk_inode.size as usize) / DIRENT_SZ; let file_count = (disk_inode.size as usize) / DIRENT_SZ;
let mut dirent = DirEntry::empty(); let mut dirent = DirEntry::empty();
for i in 0..file_count { for i in 0..file_count {
assert_eq!( assert_eq!(
disk_inode.read_at( disk_inode.read_at(DIRENT_SZ * i, dirent.as_bytes_mut(), &self.block_device,),
DIRENT_SZ * i,
dirent.as_bytes_mut(),
&self.block_device,
),
DIRENT_SZ, DIRENT_SZ,
); );
if dirent.name() == name { if dirent.name() == name {
@ -78,8 +62,7 @@ impl Inode {
pub fn find(&self, name: &str) -> Option<Arc<Inode>> { pub fn find(&self, name: &str) -> Option<Arc<Inode>> {
let fs = self.fs.lock(); let fs = self.fs.lock();
self.read_disk_inode(|disk_inode| { self.read_disk_inode(|disk_inode| {
self.find_inode_id(name, disk_inode) self.find_inode_id(name, disk_inode).map(|inode_id| {
.map(|inode_id| {
let (block_id, block_offset) = fs.get_disk_inode_pos(inode_id); let (block_id, block_offset) = fs.get_disk_inode_pos(inode_id);
Arc::new(Self::new( Arc::new(Self::new(
block_id, block_id,
@ -110,26 +93,25 @@ impl Inode {
pub fn create(&self, name: &str) -> Option<Arc<Inode>> { pub fn create(&self, name: &str) -> Option<Arc<Inode>> {
let mut fs = self.fs.lock(); let mut fs = self.fs.lock();
if self.modify_disk_inode(|root_inode| { let op = |root_inode: &mut DiskInode| {
// assert it is a directory // assert it is a directory
assert!(root_inode.is_dir()); assert!(root_inode.is_dir());
// has the file been created? // has the file been created?
self.find_inode_id(name, root_inode) self.find_inode_id(name, root_inode)
}).is_some() { };
if self.modify_disk_inode(op).is_some() {
return None; return None;
} }
// create a new file // create a new file
// alloc a inode with an indirect block // alloc a inode with an indirect block
let new_inode_id = fs.alloc_inode(); let new_inode_id = fs.alloc_inode();
// initialize inode // initialize inode
let (new_inode_block_id, new_inode_block_offset) let (new_inode_block_id, new_inode_block_offset) = fs.get_disk_inode_pos(new_inode_id);
= fs.get_disk_inode_pos(new_inode_id); get_block_cache(new_inode_block_id as usize, Arc::clone(&self.block_device))
get_block_cache( .lock()
new_inode_block_id as usize, .modify(new_inode_block_offset, |new_inode: &mut DiskInode| {
Arc::clone(&self.block_device) new_inode.initialize(DiskInodeType::File);
).lock().modify(new_inode_block_offset, |new_inode: &mut DiskInode| { });
new_inode.initialize(DiskInodeType::File);
});
self.modify_disk_inode(|root_inode| { self.modify_disk_inode(|root_inode| {
// append file in the dirent // append file in the dirent
let file_count = (root_inode.size as usize) / DIRENT_SZ; let file_count = (root_inode.size as usize) / DIRENT_SZ;
@ -165,11 +147,7 @@ impl Inode {
for i in 0..file_count { for i in 0..file_count {
let mut dirent = DirEntry::empty(); let mut dirent = DirEntry::empty();
assert_eq!( assert_eq!(
disk_inode.read_at( disk_inode.read_at(i * DIRENT_SZ, dirent.as_bytes_mut(), &self.block_device,),
i * DIRENT_SZ,
dirent.as_bytes_mut(),
&self.block_device,
),
DIRENT_SZ, DIRENT_SZ,
); );
v.push(String::from(dirent.name())); v.push(String::from(dirent.name()));
@ -180,9 +158,7 @@ impl Inode {
pub fn read_at(&self, offset: usize, buf: &mut [u8]) -> usize { pub fn read_at(&self, offset: usize, buf: &mut [u8]) -> usize {
let _fs = self.fs.lock(); let _fs = self.fs.lock();
self.read_disk_inode(|disk_inode| { self.read_disk_inode(|disk_inode| disk_inode.read_at(offset, buf, &self.block_device))
disk_inode.read_at(offset, buf, &self.block_device)
})
} }
pub fn write_at(&self, offset: usize, buf: &[u8]) -> usize { pub fn write_at(&self, offset: usize, buf: &[u8]) -> usize {

View file

@ -36,9 +36,9 @@ build: env switch-check $(KERNEL_BIN) fs-img
switch-check: switch-check:
ifeq ($(BOARD), qemu) ifeq ($(BOARD), qemu)
(which last-qemu) || (rm last-k210 -f && touch last-qemu && make clean) (which last-qemu) || (rm -f last-k210 && touch last-qemu && make clean)
else ifeq ($(BOARD), k210) else ifeq ($(BOARD), k210)
(which last-k210) || (rm last-qemu -f && touch last-k210 && make clean) (which last-k210) || (rm -f last-qemu && touch last-k210 && make clean)
endif endif
env: env:
@ -57,7 +57,7 @@ $(KERNEL_BIN): kernel
fs-img: $(APPS) fs-img: $(APPS)
@cd ../user && make build @cd ../user && make build
@rm $(FS_IMG) -f @rm -f $(FS_IMG)
@cd ../easy-fs-fuse && cargo run --release -- -s ../user/src/bin/ -t ../user/target/riscv64gc-unknown-none-elf/release/ @cd ../easy-fs-fuse && cargo run --release -- -s ../user/src/bin/ -t ../user/target/riscv64gc-unknown-none-elf/release/
$(APPS): $(APPS):

23
os/src/boards/k210.rs Normal file
View file

@ -0,0 +1,23 @@
pub const CLOCK_FREQ: usize = 403000000 / 62;
pub const MMIO: &[(usize, usize)] = &[
// we don't need clint in S priv when running
// we only need claim/complete for target0 after initializing
(0x0C00_0000, 0x3000), /* PLIC */
(0x0C20_0000, 0x1000), /* PLIC */
(0x3800_0000, 0x1000), /* UARTHS */
(0x3800_1000, 0x1000), /* GPIOHS */
(0x5020_0000, 0x1000), /* GPIO */
(0x5024_0000, 0x1000), /* SPI_SLAVE */
(0x502B_0000, 0x1000), /* FPIOA */
(0x502D_0000, 0x1000), /* TIMER0 */
(0x502E_0000, 0x1000), /* TIMER1 */
(0x502F_0000, 0x1000), /* TIMER2 */
(0x5044_0000, 0x1000), /* SYSCTL */
(0x5200_0000, 0x1000), /* SPI0 */
(0x5300_0000, 0x1000), /* SPI1 */
(0x5400_0000, 0x1000), /* SPI2 */
];
pub type BlockDeviceImpl = crate::drivers::block::SDCardWrapper;

6
os/src/boards/qemu.rs Normal file
View file

@ -0,0 +1,6 @@
pub const CLOCK_FREQ: usize = 12500000;
pub const MMIO: &[(usize, usize)] = &[(0x10001000, 0x1000)];
pub type BlockDeviceImpl = crate::drivers::block::VirtIOBlock;

View file

@ -10,33 +10,5 @@ pub const PAGE_SIZE_BITS: usize = 0xc;
pub const TRAMPOLINE: usize = usize::MAX - PAGE_SIZE + 1; pub const TRAMPOLINE: usize = usize::MAX - PAGE_SIZE + 1;
pub const TRAP_CONTEXT_BASE: usize = TRAMPOLINE - PAGE_SIZE; pub const TRAP_CONTEXT_BASE: usize = TRAMPOLINE - PAGE_SIZE;
#[cfg(feature = "board_k210")] pub use crate::board::{CLOCK_FREQ, MMIO};
pub const CLOCK_FREQ: usize = 403000000 / 62;
#[cfg(feature = "board_qemu")]
pub const CLOCK_FREQ: usize = 12500000;
#[cfg(feature = "board_qemu")]
pub const MMIO: &[(usize, usize)] = &[
(0x10001000, 0x1000),
];
#[cfg(feature = "board_k210")]
pub const MMIO: &[(usize, usize)] = &[
// we don't need clint in S priv when running
// we only need claim/complete for target0 after initializing
(0x0C00_0000, 0x3000), /* PLIC */
(0x0C20_0000, 0x1000), /* PLIC */
(0x3800_0000, 0x1000), /* UARTHS */
(0x3800_1000, 0x1000), /* GPIOHS */
(0x5020_0000, 0x1000), /* GPIO */
(0x5024_0000, 0x1000), /* SPI_SLAVE */
(0x502B_0000, 0x1000), /* FPIOA */
(0x502D_0000, 0x1000), /* TIMER0 */
(0x502E_0000, 0x1000), /* TIMER1 */
(0x502F_0000, 0x1000), /* TIMER2 */
(0x5044_0000, 0x1000), /* SYSCTL */
(0x5200_0000, 0x1000), /* SPI0 */
(0x5300_0000, 0x1000), /* SPI1 */
(0x5400_0000, 0x1000), /* SPI2 */
];

View file

@ -1,5 +1,5 @@
use core::fmt::{self, Write};
use crate::sbi::console_putchar; use crate::sbi::console_putchar;
use core::fmt::{self, Write};
struct Stdout; struct Stdout;
@ -29,5 +29,3 @@ macro_rules! println {
$crate::console::print(format_args!(concat!($fmt, "\n") $(, $($arg)+)?)) $crate::console::print(format_args!(concat!($fmt, "\n") $(, $($arg)+)?))
} }
} }

View file

@ -1,15 +1,13 @@
mod virtio_blk;
mod sdcard; mod sdcard;
mod virtio_blk;
pub use virtio_blk::VirtIOBlock;
pub use sdcard::SDCardWrapper;
use lazy_static::*;
use alloc::sync::Arc; use alloc::sync::Arc;
use easy_fs::BlockDevice; use easy_fs::BlockDevice;
use lazy_static::*;
#[cfg(feature = "board_qemu")] use crate::board::BlockDeviceImpl;
type BlockDeviceImpl = virtio_blk::VirtIOBlock;
#[cfg(feature = "board_k210")]
type BlockDeviceImpl = sdcard::SDCardWrapper;
lazy_static! { lazy_static! {
pub static ref BLOCK_DEVICE: Arc<dyn BlockDevice> = Arc::new(BlockDeviceImpl::new()); pub static ref BLOCK_DEVICE: Arc<dyn BlockDevice> = Arc::new(BlockDeviceImpl::new());
@ -21,7 +19,9 @@ pub fn block_device_test() {
let mut write_buffer = [0u8; 512]; let mut write_buffer = [0u8; 512];
let mut read_buffer = [0u8; 512]; let mut read_buffer = [0u8; 512];
for i in 0..512 { for i in 0..512 {
for byte in write_buffer.iter_mut() { *byte = i as u8; } for byte in write_buffer.iter_mut() {
*byte = i as u8;
}
block_device.write_block(i as usize, &write_buffer); block_device.write_block(i as usize, &write_buffer);
block_device.read_block(i as usize, &mut read_buffer); block_device.read_block(i as usize, &mut read_buffer);
assert_eq!(write_buffer, read_buffer); assert_eq!(write_buffer, read_buffer);

View file

@ -2,21 +2,21 @@
#![allow(non_camel_case_types)] #![allow(non_camel_case_types)]
#![allow(unused)] #![allow(unused)]
use k210_pac::{Peripherals, SPI0}; use super::BlockDevice;
use crate::sync::UPSafeCell;
use core::convert::TryInto;
use k210_hal::prelude::*; use k210_hal::prelude::*;
use k210_pac::{Peripherals, SPI0};
use k210_soc::{ use k210_soc::{
fpioa::{self, io},
//dmac::{dma_channel, DMAC, DMACExt}, //dmac::{dma_channel, DMAC, DMACExt},
gpio, gpio,
gpiohs, gpiohs,
spi::{aitm, frame_format, tmod, work_mode, SPI, SPIExt, SPIImpl},
fpioa::{self, io},
sysctl,
sleep::usleep, sleep::usleep,
spi::{aitm, frame_format, tmod, work_mode, SPIExt, SPIImpl, SPI},
sysctl,
}; };
use crate::sync::UPSafeCell;
use lazy_static::*; use lazy_static::*;
use super::BlockDevice;
use core::convert::TryInto;
pub struct SDCard<SPI> { pub struct SDCard<SPI> {
spi: SPI, spi: SPI,
@ -160,7 +160,11 @@ pub struct SDCardInfo {
} }
impl</*'a,*/ X: SPI> SDCard</*'a,*/ X> { impl</*'a,*/ X: SPI> SDCard</*'a,*/ X> {
pub fn new(spi: X, spi_cs: u32, cs_gpionum: u8/*, dmac: &'a DMAC, channel: dma_channel*/) -> Self { pub fn new(
spi: X,
spi_cs: u32,
cs_gpionum: u8, /*, dmac: &'a DMAC, channel: dma_channel*/
) -> Self {
Self { Self {
spi, spi,
spi_cs, spi_cs,
@ -310,7 +314,7 @@ impl</*'a,*/ X: SPI> SDCard</*'a,*/ X> {
timeout -= 1; timeout -= 1;
} }
/* After time out */ /* After time out */
return 0xFF; 0xFF
} }
/* /*
@ -337,7 +341,7 @@ impl</*'a,*/ X: SPI> SDCard</*'a,*/ X> {
self.read_data(response); self.read_data(response);
} }
/* Return response */ /* Return response */
return 0; 0
} }
/* /*
@ -367,7 +371,7 @@ impl</*'a,*/ X: SPI> SDCard</*'a,*/ X> {
self.read_data(&mut csd_tab); self.read_data(&mut csd_tab);
self.end_cmd(); self.end_cmd();
/* see also: https://cdn-shop.adafruit.com/datasheets/TS16GUSDHC6.pdf */ /* see also: https://cdn-shop.adafruit.com/datasheets/TS16GUSDHC6.pdf */
return Ok(SDCardCSD { Ok(SDCardCSD {
/* Byte 0 */ /* Byte 0 */
CSDStruct: (csd_tab[0] & 0xC0) >> 6, CSDStruct: (csd_tab[0] & 0xC0) >> 6,
SysSpecVersion: (csd_tab[0] & 0x3C) >> 2, SysSpecVersion: (csd_tab[0] & 0x3C) >> 2,
@ -419,8 +423,8 @@ impl</*'a,*/ X: SPI> SDCard</*'a,*/ X> {
/* Byte 15 */ /* Byte 15 */
CSD_CRC: (csd_tab[15] & 0xFE) >> 1, CSD_CRC: (csd_tab[15] & 0xFE) >> 1,
Reserved4: 1, Reserved4: 1,
/* Return the response */ /* Return the reponse */
}); })
} }
/* /*
@ -449,7 +453,7 @@ impl</*'a,*/ X: SPI> SDCard</*'a,*/ X> {
/* Get CRC bytes (not really needed by us, but required by SD) */ /* Get CRC bytes (not really needed by us, but required by SD) */
self.read_data(&mut cid_tab); self.read_data(&mut cid_tab);
self.end_cmd(); self.end_cmd();
return Ok(SDCardCID { Ok(SDCardCID {
/* Byte 0 */ /* Byte 0 */
ManufacturerID: cid_tab[0], ManufacturerID: cid_tab[0],
/* Byte 1, 2 */ /* Byte 1, 2 */
@ -474,7 +478,7 @@ impl</*'a,*/ X: SPI> SDCard</*'a,*/ X> {
/* Byte 15 */ /* Byte 15 */
CID_CRC: (cid_tab[15] & 0xFE) >> 1, CID_CRC: (cid_tab[15] & 0xFE) >> 1,
Reserved2: 1, Reserved2: 1,
}); })
} }
/* /*
@ -606,7 +610,7 @@ impl</*'a,*/ X: SPI> SDCard</*'a,*/ X> {
} }
let mut error = false; let mut error = false;
//let mut dma_chunk = [0u32; SEC_LEN]; //let mut dma_chunk = [0u32; SEC_LEN];
let mut tmp_chunk= [0u8; SEC_LEN]; let mut tmp_chunk = [0u8; SEC_LEN];
for chunk in data_buf.chunks_mut(SEC_LEN) { for chunk in data_buf.chunks_mut(SEC_LEN) {
if self.get_response() != SD_START_DATA_SINGLE_BLOCK_READ { if self.get_response() != SD_START_DATA_SINGLE_BLOCK_READ {
error = true; error = true;
@ -616,7 +620,7 @@ impl</*'a,*/ X: SPI> SDCard</*'a,*/ X> {
//self.read_data_dma(&mut dma_chunk); //self.read_data_dma(&mut dma_chunk);
self.read_data(&mut tmp_chunk); self.read_data(&mut tmp_chunk);
/* Place the data received as u32 units from DMA into the u8 target buffer */ /* Place the data received as u32 units from DMA into the u8 target buffer */
for (a, b) in chunk.iter_mut().zip(/*dma_chunk*/tmp_chunk.iter()) { for (a, b) in chunk.iter_mut().zip(/*dma_chunk*/ tmp_chunk.iter()) {
//*a = (b & 0xff) as u8; //*a = (b & 0xff) as u8;
*a = *b; *a = *b;
} }
@ -675,12 +679,12 @@ impl</*'a,*/ X: SPI> SDCard</*'a,*/ X> {
/* Send the data token to signify the start of the data */ /* Send the data token to signify the start of the data */
self.write_data(&frame); self.write_data(&frame);
/* Write the block data to SD : write count data by block */ /* Write the block data to SD : write count data by block */
for (a, &b) in /*dma_chunk*/tmp_chunk.iter_mut().zip(chunk.iter()) { for (a, &b) in /*dma_chunk*/ tmp_chunk.iter_mut().zip(chunk.iter()) {
//*a = b.into(); //*a = b.into();
*a = b; *a = b;
} }
//self.write_data_dma(&mut dma_chunk); //self.write_data_dma(&mut dma_chunk);
self.write_data(&mut tmp_chunk); self.write_data(&tmp_chunk);
/* Put dummy CRC bytes */ /* Put dummy CRC bytes */
self.write_data(&[0xff, 0xff]); self.write_data(&[0xff, 0xff]);
/* Read data response */ /* Read data response */
@ -711,9 +715,8 @@ fn io_init() {
} }
lazy_static! { lazy_static! {
static ref PERIPHERALS: UPSafeCell<Peripherals> = unsafe { static ref PERIPHERALS: UPSafeCell<Peripherals> =
UPSafeCell::new(Peripherals::take().unwrap()) unsafe { UPSafeCell::new(Peripherals::take().unwrap()) };
};
} }
fn init_sdcard() -> SDCard<SPIImpl<SPI0>> { fn init_sdcard() -> SDCard<SPIImpl<SPI0>> {
@ -747,9 +750,15 @@ impl SDCardWrapper {
impl BlockDevice for SDCardWrapper { impl BlockDevice for SDCardWrapper {
fn read_block(&self, block_id: usize, buf: &mut [u8]) { fn read_block(&self, block_id: usize, buf: &mut [u8]) {
self.0.exclusive_access().read_sector(buf,block_id as u32).unwrap(); self.0
.exclusive_access()
.read_sector(buf, block_id as u32)
.unwrap();
} }
fn write_block(&self, block_id: usize, buf: &[u8]) { fn write_block(&self, block_id: usize, buf: &[u8]) {
self.0.exclusive_access().write_sector(buf,block_id as u32).unwrap(); self.0
.exclusive_access()
.write_sector(buf, block_id as u32)
.unwrap();
} }
} }

View file

@ -1,20 +1,12 @@
use virtio_drivers::{VirtIOBlk, VirtIOHeader};
use crate::mm::{
PhysAddr,
VirtAddr,
frame_alloc,
frame_dealloc,
PhysPageNum,
FrameTracker,
StepByOne,
PageTable,
kernel_token,
};
use super::BlockDevice; use super::BlockDevice;
use crate::mm::{
frame_alloc, frame_dealloc, kernel_token, FrameTracker, PageTable, PhysAddr, PhysPageNum,
StepByOne, VirtAddr,
};
use crate::sync::UPSafeCell; use crate::sync::UPSafeCell;
use alloc::vec::Vec; use alloc::vec::Vec;
use lazy_static::*; use lazy_static::*;
use virtio_drivers::{VirtIOBlk, VirtIOHeader};
#[allow(unused)] #[allow(unused)]
const VIRTIO0: usize = 0x10001000; const VIRTIO0: usize = 0x10001000;
@ -22,21 +14,21 @@ const VIRTIO0: usize = 0x10001000;
pub struct VirtIOBlock(UPSafeCell<VirtIOBlk<'static>>); pub struct VirtIOBlock(UPSafeCell<VirtIOBlk<'static>>);
lazy_static! { lazy_static! {
static ref QUEUE_FRAMES: UPSafeCell<Vec<FrameTracker>> = unsafe { static ref QUEUE_FRAMES: UPSafeCell<Vec<FrameTracker>> = unsafe { UPSafeCell::new(Vec::new()) };
UPSafeCell::new(Vec::new())
};
} }
impl BlockDevice for VirtIOBlock { impl BlockDevice for VirtIOBlock {
fn read_block(&self, block_id: usize, buf: &mut [u8]) { fn read_block(&self, block_id: usize, buf: &mut [u8]) {
self.0.exclusive_access() self.0
.read_block(block_id, buf) .exclusive_access()
.expect("Error when reading VirtIOBlk"); .read_block(block_id, buf)
.expect("Error when reading VirtIOBlk");
} }
fn write_block(&self, block_id: usize, buf: &[u8]) { fn write_block(&self, block_id: usize, buf: &[u8]) {
self.0.exclusive_access() self.0
.write_block(block_id, buf) .exclusive_access()
.expect("Error when writing VirtIOBlk"); .write_block(block_id, buf)
.expect("Error when writing VirtIOBlk");
} }
} }
@ -44,9 +36,9 @@ impl VirtIOBlock {
#[allow(unused)] #[allow(unused)]
pub fn new() -> Self { pub fn new() -> Self {
unsafe { unsafe {
Self(UPSafeCell::new(VirtIOBlk::new( Self(UPSafeCell::new(
&mut *(VIRTIO0 as *mut VirtIOHeader) VirtIOBlk::new(&mut *(VIRTIO0 as *mut VirtIOHeader)).unwrap(),
).unwrap())) ))
} }
} }
} }
@ -56,7 +48,9 @@ pub extern "C" fn virtio_dma_alloc(pages: usize) -> PhysAddr {
let mut ppn_base = PhysPageNum(0); let mut ppn_base = PhysPageNum(0);
for i in 0..pages { for i in 0..pages {
let frame = frame_alloc().unwrap(); let frame = frame_alloc().unwrap();
if i == 0 { ppn_base = frame.ppn; } if i == 0 {
ppn_base = frame.ppn;
}
assert_eq!(frame.ppn.0, ppn_base.0 + i); assert_eq!(frame.ppn.0, ppn_base.0 + i);
QUEUE_FRAMES.exclusive_access().push(frame); QUEUE_FRAMES.exclusive_access().push(frame);
} }
@ -80,5 +74,7 @@ pub extern "C" fn virtio_phys_to_virt(paddr: PhysAddr) -> VirtAddr {
#[no_mangle] #[no_mangle]
pub extern "C" fn virtio_virt_to_phys(vaddr: VirtAddr) -> PhysAddr { pub extern "C" fn virtio_virt_to_phys(vaddr: VirtAddr) -> PhysAddr {
PageTable::from_token(kernel_token()).translate_va(vaddr).unwrap() PageTable::from_token(kernel_token())
.translate_va(vaddr)
.unwrap()
} }

View file

@ -1,3 +1,3 @@
mod block; pub mod block;
pub use block::BLOCK_DEVICE; pub use block::BLOCK_DEVICE;

View file

@ -1,15 +1,12 @@
use easy_fs::{ use super::File;
EasyFileSystem,
Inode,
};
use crate::drivers::BLOCK_DEVICE; use crate::drivers::BLOCK_DEVICE;
use crate::mm::UserBuffer;
use crate::sync::UPSafeCell; use crate::sync::UPSafeCell;
use alloc::sync::Arc; use alloc::sync::Arc;
use lazy_static::*;
use bitflags::*;
use alloc::vec::Vec; use alloc::vec::Vec;
use super::File; use bitflags::*;
use crate::mm::UserBuffer; use easy_fs::{EasyFileSystem, Inode};
use lazy_static::*;
pub struct OSInode { pub struct OSInode {
readable: bool, readable: bool,
@ -23,18 +20,11 @@ pub struct OSInodeInner {
} }
impl OSInode { impl OSInode {
pub fn new( pub fn new(readable: bool, writable: bool, inode: Arc<Inode>) -> Self {
readable: bool,
writable: bool,
inode: Arc<Inode>,
) -> Self {
Self { Self {
readable, readable,
writable, writable,
inner: unsafe { UPSafeCell::new(OSInodeInner { inner: unsafe { UPSafeCell::new(OSInodeInner { offset: 0, inode }) },
offset: 0,
inode,
})},
} }
} }
pub fn read_all(&self) -> Vec<u8> { pub fn read_all(&self) -> Vec<u8> {
@ -98,40 +88,30 @@ pub fn open_file(name: &str, flags: OpenFlags) -> Option<Arc<OSInode>> {
if let Some(inode) = ROOT_INODE.find(name) { if let Some(inode) = ROOT_INODE.find(name) {
// clear size // clear size
inode.clear(); inode.clear();
Some(Arc::new(OSInode::new( Some(Arc::new(OSInode::new(readable, writable, inode)))
readable,
writable,
inode,
)))
} else { } else {
// create file // create file
ROOT_INODE.create(name) ROOT_INODE
.map(|inode| { .create(name)
Arc::new(OSInode::new( .map(|inode| Arc::new(OSInode::new(readable, writable, inode)))
readable,
writable,
inode,
))
})
} }
} else { } else {
ROOT_INODE.find(name) ROOT_INODE.find(name).map(|inode| {
.map(|inode| { if flags.contains(OpenFlags::TRUNC) {
if flags.contains(OpenFlags::TRUNC) { inode.clear();
inode.clear(); }
} Arc::new(OSInode::new(readable, writable, inode))
Arc::new(OSInode::new( })
readable,
writable,
inode
))
})
} }
} }
impl File for OSInode { impl File for OSInode {
fn readable(&self) -> bool { self.readable } fn readable(&self) -> bool {
fn writable(&self) -> bool { self.writable } self.readable
}
fn writable(&self) -> bool {
self.writable
}
fn read(&self, mut buf: UserBuffer) -> usize { fn read(&self, mut buf: UserBuffer) -> usize {
let mut inner = self.inner.exclusive_access(); let mut inner = self.inner.exclusive_access();
let mut total_read_size = 0usize; let mut total_read_size = 0usize;

View file

@ -1,16 +1,16 @@
mod inode;
mod pipe; mod pipe;
mod stdio; mod stdio;
mod inode;
use crate::mm::UserBuffer; use crate::mm::UserBuffer;
pub trait File : Send + Sync { pub trait File: Send + Sync {
fn readable(&self) -> bool; fn readable(&self) -> bool;
fn writable(&self) -> bool; fn writable(&self) -> bool;
fn read(&self, buf: UserBuffer) -> usize; fn read(&self, buf: UserBuffer) -> usize;
fn write(&self, buf: UserBuffer) -> usize; fn write(&self, buf: UserBuffer) -> usize;
} }
pub use pipe::{Pipe, make_pipe}; pub use inode::{list_apps, open_file, OSInode, OpenFlags};
pub use pipe::{make_pipe, Pipe};
pub use stdio::{Stdin, Stdout}; pub use stdio::{Stdin, Stdout};
pub use inode::{OSInode, open_file, OpenFlags, list_apps};

View file

@ -1,7 +1,7 @@
use super::File; use super::File;
use alloc::sync::{Arc, Weak};
use crate::sync::UPSafeCell;
use crate::mm::UserBuffer; use crate::mm::UserBuffer;
use crate::sync::UPSafeCell;
use alloc::sync::{Arc, Weak};
use crate::task::suspend_current_and_run_next; use crate::task::suspend_current_and_run_next;
@ -32,9 +32,9 @@ const RING_BUFFER_SIZE: usize = 32;
#[derive(Copy, Clone, PartialEq)] #[derive(Copy, Clone, PartialEq)]
enum RingBufferStatus { enum RingBufferStatus {
FULL, Full,
EMPTY, Empty,
NORMAL, Normal,
} }
pub struct PipeRingBuffer { pub struct PipeRingBuffer {
@ -51,7 +51,7 @@ impl PipeRingBuffer {
arr: [0; RING_BUFFER_SIZE], arr: [0; RING_BUFFER_SIZE],
head: 0, head: 0,
tail: 0, tail: 0,
status: RingBufferStatus::EMPTY, status: RingBufferStatus::Empty,
write_end: None, write_end: None,
} }
} }
@ -59,35 +59,33 @@ impl PipeRingBuffer {
self.write_end = Some(Arc::downgrade(write_end)); self.write_end = Some(Arc::downgrade(write_end));
} }
pub fn write_byte(&mut self, byte: u8) { pub fn write_byte(&mut self, byte: u8) {
self.status = RingBufferStatus::NORMAL; self.status = RingBufferStatus::Normal;
self.arr[self.tail] = byte; self.arr[self.tail] = byte;
self.tail = (self.tail + 1) % RING_BUFFER_SIZE; self.tail = (self.tail + 1) % RING_BUFFER_SIZE;
if self.tail == self.head { if self.tail == self.head {
self.status = RingBufferStatus::FULL; self.status = RingBufferStatus::Full;
} }
} }
pub fn read_byte(&mut self) -> u8 { pub fn read_byte(&mut self) -> u8 {
self.status = RingBufferStatus::NORMAL; self.status = RingBufferStatus::Normal;
let c = self.arr[self.head]; let c = self.arr[self.head];
self.head = (self.head + 1) % RING_BUFFER_SIZE; self.head = (self.head + 1) % RING_BUFFER_SIZE;
if self.head == self.tail { if self.head == self.tail {
self.status = RingBufferStatus::EMPTY; self.status = RingBufferStatus::Empty;
} }
c c
} }
pub fn available_read(&self) -> usize { pub fn available_read(&self) -> usize {
if self.status == RingBufferStatus::EMPTY { if self.status == RingBufferStatus::Empty {
0 0
} else if self.tail > self.head {
self.tail - self.head
} else { } else {
if self.tail > self.head { self.tail + RING_BUFFER_SIZE - self.head
self.tail - self.head
} else {
self.tail + RING_BUFFER_SIZE - self.head
}
} }
} }
pub fn available_write(&self) -> usize { pub fn available_write(&self) -> usize {
if self.status == RingBufferStatus::FULL { if self.status == RingBufferStatus::Full {
0 0
} else { } else {
RING_BUFFER_SIZE - self.available_read() RING_BUFFER_SIZE - self.available_read()
@ -100,24 +98,22 @@ impl PipeRingBuffer {
/// Return (read_end, write_end) /// Return (read_end, write_end)
pub fn make_pipe() -> (Arc<Pipe>, Arc<Pipe>) { pub fn make_pipe() -> (Arc<Pipe>, Arc<Pipe>) {
let buffer = Arc::new(unsafe { let buffer = Arc::new(unsafe { UPSafeCell::new(PipeRingBuffer::new()) });
UPSafeCell::new(PipeRingBuffer::new()) let read_end = Arc::new(Pipe::read_end_with_buffer(buffer.clone()));
}); let write_end = Arc::new(Pipe::write_end_with_buffer(buffer.clone()));
let read_end = Arc::new(
Pipe::read_end_with_buffer(buffer.clone())
);
let write_end = Arc::new(
Pipe::write_end_with_buffer(buffer.clone())
);
buffer.exclusive_access().set_write_end(&write_end); buffer.exclusive_access().set_write_end(&write_end);
(read_end, write_end) (read_end, write_end)
} }
impl File for Pipe { impl File for Pipe {
fn readable(&self) -> bool { self.readable } fn readable(&self) -> bool {
fn writable(&self) -> bool { self.writable } self.readable
}
fn writable(&self) -> bool {
self.writable
}
fn read(&self, buf: UserBuffer) -> usize { fn read(&self, buf: UserBuffer) -> usize {
assert_eq!(self.readable(), true); assert!(self.readable());
let mut buf_iter = buf.into_iter(); let mut buf_iter = buf.into_iter();
let mut read_size = 0usize; let mut read_size = 0usize;
loop { loop {
@ -134,7 +130,9 @@ impl File for Pipe {
// read at most loop_read bytes // read at most loop_read bytes
for _ in 0..loop_read { for _ in 0..loop_read {
if let Some(byte_ref) = buf_iter.next() { if let Some(byte_ref) = buf_iter.next() {
unsafe { *byte_ref = ring_buffer.read_byte(); } unsafe {
*byte_ref = ring_buffer.read_byte();
}
read_size += 1; read_size += 1;
} else { } else {
return read_size; return read_size;
@ -143,7 +141,7 @@ impl File for Pipe {
} }
} }
fn write(&self, buf: UserBuffer) -> usize { fn write(&self, buf: UserBuffer) -> usize {
assert_eq!(self.writable(), true); assert!(self.writable());
let mut buf_iter = buf.into_iter(); let mut buf_iter = buf.into_iter();
let mut write_size = 0usize; let mut write_size = 0usize;
loop { loop {

View file

@ -1,5 +1,5 @@
use super::File; use super::File;
use crate::mm::{UserBuffer}; use crate::mm::UserBuffer;
use crate::sbi::console_getchar; use crate::sbi::console_getchar;
use crate::task::suspend_current_and_run_next; use crate::task::suspend_current_and_run_next;
@ -8,8 +8,12 @@ pub struct Stdin;
pub struct Stdout; pub struct Stdout;
impl File for Stdin { impl File for Stdin {
fn readable(&self) -> bool { true } fn readable(&self) -> bool {
fn writable(&self) -> bool { false } true
}
fn writable(&self) -> bool {
false
}
fn read(&self, mut user_buf: UserBuffer) -> usize { fn read(&self, mut user_buf: UserBuffer) -> usize {
assert_eq!(user_buf.len(), 1); assert_eq!(user_buf.len(), 1);
// busy loop // busy loop
@ -24,7 +28,9 @@ impl File for Stdin {
} }
} }
let ch = c as u8; let ch = c as u8;
unsafe { user_buf.buffers[0].as_mut_ptr().write_volatile(ch); } unsafe {
user_buf.buffers[0].as_mut_ptr().write_volatile(ch);
}
1 1
} }
fn write(&self, _user_buf: UserBuffer) -> usize { fn write(&self, _user_buf: UserBuffer) -> usize {
@ -33,9 +39,13 @@ impl File for Stdin {
} }
impl File for Stdout { impl File for Stdout {
fn readable(&self) -> bool { false } fn readable(&self) -> bool {
fn writable(&self) -> bool { true } false
fn read(&self, _user_buf: UserBuffer) -> usize{ }
fn writable(&self) -> bool {
true
}
fn read(&self, _user_buf: UserBuffer) -> usize {
panic!("Cannot read from stdout!"); panic!("Cannot read from stdout!");
} }
fn write(&self, user_buf: UserBuffer) -> usize { fn write(&self, user_buf: UserBuffer) -> usize {

View file

@ -1,22 +1,23 @@
use core::panic::PanicInfo;
use core::arch::asm;
use crate::sbi::shutdown; use crate::sbi::shutdown;
use crate::task::current_kstack_top; use crate::task::current_kstack_top;
use core::arch::asm;
use core::panic::PanicInfo;
#[panic_handler] #[panic_handler]
fn panic(info: &PanicInfo) -> ! { fn panic(info: &PanicInfo) -> ! {
match info.location() { if let Some(location) = info.location() {
Some(location) => { println!(
println!("[kernel] panicked at '{}', {}:{}:{}", "[kernel] Panicked at {}:{} {}",
info.message().unwrap(), location.file(),
location.file(), location.line(),
location.line(), info.message().unwrap()
location.column() );
); } else {
} println!("[kernel] Panicked: {}", info.message().unwrap());
None => println!("[kernel] panicked at '{}'", info.message().unwrap()) }
unsafe {
backtrace();
} }
unsafe { backtrace(); }
shutdown() shutdown()
} }
@ -26,9 +27,11 @@ unsafe fn backtrace() {
asm!("mv {}, s0", out(reg) fp); asm!("mv {}, s0", out(reg) fp);
println!("---START BACKTRACE---"); println!("---START BACKTRACE---");
for i in 0..10 { for i in 0..10 {
if fp == stop { break; } if fp == stop {
println!("#{}:ra={:#x}", i, *((fp-8) as *const usize)); break;
fp = *((fp-16) as *const usize); }
println!("#{}:ra={:#x}", i, *((fp - 8) as *const usize));
fp = *((fp - 16) as *const usize);
} }
println!("---END BACKTRACE---"); println!("---END BACKTRACE---");
} }

View file

@ -1,62 +0,0 @@
use alloc::vec::Vec;
use lazy_static::*;
pub fn get_num_app() -> usize {
extern "C" { fn _num_app(); }
unsafe { (_num_app as usize as *const usize).read_volatile() }
}
pub fn get_app_data(app_id: usize) -> &'static [u8] {
extern "C" { fn _num_app(); }
let num_app_ptr = _num_app as usize as *const usize;
let num_app = get_num_app();
let app_start = unsafe {
core::slice::from_raw_parts(num_app_ptr.add(1), num_app + 1)
};
assert!(app_id < num_app);
unsafe {
core::slice::from_raw_parts(
app_start[app_id] as *const u8,
app_start[app_id + 1] - app_start[app_id]
)
}
}
lazy_static! {
static ref APP_NAMES: Vec<&'static str> = {
let num_app = get_num_app();
extern "C" { fn _app_names(); }
let mut start = _app_names as usize as *const u8;
let mut v = Vec::new();
unsafe {
for _ in 0..num_app {
let mut end = start;
while end.read_volatile() != '\0' as u8 {
end = end.add(1);
}
let slice = core::slice::from_raw_parts(start, end as usize - start as usize);
let str = core::str::from_utf8(slice).unwrap();
v.push(str);
start = end.add(1);
}
}
v
};
}
#[allow(unused)]
pub fn get_app_data_by_name(name: &str) -> Option<&'static [u8]> {
let num_app = get_num_app();
(0..num_app)
.find(|&i| APP_NAMES[i] == name)
.map(|i| get_app_data(i))
}
pub fn list_apps() {
println!("/**** APPS ****");
for app in APP_NAMES.iter() {
println!("{}", app);
}
println!("**************/");
}

View file

@ -8,21 +8,28 @@ extern crate alloc;
#[macro_use] #[macro_use]
extern crate bitflags; extern crate bitflags;
use core::arch::global_asm; #[cfg(feature = "board_k210")]
#[path = "boards/k210.rs"]
mod board;
#[cfg(not(any(feature = "board_k210")))]
#[path = "boards/qemu.rs"]
mod board;
#[macro_use] #[macro_use]
mod console; mod console;
mod lang_items;
mod sbi;
mod syscall;
mod trap;
mod config; mod config;
mod drivers;
mod fs;
mod lang_items;
mod mm;
mod sbi;
mod sync;
mod syscall;
mod task; mod task;
mod timer; mod timer;
mod sync; mod trap;
mod mm;
mod fs; use core::arch::global_asm;
mod drivers;
global_asm!(include_str!("entry.asm")); global_asm!(include_str!("entry.asm"));
@ -32,10 +39,8 @@ fn clear_bss() {
fn ebss(); fn ebss();
} }
unsafe { unsafe {
core::slice::from_raw_parts_mut( core::slice::from_raw_parts_mut(sbss as usize as *mut u8, ebss as usize - sbss as usize)
sbss as usize as *mut u8, .fill(0);
ebss as usize - sbss as usize,
).fill(0);
} }
} }

View file

@ -1,5 +1,5 @@
use crate::config::{PAGE_SIZE, PAGE_SIZE_BITS};
use super::PageTableEntry; use super::PageTableEntry;
use crate::config::{PAGE_SIZE, PAGE_SIZE_BITS};
use core::fmt::{self, Debug, Formatter}; use core::fmt::{self, Debug, Formatter};
const PA_WIDTH_SV39: usize = 56; const PA_WIDTH_SV39: usize = 56;
@ -52,35 +52,59 @@ impl Debug for PhysPageNum {
/// usize -> T: usize.into() /// usize -> T: usize.into()
impl From<usize> for PhysAddr { impl From<usize> for PhysAddr {
fn from(v: usize) -> Self { Self(v & ( (1 << PA_WIDTH_SV39) - 1 )) } fn from(v: usize) -> Self {
Self(v & ((1 << PA_WIDTH_SV39) - 1))
}
} }
impl From<usize> for PhysPageNum { impl From<usize> for PhysPageNum {
fn from(v: usize) -> Self { Self(v & ( (1 << PPN_WIDTH_SV39) - 1 )) } fn from(v: usize) -> Self {
Self(v & ((1 << PPN_WIDTH_SV39) - 1))
}
} }
impl From<usize> for VirtAddr { impl From<usize> for VirtAddr {
fn from(v: usize) -> Self { Self(v & ( (1 << VA_WIDTH_SV39) - 1 )) } fn from(v: usize) -> Self {
Self(v & ((1 << VA_WIDTH_SV39) - 1))
}
} }
impl From<usize> for VirtPageNum { impl From<usize> for VirtPageNum {
fn from(v: usize) -> Self { Self(v & ( (1 << VPN_WIDTH_SV39) - 1 )) } fn from(v: usize) -> Self {
Self(v & ((1 << VPN_WIDTH_SV39) - 1))
}
} }
impl From<PhysAddr> for usize { impl From<PhysAddr> for usize {
fn from(v: PhysAddr) -> Self { v.0 } fn from(v: PhysAddr) -> Self {
v.0
}
} }
impl From<PhysPageNum> for usize { impl From<PhysPageNum> for usize {
fn from(v: PhysPageNum) -> Self { v.0 } fn from(v: PhysPageNum) -> Self {
v.0
}
} }
impl From<VirtAddr> for usize { impl From<VirtAddr> for usize {
fn from(v: VirtAddr) -> Self { v.0 } fn from(v: VirtAddr) -> Self {
v.0
}
} }
impl From<VirtPageNum> for usize { impl From<VirtPageNum> for usize {
fn from(v: VirtPageNum) -> Self { v.0 } fn from(v: VirtPageNum) -> Self {
v.0
}
} }
impl VirtAddr { impl VirtAddr {
pub fn floor(&self) -> VirtPageNum { VirtPageNum(self.0 / PAGE_SIZE) } pub fn floor(&self) -> VirtPageNum {
pub fn ceil(&self) -> VirtPageNum { VirtPageNum((self.0 - 1 + PAGE_SIZE) / PAGE_SIZE) } VirtPageNum(self.0 / PAGE_SIZE)
pub fn page_offset(&self) -> usize { self.0 & (PAGE_SIZE - 1) } }
pub fn aligned(&self) -> bool { self.page_offset() == 0 } pub fn ceil(&self) -> VirtPageNum {
VirtPageNum((self.0 - 1 + PAGE_SIZE) / PAGE_SIZE)
}
pub fn page_offset(&self) -> usize {
self.0 & (PAGE_SIZE - 1)
}
pub fn aligned(&self) -> bool {
self.page_offset() == 0
}
} }
impl From<VirtAddr> for VirtPageNum { impl From<VirtAddr> for VirtPageNum {
fn from(v: VirtAddr) -> Self { fn from(v: VirtAddr) -> Self {
@ -89,13 +113,23 @@ impl From<VirtAddr> for VirtPageNum {
} }
} }
impl From<VirtPageNum> for VirtAddr { impl From<VirtPageNum> for VirtAddr {
fn from(v: VirtPageNum) -> Self { Self(v.0 << PAGE_SIZE_BITS) } fn from(v: VirtPageNum) -> Self {
Self(v.0 << PAGE_SIZE_BITS)
}
} }
impl PhysAddr { impl PhysAddr {
pub fn floor(&self) -> PhysPageNum { PhysPageNum(self.0 / PAGE_SIZE) } pub fn floor(&self) -> PhysPageNum {
pub fn ceil(&self) -> PhysPageNum { PhysPageNum((self.0 - 1 + PAGE_SIZE) / PAGE_SIZE) } PhysPageNum(self.0 / PAGE_SIZE)
pub fn page_offset(&self) -> usize { self.0 & (PAGE_SIZE - 1) } }
pub fn aligned(&self) -> bool { self.page_offset() == 0 } pub fn ceil(&self) -> PhysPageNum {
PhysPageNum((self.0 - 1 + PAGE_SIZE) / PAGE_SIZE)
}
pub fn page_offset(&self) -> usize {
self.0 & (PAGE_SIZE - 1)
}
pub fn aligned(&self) -> bool {
self.page_offset() == 0
}
} }
impl From<PhysAddr> for PhysPageNum { impl From<PhysAddr> for PhysPageNum {
fn from(v: PhysAddr) -> Self { fn from(v: PhysAddr) -> Self {
@ -104,7 +138,9 @@ impl From<PhysAddr> for PhysPageNum {
} }
} }
impl From<PhysPageNum> for PhysAddr { impl From<PhysPageNum> for PhysAddr {
fn from(v: PhysPageNum) -> Self { Self(v.0 << PAGE_SIZE_BITS) } fn from(v: PhysPageNum) -> Self {
Self(v.0 << PAGE_SIZE_BITS)
}
} }
impl VirtPageNum { impl VirtPageNum {
@ -121,31 +157,23 @@ impl VirtPageNum {
impl PhysAddr { impl PhysAddr {
pub fn get_ref<T>(&self) -> &'static T { pub fn get_ref<T>(&self) -> &'static T {
unsafe { unsafe { (self.0 as *const T).as_ref().unwrap() }
(self.0 as *const T).as_ref().unwrap()
}
} }
pub fn get_mut<T>(&self) -> &'static mut T { pub fn get_mut<T>(&self) -> &'static mut T {
unsafe { unsafe { (self.0 as *mut T).as_mut().unwrap() }
(self.0 as *mut T).as_mut().unwrap()
}
} }
} }
impl PhysPageNum { impl PhysPageNum {
pub fn get_pte_array(&self) -> &'static mut [PageTableEntry] { pub fn get_pte_array(&self) -> &'static mut [PageTableEntry] {
let pa: PhysAddr = self.clone().into(); let pa: PhysAddr = (*self).into();
unsafe { unsafe { core::slice::from_raw_parts_mut(pa.0 as *mut PageTableEntry, 512) }
core::slice::from_raw_parts_mut(pa.0 as *mut PageTableEntry, 512)
}
} }
pub fn get_bytes_array(&self) -> &'static mut [u8] { pub fn get_bytes_array(&self) -> &'static mut [u8] {
let pa: PhysAddr = self.clone().into(); let pa: PhysAddr = (*self).into();
unsafe { unsafe { core::slice::from_raw_parts_mut(pa.0 as *mut u8, 4096) }
core::slice::from_raw_parts_mut(pa.0 as *mut u8, 4096)
}
} }
pub fn get_mut<T>(&self) -> &'static mut T { pub fn get_mut<T>(&self) -> &'static mut T {
let pa: PhysAddr = self.clone().into(); let pa: PhysAddr = (*self).into();
pa.get_mut() pa.get_mut()
} }
} }
@ -165,41 +193,57 @@ impl StepByOne for PhysPageNum {
} }
#[derive(Copy, Clone)] #[derive(Copy, Clone)]
pub struct SimpleRange<T> where pub struct SimpleRange<T>
T: StepByOne + Copy + PartialEq + PartialOrd + Debug, { where
T: StepByOne + Copy + PartialEq + PartialOrd + Debug,
{
l: T, l: T,
r: T, r: T,
} }
impl<T> SimpleRange<T> where impl<T> SimpleRange<T>
T: StepByOne + Copy + PartialEq + PartialOrd + Debug, { where
T: StepByOne + Copy + PartialEq + PartialOrd + Debug,
{
pub fn new(start: T, end: T) -> Self { pub fn new(start: T, end: T) -> Self {
assert!(start <= end, "start {:?} > end {:?}!", start, end); assert!(start <= end, "start {:?} > end {:?}!", start, end);
Self { l: start, r: end } Self { l: start, r: end }
} }
pub fn get_start(&self) -> T { self.l } pub fn get_start(&self) -> T {
pub fn get_end(&self) -> T { self.r } self.l
}
pub fn get_end(&self) -> T {
self.r
}
} }
impl<T> IntoIterator for SimpleRange<T> where impl<T> IntoIterator for SimpleRange<T>
T: StepByOne + Copy + PartialEq + PartialOrd + Debug, { where
T: StepByOne + Copy + PartialEq + PartialOrd + Debug,
{
type Item = T; type Item = T;
type IntoIter = SimpleRangeIterator<T>; type IntoIter = SimpleRangeIterator<T>;
fn into_iter(self) -> Self::IntoIter { fn into_iter(self) -> Self::IntoIter {
SimpleRangeIterator::new(self.l, self.r) SimpleRangeIterator::new(self.l, self.r)
} }
} }
pub struct SimpleRangeIterator<T> where pub struct SimpleRangeIterator<T>
T: StepByOne + Copy + PartialEq + PartialOrd + Debug, { where
T: StepByOne + Copy + PartialEq + PartialOrd + Debug,
{
current: T, current: T,
end: T, end: T,
} }
impl<T> SimpleRangeIterator<T> where impl<T> SimpleRangeIterator<T>
T: StepByOne + Copy + PartialEq + PartialOrd + Debug, { where
T: StepByOne + Copy + PartialEq + PartialOrd + Debug,
{
pub fn new(l: T, r: T) -> Self { pub fn new(l: T, r: T) -> Self {
Self { current: l, end: r, } Self { current: l, end: r }
} }
} }
impl<T> Iterator for SimpleRangeIterator<T> where impl<T> Iterator for SimpleRangeIterator<T>
T: StepByOne + Copy + PartialEq + PartialOrd + Debug, { where
T: StepByOne + Copy + PartialEq + PartialOrd + Debug,
{
type Item = T; type Item = T;
fn next(&mut self) -> Option<Self::Item> { fn next(&mut self) -> Option<Self::Item> {
if self.current == self.end { if self.current == self.end {

View file

@ -1,9 +1,9 @@
use super::{PhysAddr, PhysPageNum}; use super::{PhysAddr, PhysPageNum};
use alloc::vec::Vec;
use crate::sync::UPSafeCell;
use crate::config::MEMORY_END; use crate::config::MEMORY_END;
use lazy_static::*; use crate::sync::UPSafeCell;
use alloc::vec::Vec;
use core::fmt::{self, Debug, Formatter}; use core::fmt::{self, Debug, Formatter};
use lazy_static::*;
pub struct FrameTracker { pub struct FrameTracker {
pub ppn: PhysPageNum, pub ppn: PhysPageNum,
@ -62,22 +62,17 @@ impl FrameAllocator for StackFrameAllocator {
fn alloc(&mut self) -> Option<PhysPageNum> { fn alloc(&mut self) -> Option<PhysPageNum> {
if let Some(ppn) = self.recycled.pop() { if let Some(ppn) = self.recycled.pop() {
Some(ppn.into()) Some(ppn.into())
} else if self.current == self.end {
None
} else { } else {
if self.current == self.end { self.current += 1;
None Some((self.current - 1).into())
} else {
self.current += 1;
Some((self.current - 1).into())
}
} }
} }
fn dealloc(&mut self, ppn: PhysPageNum) { fn dealloc(&mut self, ppn: PhysPageNum) {
let ppn = ppn.0; let ppn = ppn.0;
// validity check // validity check
if ppn >= self.current || self.recycled if ppn >= self.current || self.recycled.iter().any(|&v| v == ppn) {
.iter()
.find(|&v| {*v == ppn})
.is_some() {
panic!("Frame ppn={:#x} has not been allocated!", ppn); panic!("Frame ppn={:#x} has not been allocated!", ppn);
} }
// recycle // recycle
@ -88,31 +83,29 @@ impl FrameAllocator for StackFrameAllocator {
type FrameAllocatorImpl = StackFrameAllocator; type FrameAllocatorImpl = StackFrameAllocator;
lazy_static! { lazy_static! {
pub static ref FRAME_ALLOCATOR: UPSafeCell<FrameAllocatorImpl> = unsafe { pub static ref FRAME_ALLOCATOR: UPSafeCell<FrameAllocatorImpl> =
UPSafeCell::new(FrameAllocatorImpl::new()) unsafe { UPSafeCell::new(FrameAllocatorImpl::new()) };
};
} }
pub fn init_frame_allocator() { pub fn init_frame_allocator() {
extern "C" { extern "C" {
fn ekernel(); fn ekernel();
} }
FRAME_ALLOCATOR FRAME_ALLOCATOR.exclusive_access().init(
.exclusive_access() PhysAddr::from(ekernel as usize).ceil(),
.init(PhysAddr::from(ekernel as usize).ceil(), PhysAddr::from(MEMORY_END).floor()); PhysAddr::from(MEMORY_END).floor(),
);
} }
pub fn frame_alloc() -> Option<FrameTracker> { pub fn frame_alloc() -> Option<FrameTracker> {
FRAME_ALLOCATOR FRAME_ALLOCATOR
.exclusive_access() .exclusive_access()
.alloc() .alloc()
.map(|ppn| FrameTracker::new(ppn)) .map(FrameTracker::new)
} }
pub fn frame_dealloc(ppn: PhysPageNum) { pub fn frame_dealloc(ppn: PhysPageNum) {
FRAME_ALLOCATOR FRAME_ALLOCATOR.exclusive_access().dealloc(ppn);
.exclusive_access()
.dealloc(ppn);
} }
#[allow(unused)] #[allow(unused)]

View file

@ -1,5 +1,5 @@
use buddy_system_allocator::LockedHeap;
use crate::config::KERNEL_HEAP_SIZE; use crate::config::KERNEL_HEAP_SIZE;
use buddy_system_allocator::LockedHeap;
#[global_allocator] #[global_allocator]
static HEAP_ALLOCATOR: LockedHeap = LockedHeap::empty(); static HEAP_ALLOCATOR: LockedHeap = LockedHeap::empty();
@ -36,8 +36,8 @@ pub fn heap_test() {
for i in 0..500 { for i in 0..500 {
v.push(i); v.push(i);
} }
for i in 0..500 { for (i, val) in v.iter().take(500).enumerate() {
assert_eq!(v[i], i); assert_eq!(*val, i);
} }
assert!(bss_range.contains(&(v.as_ptr() as usize))); assert!(bss_range.contains(&(v.as_ptr() as usize)));
drop(v); drop(v);

View file

@ -1,20 +1,15 @@
use super::{PageTable, PageTableEntry, PTEFlags}; use super::{frame_alloc, FrameTracker};
use super::{VirtPageNum, VirtAddr, PhysPageNum, PhysAddr}; use super::{PTEFlags, PageTable, PageTableEntry};
use super::{FrameTracker, frame_alloc}; use super::{PhysAddr, PhysPageNum, VirtAddr, VirtPageNum};
use super::{VPNRange, StepByOne}; use super::{StepByOne, VPNRange};
use alloc::collections::BTreeMap; use crate::config::{MEMORY_END, MMIO, PAGE_SIZE, TRAMPOLINE};
use alloc::vec::Vec;
use riscv::register::satp;
use alloc::sync::Arc;
use lazy_static::*;
use crate::sync::UPSafeCell; use crate::sync::UPSafeCell;
use crate::config::{ use alloc::collections::BTreeMap;
MEMORY_END, use alloc::sync::Arc;
PAGE_SIZE, use alloc::vec::Vec;
TRAMPOLINE,
MMIO,
};
use core::arch::asm; use core::arch::asm;
use lazy_static::*;
use riscv::register::satp;
extern "C" { extern "C" {
fn stext(); fn stext();
@ -30,9 +25,8 @@ extern "C" {
} }
lazy_static! { lazy_static! {
pub static ref KERNEL_SPACE: Arc<UPSafeCell<MemorySet>> = Arc::new(unsafe { pub static ref KERNEL_SPACE: Arc<UPSafeCell<MemorySet>> =
UPSafeCell::new(MemorySet::new_kernel()) Arc::new(unsafe { UPSafeCell::new(MemorySet::new_kernel()) });
});
} }
pub fn kernel_token() -> usize { pub fn kernel_token() -> usize {
@ -55,17 +49,24 @@ impl MemorySet {
self.page_table.token() self.page_table.token()
} }
/// Assume that no conflicts. /// Assume that no conflicts.
pub fn insert_framed_area(&mut self, start_va: VirtAddr, end_va: VirtAddr, permission: MapPermission) { pub fn insert_framed_area(
self.push(MapArea::new( &mut self,
start_va, start_va: VirtAddr,
end_va, end_va: VirtAddr,
MapType::Framed, permission: MapPermission,
permission, ) {
), None); self.push(
MapArea::new(start_va, end_va, MapType::Framed, permission),
None,
);
} }
pub fn remove_area_with_start_vpn(&mut self, start_vpn: VirtPageNum) { pub fn remove_area_with_start_vpn(&mut self, start_vpn: VirtPageNum) {
if let Some((idx, area)) = self.areas.iter_mut().enumerate() if let Some((idx, area)) = self
.find(|(_, area)| area.vpn_range.get_start() == start_vpn) { .areas
.iter_mut()
.enumerate()
.find(|(_, area)| area.vpn_range.get_start() == start_vpn)
{
area.unmap(&mut self.page_table); area.unmap(&mut self.page_table);
self.areas.remove(idx); self.areas.remove(idx);
} }
@ -94,50 +95,71 @@ impl MemorySet {
println!(".text [{:#x}, {:#x})", stext as usize, etext as usize); println!(".text [{:#x}, {:#x})", stext as usize, etext as usize);
println!(".rodata [{:#x}, {:#x})", srodata as usize, erodata as usize); println!(".rodata [{:#x}, {:#x})", srodata as usize, erodata as usize);
println!(".data [{:#x}, {:#x})", sdata as usize, edata as usize); println!(".data [{:#x}, {:#x})", sdata as usize, edata as usize);
println!(".bss [{:#x}, {:#x})", sbss_with_stack as usize, ebss as usize); println!(
".bss [{:#x}, {:#x})",
sbss_with_stack as usize, ebss as usize
);
println!("mapping .text section"); println!("mapping .text section");
memory_set.push(MapArea::new( memory_set.push(
(stext as usize).into(), MapArea::new(
(etext as usize).into(), (stext as usize).into(),
MapType::Identical, (etext as usize).into(),
MapPermission::R | MapPermission::X, MapType::Identical,
), None); MapPermission::R | MapPermission::X,
),
None,
);
println!("mapping .rodata section"); println!("mapping .rodata section");
memory_set.push(MapArea::new( memory_set.push(
(srodata as usize).into(), MapArea::new(
(erodata as usize).into(), (srodata as usize).into(),
MapType::Identical, (erodata as usize).into(),
MapPermission::R, MapType::Identical,
), None); MapPermission::R,
),
None,
);
println!("mapping .data section"); println!("mapping .data section");
memory_set.push(MapArea::new( memory_set.push(
(sdata as usize).into(), MapArea::new(
(edata as usize).into(), (sdata as usize).into(),
MapType::Identical, (edata as usize).into(),
MapPermission::R | MapPermission::W,
), None);
println!("mapping .bss section");
memory_set.push(MapArea::new(
(sbss_with_stack as usize).into(),
(ebss as usize).into(),
MapType::Identical,
MapPermission::R | MapPermission::W,
), None);
println!("mapping physical memory");
memory_set.push(MapArea::new(
(ekernel as usize).into(),
MEMORY_END.into(),
MapType::Identical,
MapPermission::R | MapPermission::W,
), None);
println!("mapping memory-mapped registers");
for pair in MMIO {
memory_set.push(MapArea::new(
(*pair).0.into(),
((*pair).0 + (*pair).1).into(),
MapType::Identical, MapType::Identical,
MapPermission::R | MapPermission::W, MapPermission::R | MapPermission::W,
), None); ),
None,
);
println!("mapping .bss section");
memory_set.push(
MapArea::new(
(sbss_with_stack as usize).into(),
(ebss as usize).into(),
MapType::Identical,
MapPermission::R | MapPermission::W,
),
None,
);
println!("mapping physical memory");
memory_set.push(
MapArea::new(
(ekernel as usize).into(),
MEMORY_END.into(),
MapType::Identical,
MapPermission::R | MapPermission::W,
),
None,
);
println!("mapping memory-mapped registers");
for pair in MMIO {
memory_set.push(
MapArea::new(
(*pair).0.into(),
((*pair).0 + (*pair).1).into(),
MapType::Identical,
MapPermission::R | MapPermission::W,
),
None,
);
} }
memory_set memory_set
} }
@ -161,26 +183,31 @@ impl MemorySet {
let end_va: VirtAddr = ((ph.virtual_addr() + ph.mem_size()) as usize).into(); let end_va: VirtAddr = ((ph.virtual_addr() + ph.mem_size()) as usize).into();
let mut map_perm = MapPermission::U; let mut map_perm = MapPermission::U;
let ph_flags = ph.flags(); let ph_flags = ph.flags();
if ph_flags.is_read() { map_perm |= MapPermission::R; } if ph_flags.is_read() {
if ph_flags.is_write() { map_perm |= MapPermission::W; } map_perm |= MapPermission::R;
if ph_flags.is_execute() { map_perm |= MapPermission::X; } }
let map_area = MapArea::new( if ph_flags.is_write() {
start_va, map_perm |= MapPermission::W;
end_va, }
MapType::Framed, if ph_flags.is_execute() {
map_perm, map_perm |= MapPermission::X;
); }
let map_area = MapArea::new(start_va, end_va, MapType::Framed, map_perm);
max_end_vpn = map_area.vpn_range.get_end(); max_end_vpn = map_area.vpn_range.get_end();
memory_set.push( memory_set.push(
map_area, map_area,
Some(&elf.input[ph.offset() as usize..(ph.offset() + ph.file_size()) as usize]) Some(&elf.input[ph.offset() as usize..(ph.offset() + ph.file_size()) as usize]),
); );
} }
} }
let max_end_va: VirtAddr = max_end_vpn.into(); let max_end_va: VirtAddr = max_end_vpn.into();
let mut user_stack_base: usize = max_end_va.into(); let mut user_stack_base: usize = max_end_va.into();
user_stack_base += PAGE_SIZE; user_stack_base += PAGE_SIZE;
(memory_set, user_stack_base, elf.header.pt2.entry_point() as usize) (
memory_set,
user_stack_base,
elf.header.pt2.entry_point() as usize,
)
} }
pub fn from_existed_user(user_space: &MemorySet) -> MemorySet { pub fn from_existed_user(user_space: &MemorySet) -> MemorySet {
let mut memory_set = Self::new_bare(); let mut memory_set = Self::new_bare();
@ -194,7 +221,9 @@ impl MemorySet {
for vpn in area.vpn_range { for vpn in area.vpn_range {
let src_ppn = user_space.translate(vpn).unwrap().ppn(); let src_ppn = user_space.translate(vpn).unwrap().ppn();
let dst_ppn = memory_set.translate(vpn).unwrap().ppn(); let dst_ppn = memory_set.translate(vpn).unwrap().ppn();
dst_ppn.get_bytes_array().copy_from_slice(src_ppn.get_bytes_array()); dst_ppn
.get_bytes_array()
.copy_from_slice(src_ppn.get_bytes_array());
} }
} }
memory_set memory_set
@ -227,7 +256,7 @@ impl MapArea {
start_va: VirtAddr, start_va: VirtAddr,
end_va: VirtAddr, end_va: VirtAddr,
map_type: MapType, map_type: MapType,
map_perm: MapPermission map_perm: MapPermission,
) -> Self { ) -> Self {
let start_vpn: VirtPageNum = start_va.floor(); let start_vpn: VirtPageNum = start_va.floor();
let end_vpn: VirtPageNum = end_va.ceil(); let end_vpn: VirtPageNum = end_va.ceil();
@ -262,11 +291,8 @@ impl MapArea {
page_table.map(vpn, ppn, pte_flags); page_table.map(vpn, ppn, pte_flags);
} }
pub fn unmap_one(&mut self, page_table: &mut PageTable, vpn: VirtPageNum) { pub fn unmap_one(&mut self, page_table: &mut PageTable, vpn: VirtPageNum) {
match self.map_type { if self.map_type == MapType::Framed {
MapType::Framed => { self.data_frames.remove(&vpn);
self.data_frames.remove(&vpn);
}
_ => {}
} }
page_table.unmap(vpn); page_table.unmap(vpn);
} }
@ -325,17 +351,26 @@ pub fn remap_test() {
let mid_text: VirtAddr = ((stext as usize + etext as usize) / 2).into(); let mid_text: VirtAddr = ((stext as usize + etext as usize) / 2).into();
let mid_rodata: VirtAddr = ((srodata as usize + erodata as usize) / 2).into(); let mid_rodata: VirtAddr = ((srodata as usize + erodata as usize) / 2).into();
let mid_data: VirtAddr = ((sdata as usize + edata as usize) / 2).into(); let mid_data: VirtAddr = ((sdata as usize + edata as usize) / 2).into();
assert_eq!( assert!(
kernel_space.page_table.translate(mid_text.floor()).unwrap().writable(), !kernel_space
false .page_table
.translate(mid_text.floor())
.unwrap()
.writable(),
); );
assert_eq!( assert!(
kernel_space.page_table.translate(mid_rodata.floor()).unwrap().writable(), !kernel_space
false, .page_table
.translate(mid_rodata.floor())
.unwrap()
.writable(),
); );
assert_eq!( assert!(
kernel_space.page_table.translate(mid_data.floor()).unwrap().executable(), !kernel_space
false, .page_table
.translate(mid_data.floor())
.unwrap()
.executable(),
); );
println!("remap_test passed!"); println!("remap_test passed!");
} }

View file

@ -1,25 +1,19 @@
mod heap_allocator;
mod address; mod address;
mod frame_allocator; mod frame_allocator;
mod page_table; mod heap_allocator;
mod memory_set; mod memory_set;
mod page_table;
use page_table::PTEFlags;
use address::VPNRange; use address::VPNRange;
pub use address::{PhysAddr, VirtAddr, PhysPageNum, VirtPageNum, StepByOne}; pub use address::{PhysAddr, PhysPageNum, StepByOne, VirtAddr, VirtPageNum};
pub use frame_allocator::{FrameTracker, frame_alloc, frame_dealloc,}; pub use frame_allocator::{frame_alloc, frame_dealloc, FrameTracker};
pub use page_table::{
PageTable,
PageTableEntry,
translated_byte_buffer,
translated_str,
translated_ref,
translated_refmut,
UserBuffer,
UserBufferIterator,
};
pub use memory_set::{MemorySet, KERNEL_SPACE, MapPermission, kernel_token};
pub use memory_set::remap_test; pub use memory_set::remap_test;
pub use memory_set::{kernel_token, MapPermission, MemorySet, KERNEL_SPACE};
use page_table::PTEFlags;
pub use page_table::{
translated_byte_buffer, translated_ref, translated_refmut, translated_str, PageTable,
PageTableEntry, UserBuffer, UserBufferIterator,
};
pub fn init() { pub fn init() {
heap_allocator::init_heap(); heap_allocator::init_heap();

View file

@ -1,15 +1,7 @@
use super::{ use super::{frame_alloc, FrameTracker, PhysAddr, PhysPageNum, StepByOne, VirtAddr, VirtPageNum};
frame_alloc,
PhysPageNum,
FrameTracker,
VirtPageNum,
VirtAddr,
PhysAddr,
StepByOne
};
use alloc::vec::Vec;
use alloc::vec;
use alloc::string::String; use alloc::string::String;
use alloc::vec;
use alloc::vec::Vec;
use bitflags::*; use bitflags::*;
bitflags! { bitflags! {
@ -38,9 +30,7 @@ impl PageTableEntry {
} }
} }
pub fn empty() -> Self { pub fn empty() -> Self {
PageTableEntry { PageTableEntry { bits: 0 }
bits: 0,
}
} }
pub fn ppn(&self) -> PhysPageNum { pub fn ppn(&self) -> PhysPageNum {
(self.bits >> 10 & ((1usize << 44) - 1)).into() (self.bits >> 10 & ((1usize << 44) - 1)).into()
@ -87,8 +77,8 @@ impl PageTable {
let idxs = vpn.indexes(); let idxs = vpn.indexes();
let mut ppn = self.root_ppn; let mut ppn = self.root_ppn;
let mut result: Option<&mut PageTableEntry> = None; let mut result: Option<&mut PageTableEntry> = None;
for i in 0..3 { for (i, idx) in idxs.iter().enumerate() {
let pte = &mut ppn.get_pte_array()[idxs[i]]; let pte = &mut ppn.get_pte_array()[*idx];
if i == 2 { if i == 2 {
result = Some(pte); result = Some(pte);
break; break;
@ -106,8 +96,8 @@ impl PageTable {
let idxs = vpn.indexes(); let idxs = vpn.indexes();
let mut ppn = self.root_ppn; let mut ppn = self.root_ppn;
let mut result: Option<&mut PageTableEntry> = None; let mut result: Option<&mut PageTableEntry> = None;
for i in 0..3 { for (i, idx) in idxs.iter().enumerate() {
let pte = &mut ppn.get_pte_array()[idxs[i]]; let pte = &mut ppn.get_pte_array()[*idx];
if i == 2 { if i == 2 {
result = Some(pte); result = Some(pte);
break; break;
@ -132,17 +122,15 @@ impl PageTable {
*pte = PageTableEntry::empty(); *pte = PageTableEntry::empty();
} }
pub fn translate(&self, vpn: VirtPageNum) -> Option<PageTableEntry> { pub fn translate(&self, vpn: VirtPageNum) -> Option<PageTableEntry> {
self.find_pte(vpn) self.find_pte(vpn).map(|pte| *pte)
.map(|pte| {pte.clone()})
} }
pub fn translate_va(&self, va: VirtAddr) -> Option<PhysAddr> { pub fn translate_va(&self, va: VirtAddr) -> Option<PhysAddr> {
self.find_pte(va.clone().floor()) self.find_pte(va.clone().floor()).map(|pte| {
.map(|pte| { let aligned_pa: PhysAddr = pte.ppn().into();
let aligned_pa: PhysAddr = pte.ppn().into(); let offset = va.page_offset();
let offset = va.page_offset(); let aligned_pa_usize: usize = aligned_pa.into();
let aligned_pa_usize: usize = aligned_pa.into(); (aligned_pa_usize + offset).into()
(aligned_pa_usize + offset).into() })
})
} }
pub fn token(&self) -> usize { pub fn token(&self) -> usize {
8usize << 60 | self.root_ppn.0 8usize << 60 | self.root_ppn.0
@ -157,10 +145,7 @@ pub fn translated_byte_buffer(token: usize, ptr: *const u8, len: usize) -> Vec<&
while start < end { while start < end {
let start_va = VirtAddr::from(start); let start_va = VirtAddr::from(start);
let mut vpn = start_va.floor(); let mut vpn = start_va.floor();
let ppn = page_table let ppn = page_table.translate(vpn).unwrap().ppn();
.translate(vpn)
.unwrap()
.ppn();
vpn.step(); vpn.step();
let mut end_va: VirtAddr = vpn.into(); let mut end_va: VirtAddr = vpn.into();
end_va = end_va.min(VirtAddr::from(end)); end_va = end_va.min(VirtAddr::from(end));
@ -180,7 +165,10 @@ pub fn translated_str(token: usize, ptr: *const u8) -> String {
let mut string = String::new(); let mut string = String::new();
let mut va = ptr as usize; let mut va = ptr as usize;
loop { loop {
let ch: u8 = *(page_table.translate_va(VirtAddr::from(va)).unwrap().get_mut()); let ch: u8 = *(page_table
.translate_va(VirtAddr::from(va))
.unwrap()
.get_mut());
if ch == 0 { if ch == 0 {
break; break;
} }
@ -192,13 +180,19 @@ pub fn translated_str(token: usize, ptr: *const u8) -> String {
pub fn translated_ref<T>(token: usize, ptr: *const T) -> &'static T { pub fn translated_ref<T>(token: usize, ptr: *const T) -> &'static T {
let page_table = PageTable::from_token(token); let page_table = PageTable::from_token(token);
page_table.translate_va(VirtAddr::from(ptr as usize)).unwrap().get_ref() page_table
.translate_va(VirtAddr::from(ptr as usize))
.unwrap()
.get_ref()
} }
pub fn translated_refmut<T>(token: usize, ptr: *mut T) -> &'static mut T { pub fn translated_refmut<T>(token: usize, ptr: *mut T) -> &'static mut T {
let page_table = PageTable::from_token(token); let page_table = PageTable::from_token(token);
let va = ptr as usize; let va = ptr as usize;
page_table.translate_va(VirtAddr::from(va)).unwrap().get_mut() page_table
.translate_va(VirtAddr::from(va))
.unwrap()
.get_mut()
} }
pub struct UserBuffer { pub struct UserBuffer {

View file

@ -1,5 +1,7 @@
#![allow(unused)] #![allow(unused)]
use core::arch::asm;
const SBI_SET_TIMER: usize = 0; const SBI_SET_TIMER: usize = 0;
const SBI_CONSOLE_PUTCHAR: usize = 1; const SBI_CONSOLE_PUTCHAR: usize = 1;
const SBI_CONSOLE_GETCHAR: usize = 2; const SBI_CONSOLE_GETCHAR: usize = 2;
@ -41,4 +43,3 @@ pub fn shutdown() -> ! {
sbi_call(SBI_SHUTDOWN, 0, 0, 0); sbi_call(SBI_SHUTDOWN, 0, 0, 0);
panic!("It should shutdown!"); panic!("It should shutdown!");
} }

39
os/src/sync/condvar.rs Normal file
View file

@ -0,0 +1,39 @@
use crate::sync::{Mutex, UPSafeCell};
use crate::task::{add_task, block_current_and_run_next, current_task, TaskControlBlock};
use alloc::{collections::VecDeque, sync::Arc};
pub struct Condvar {
pub inner: UPSafeCell<CondvarInner>,
}
pub struct CondvarInner {
pub wait_queue: VecDeque<Arc<TaskControlBlock>>,
}
impl Condvar {
pub fn new() -> Self {
Self {
inner: unsafe {
UPSafeCell::new(CondvarInner {
wait_queue: VecDeque::new(),
})
},
}
}
pub fn signal(&self) {
let mut inner = self.inner.exclusive_access();
if let Some(task) = inner.wait_queue.pop_front() {
add_task(task);
}
}
pub fn wait(&self, mutex: Arc<dyn Mutex>) {
mutex.unlock();
let mut inner = self.inner.exclusive_access();
inner.wait_queue.push_back(current_task().unwrap());
drop(inner);
block_current_and_run_next();
mutex.lock();
}
}

View file

@ -1,7 +1,9 @@
mod up; mod condvar;
mod mutex; mod mutex;
mod semaphore; mod semaphore;
mod up;
pub use up::UPSafeCell; pub use condvar::Condvar;
pub use mutex::{Mutex, MutexSpin, MutexBlocking}; pub use mutex::{Mutex, MutexBlocking, MutexSpin};
pub use semaphore::Semaphore; pub use semaphore::Semaphore;
pub use up::UPSafeCell;

View file

@ -1,8 +1,8 @@
use super::UPSafeCell; use super::UPSafeCell;
use crate::task::{block_current_and_run_next, suspend_current_and_run_next};
use crate::task::TaskControlBlock; use crate::task::TaskControlBlock;
use crate::task::{add_task, current_task}; use crate::task::{add_task, current_task};
use alloc::{sync::Arc, collections::VecDeque}; use crate::task::{block_current_and_run_next, suspend_current_and_run_next};
use alloc::{collections::VecDeque, sync::Arc};
pub trait Mutex: Sync + Send { pub trait Mutex: Sync + Send {
fn lock(&self); fn lock(&self);
@ -78,10 +78,11 @@ impl Mutex for MutexBlocking {
fn unlock(&self) { fn unlock(&self) {
let mut mutex_inner = self.inner.exclusive_access(); let mut mutex_inner = self.inner.exclusive_access();
assert_eq!(mutex_inner.locked, true); assert!(mutex_inner.locked);
mutex_inner.locked = false;
if let Some(waking_task) = mutex_inner.wait_queue.pop_front() { if let Some(waking_task) = mutex_inner.wait_queue.pop_front() {
add_task(waking_task); add_task(waking_task);
} else {
mutex_inner.locked = false;
} }
} }
} }

View file

@ -1,6 +1,6 @@
use alloc::{sync::Arc, collections::VecDeque};
use crate::task::{add_task, TaskControlBlock, current_task, block_current_and_run_next};
use crate::sync::UPSafeCell; use crate::sync::UPSafeCell;
use crate::task::{add_task, block_current_and_run_next, current_task, TaskControlBlock};
use alloc::{collections::VecDeque, sync::Arc};
pub struct Semaphore { pub struct Semaphore {
pub inner: UPSafeCell<SemaphoreInner>, pub inner: UPSafeCell<SemaphoreInner>,
@ -14,12 +14,12 @@ pub struct SemaphoreInner {
impl Semaphore { impl Semaphore {
pub fn new(res_count: usize) -> Self { pub fn new(res_count: usize) -> Self {
Self { Self {
inner: unsafe { UPSafeCell::new( inner: unsafe {
SemaphoreInner { UPSafeCell::new(SemaphoreInner {
count: res_count as isize, count: res_count as isize,
wait_queue: VecDeque::new(), wait_queue: VecDeque::new(),
} })
)}, },
} }
} }

View file

@ -18,7 +18,9 @@ impl<T> UPSafeCell<T> {
/// User is responsible to guarantee that inner struct is only used in /// User is responsible to guarantee that inner struct is only used in
/// uniprocessor. /// uniprocessor.
pub unsafe fn new(value: T) -> Self { pub unsafe fn new(value: T) -> Self {
Self { inner: RefCell::new(value) } Self {
inner: RefCell::new(value),
}
} }
/// Panic if the data has been borrowed. /// Panic if the data has been borrowed.
pub fn exclusive_access(&self) -> RefMut<'_, T> { pub fn exclusive_access(&self) -> RefMut<'_, T> {

View file

@ -1,11 +1,6 @@
use crate::mm::{ use crate::fs::{make_pipe, open_file, OpenFlags};
UserBuffer, use crate::mm::{translated_byte_buffer, translated_refmut, translated_str, UserBuffer};
translated_byte_buffer, use crate::task::{current_process, current_user_token};
translated_refmut,
translated_str,
};
use crate::task::{current_user_token, current_process};
use crate::fs::{make_pipe, OpenFlags, open_file};
use alloc::sync::Arc; use alloc::sync::Arc;
pub fn sys_write(fd: usize, buf: *const u8, len: usize) -> isize { pub fn sys_write(fd: usize, buf: *const u8, len: usize) -> isize {
@ -22,9 +17,7 @@ pub fn sys_write(fd: usize, buf: *const u8, len: usize) -> isize {
let file = file.clone(); let file = file.clone();
// release current task TCB manually to avoid multi-borrow // release current task TCB manually to avoid multi-borrow
drop(inner); drop(inner);
file.write( file.write(UserBuffer::new(translated_byte_buffer(token, buf, len))) as isize
UserBuffer::new(translated_byte_buffer(token, buf, len))
) as isize
} else { } else {
-1 -1
} }
@ -44,9 +37,7 @@ pub fn sys_read(fd: usize, buf: *const u8, len: usize) -> isize {
} }
// release current task TCB manually to avoid multi-borrow // release current task TCB manually to avoid multi-borrow
drop(inner); drop(inner);
file.read( file.read(UserBuffer::new(translated_byte_buffer(token, buf, len))) as isize
UserBuffer::new(translated_byte_buffer(token, buf, len))
) as isize
} else { } else {
-1 -1
} }
@ -56,10 +47,7 @@ pub fn sys_open(path: *const u8, flags: u32) -> isize {
let process = current_process(); let process = current_process();
let token = current_user_token(); let token = current_user_token();
let path = translated_str(token, path); let path = translated_str(token, path);
if let Some(inode) = open_file( if let Some(inode) = open_file(path.as_str(), OpenFlags::from_bits(flags).unwrap()) {
path.as_str(),
OpenFlags::from_bits(flags).unwrap()
) {
let mut inner = process.inner_exclusive_access(); let mut inner = process.inner_exclusive_access();
let fd = inner.alloc_fd(); let fd = inner.alloc_fd();
inner.fd_table[fd] = Some(inode); inner.fd_table[fd] = Some(inode);

View file

@ -7,6 +7,7 @@ const SYSCALL_WRITE: usize = 64;
const SYSCALL_EXIT: usize = 93; const SYSCALL_EXIT: usize = 93;
const SYSCALL_SLEEP: usize = 101; const SYSCALL_SLEEP: usize = 101;
const SYSCALL_YIELD: usize = 124; const SYSCALL_YIELD: usize = 124;
const SYSCALL_KILL: usize = 129;
const SYSCALL_GET_TIME: usize = 169; const SYSCALL_GET_TIME: usize = 169;
const SYSCALL_GETPID: usize = 172; const SYSCALL_GETPID: usize = 172;
const SYSCALL_FORK: usize = 220; const SYSCALL_FORK: usize = 220;
@ -21,20 +22,23 @@ const SYSCALL_MUTEX_UNLOCK: usize = 1012;
const SYSCALL_SEMAPHORE_CREATE: usize = 1020; const SYSCALL_SEMAPHORE_CREATE: usize = 1020;
const SYSCALL_SEMAPHORE_UP: usize = 1021; const SYSCALL_SEMAPHORE_UP: usize = 1021;
const SYSCALL_SEMAPHORE_DOWN: usize = 1022; const SYSCALL_SEMAPHORE_DOWN: usize = 1022;
const SYSCALL_CONDVAR_CREATE: usize = 1030;
const SYSCALL_CONDVAR_SIGNAL: usize = 1031;
const SYSCALL_CONDVAR_WAIT: usize = 1032;
mod fs; mod fs;
mod process; mod process;
mod thread;
mod sync; mod sync;
mod thread;
use fs::*; use fs::*;
use process::*; use process::*;
use thread::*;
use sync::*; use sync::*;
use thread::*;
pub fn syscall(syscall_id: usize, args: [usize; 3]) -> isize { pub fn syscall(syscall_id: usize, args: [usize; 3]) -> isize {
match syscall_id { match syscall_id {
SYSCALL_DUP=> sys_dup(args[0]), SYSCALL_DUP => sys_dup(args[0]),
SYSCALL_OPEN => sys_open(args[0] as *const u8, args[1] as u32), SYSCALL_OPEN => sys_open(args[0] as *const u8, args[1] as u32),
SYSCALL_CLOSE => sys_close(args[0]), SYSCALL_CLOSE => sys_close(args[0]),
SYSCALL_PIPE => sys_pipe(args[0] as *mut usize), SYSCALL_PIPE => sys_pipe(args[0] as *mut usize),
@ -43,6 +47,7 @@ pub fn syscall(syscall_id: usize, args: [usize; 3]) -> isize {
SYSCALL_EXIT => sys_exit(args[0] as i32), SYSCALL_EXIT => sys_exit(args[0] as i32),
SYSCALL_SLEEP => sys_sleep(args[0]), SYSCALL_SLEEP => sys_sleep(args[0]),
SYSCALL_YIELD => sys_yield(), SYSCALL_YIELD => sys_yield(),
SYSCALL_KILL => sys_kill(args[0], args[1] as u32),
SYSCALL_GET_TIME => sys_get_time(), SYSCALL_GET_TIME => sys_get_time(),
SYSCALL_GETPID => sys_getpid(), SYSCALL_GETPID => sys_getpid(),
SYSCALL_FORK => sys_fork(), SYSCALL_FORK => sys_fork(),
@ -54,10 +59,12 @@ pub fn syscall(syscall_id: usize, args: [usize; 3]) -> isize {
SYSCALL_MUTEX_CREATE => sys_mutex_create(args[0] == 1), SYSCALL_MUTEX_CREATE => sys_mutex_create(args[0] == 1),
SYSCALL_MUTEX_LOCK => sys_mutex_lock(args[0]), SYSCALL_MUTEX_LOCK => sys_mutex_lock(args[0]),
SYSCALL_MUTEX_UNLOCK => sys_mutex_unlock(args[0]), SYSCALL_MUTEX_UNLOCK => sys_mutex_unlock(args[0]),
SYSCALL_SEMAPHORE_CREATE => sys_semaphore_creare(args[0]), SYSCALL_SEMAPHORE_CREATE => sys_semaphore_create(args[0]),
SYSCALL_SEMAPHORE_UP => sys_semaphore_up(args[0]), SYSCALL_SEMAPHORE_UP => sys_semaphore_up(args[0]),
SYSCALL_SEMAPHORE_DOWN => sys_semaphore_down(args[0]), SYSCALL_SEMAPHORE_DOWN => sys_semaphore_down(args[0]),
SYSCALL_CONDVAR_CREATE => sys_condvar_create(args[0]),
SYSCALL_CONDVAR_SIGNAL => sys_condvar_signal(args[0]),
SYSCALL_CONDVAR_WAIT => sys_condvar_wait(args[0], args[1]),
_ => panic!("Unsupported syscall_id: {}", syscall_id), _ => panic!("Unsupported syscall_id: {}", syscall_id),
} }
} }

View file

@ -1,23 +1,13 @@
use crate::fs::{open_file, OpenFlags};
use crate::mm::{translated_ref, translated_refmut, translated_str};
use crate::task::{ use crate::task::{
suspend_current_and_run_next, current_process, current_task, current_user_token, exit_current_and_run_next, pid2process,
exit_current_and_run_next, suspend_current_and_run_next, SignalFlags,
current_task,
current_process,
current_user_token,
}; };
use crate::timer::get_time_ms; use crate::timer::get_time_ms;
use crate::mm::{ use alloc::string::String;
translated_str,
translated_refmut,
translated_ref,
};
use crate::fs::{
open_file,
OpenFlags,
};
use alloc::sync::Arc; use alloc::sync::Arc;
use alloc::vec::Vec; use alloc::vec::Vec;
use alloc::string::String;
pub fn sys_exit(exit_code: i32) -> ! { pub fn sys_exit(exit_code: i32) -> ! {
exit_current_and_run_next(exit_code); exit_current_and_run_next(exit_code);
@ -61,7 +51,9 @@ pub fn sys_exec(path: *const u8, mut args: *const usize) -> isize {
break; break;
} }
args_vec.push(translated_str(token, arg_str_ptr as *const u8)); args_vec.push(translated_str(token, arg_str_ptr as *const u8));
unsafe { args = args.add(1); } unsafe {
args = args.add(1);
}
} }
if let Some(app_inode) = open_file(path.as_str(), OpenFlags::RDONLY) { if let Some(app_inode) = open_file(path.as_str(), OpenFlags::RDONLY) {
let all_data = app_inode.read_all(); let all_data = app_inode.read_all();
@ -82,21 +74,19 @@ pub fn sys_waitpid(pid: isize, exit_code_ptr: *mut i32) -> isize {
// find a child process // find a child process
let mut inner = process.inner_exclusive_access(); let mut inner = process.inner_exclusive_access();
if inner.children if !inner
.children
.iter() .iter()
.find(|p| {pid == -1 || pid as usize == p.getpid()}) .any(|p| pid == -1 || pid as usize == p.getpid())
.is_none() { {
return -1; return -1;
// ---- release current PCB // ---- release current PCB
} }
let pair = inner.children let pair = inner.children.iter().enumerate().find(|(_, p)| {
.iter() // ++++ temporarily access child PCB exclusively
.enumerate() p.inner_exclusive_access().is_zombie && (pid == -1 || pid as usize == p.getpid())
.find(|(_, p)| { // ++++ release child PCB
// ++++ temporarily access child PCB exclusively });
p.inner_exclusive_access().is_zombie && (pid == -1 || pid as usize == p.getpid())
// ++++ release child PCB
});
if let Some((idx, _)) = pair { if let Some((idx, _)) = pair {
let child = inner.children.remove(idx); let child = inner.children.remove(idx);
// confirm that child will be deallocated after being removed from children list // confirm that child will be deallocated after being removed from children list
@ -112,3 +102,16 @@ pub fn sys_waitpid(pid: isize, exit_code_ptr: *mut i32) -> isize {
} }
// ---- release current PCB automatically // ---- release current PCB automatically
} }
pub fn sys_kill(pid: usize, signal: u32) -> isize {
if let Some(process) = pid2process(pid) {
if let Some(flag) = SignalFlags::from_bits(signal) {
process.inner_exclusive_access().signals |= flag;
0
} else {
-1
}
} else {
-1
}
}

View file

@ -1,6 +1,6 @@
use crate::task::{current_task, current_process, block_current_and_run_next}; use crate::sync::{Condvar, Mutex, MutexBlocking, MutexSpin, Semaphore};
use crate::sync::{MutexSpin, MutexBlocking, Semaphore}; use crate::task::{block_current_and_run_next, current_process, current_task};
use crate::timer::{get_time_ms, add_timer}; use crate::timer::{add_timer, get_time_ms};
use alloc::sync::Arc; use alloc::sync::Arc;
pub fn sys_sleep(ms: usize) -> isize { pub fn sys_sleep(ms: usize) -> isize {
@ -13,21 +13,23 @@ pub fn sys_sleep(ms: usize) -> isize {
pub fn sys_mutex_create(blocking: bool) -> isize { pub fn sys_mutex_create(blocking: bool) -> isize {
let process = current_process(); let process = current_process();
let mutex: Option<Arc<dyn Mutex>> = if !blocking {
Some(Arc::new(MutexSpin::new()))
} else {
Some(Arc::new(MutexBlocking::new()))
};
let mut process_inner = process.inner_exclusive_access(); let mut process_inner = process.inner_exclusive_access();
if let Some(id) = process_inner if let Some(id) = process_inner
.mutex_list .mutex_list
.iter() .iter()
.enumerate() .enumerate()
.find(|(_, item)| item.is_none()) .find(|(_, item)| item.is_none())
.map(|(id, _)| id) { .map(|(id, _)| id)
process_inner.mutex_list[id] = if !blocking { {
Some(Arc::new(MutexSpin::new())) process_inner.mutex_list[id] = mutex;
} else {
Some(Arc::new(MutexBlocking::new()))
};
id as isize id as isize
} else { } else {
process_inner.mutex_list.push(Some(Arc::new(MutexSpin::new()))); process_inner.mutex_list.push(mutex);
process_inner.mutex_list.len() as isize - 1 process_inner.mutex_list.len() as isize - 1
} }
} }
@ -52,7 +54,7 @@ pub fn sys_mutex_unlock(mutex_id: usize) -> isize {
0 0
} }
pub fn sys_semaphore_creare(res_count: usize) -> isize { pub fn sys_semaphore_create(res_count: usize) -> isize {
let process = current_process(); let process = current_process();
let mut process_inner = process.inner_exclusive_access(); let mut process_inner = process.inner_exclusive_access();
let id = if let Some(id) = process_inner let id = if let Some(id) = process_inner
@ -60,11 +62,14 @@ pub fn sys_semaphore_creare(res_count: usize) -> isize {
.iter() .iter()
.enumerate() .enumerate()
.find(|(_, item)| item.is_none()) .find(|(_, item)| item.is_none())
.map(|(id, _)| id) { .map(|(id, _)| id)
{
process_inner.semaphore_list[id] = Some(Arc::new(Semaphore::new(res_count))); process_inner.semaphore_list[id] = Some(Arc::new(Semaphore::new(res_count)));
id id
} else { } else {
process_inner.semaphore_list.push(Some(Arc::new(Semaphore::new(res_count)))); process_inner
.semaphore_list
.push(Some(Arc::new(Semaphore::new(res_count))));
process_inner.semaphore_list.len() - 1 process_inner.semaphore_list.len() - 1
}; };
id as isize id as isize
@ -87,3 +92,43 @@ pub fn sys_semaphore_down(sem_id: usize) -> isize {
sem.down(); sem.down();
0 0
} }
pub fn sys_condvar_create(_arg: usize) -> isize {
let process = current_process();
let mut process_inner = process.inner_exclusive_access();
let id = if let Some(id) = process_inner
.condvar_list
.iter()
.enumerate()
.find(|(_, item)| item.is_none())
.map(|(id, _)| id)
{
process_inner.condvar_list[id] = Some(Arc::new(Condvar::new()));
id
} else {
process_inner
.condvar_list
.push(Some(Arc::new(Condvar::new())));
process_inner.condvar_list.len() - 1
};
id as isize
}
pub fn sys_condvar_signal(condvar_id: usize) -> isize {
let process = current_process();
let process_inner = process.inner_exclusive_access();
let condvar = Arc::clone(process_inner.condvar_list[condvar_id].as_ref().unwrap());
drop(process_inner);
condvar.signal();
0
}
pub fn sys_condvar_wait(condvar_id: usize, mutex_id: usize) -> isize {
let process = current_process();
let process_inner = process.inner_exclusive_access();
let condvar = Arc::clone(process_inner.condvar_list[condvar_id].as_ref().unwrap());
let mutex = Arc::clone(process_inner.mutex_list[mutex_id].as_ref().unwrap());
drop(process_inner);
condvar.wait(mutex);
0
}

View file

@ -1,5 +1,9 @@
use crate::{
mm::kernel_token,
task::{add_task, current_task, TaskControlBlock},
trap::{trap_handler, TrapContext},
};
use alloc::sync::Arc; use alloc::sync::Arc;
use crate::{mm::kernel_token, task::{TaskControlBlock, add_task, current_task}, trap::{TrapContext, trap_handler}};
pub fn sys_thread_create(entry: usize, arg: usize) -> isize { pub fn sys_thread_create(entry: usize, arg: usize) -> isize {
let task = current_task().unwrap(); let task = current_task().unwrap();
@ -7,7 +11,11 @@ pub fn sys_thread_create(entry: usize, arg: usize) -> isize {
// create a new thread // create a new thread
let new_task = Arc::new(TaskControlBlock::new( let new_task = Arc::new(TaskControlBlock::new(
Arc::clone(&process), Arc::clone(&process),
task.inner_exclusive_access().res.as_ref().unwrap().ustack_base, task.inner_exclusive_access()
.res
.as_ref()
.unwrap()
.ustack_base,
true, true,
)); ));
// add new task to scheduler // add new task to scheduler
@ -35,7 +43,13 @@ pub fn sys_thread_create(entry: usize, arg: usize) -> isize {
} }
pub fn sys_gettid() -> isize { pub fn sys_gettid() -> isize {
current_task().unwrap().inner_exclusive_access().res.as_ref().unwrap().tid as isize current_task()
.unwrap()
.inner_exclusive_access()
.res
.as_ref()
.unwrap()
.tid as isize
} }
/// thread does not exist, return -1 /// thread does not exist, return -1

View file

@ -23,4 +23,3 @@ impl TaskContext {
} }
} }
} }

View file

@ -1,9 +1,12 @@
use alloc::{vec::Vec, sync::{Arc, Weak}};
use lazy_static::*;
use crate::sync::UPSafeCell;
use crate::mm::{KERNEL_SPACE, MapPermission, PhysPageNum, VirtAddr};
use crate::config::{KERNEL_STACK_SIZE, PAGE_SIZE, TRAMPOLINE, TRAP_CONTEXT_BASE, USER_STACK_SIZE};
use super::ProcessControlBlock; use super::ProcessControlBlock;
use crate::config::{KERNEL_STACK_SIZE, PAGE_SIZE, TRAMPOLINE, TRAP_CONTEXT_BASE, USER_STACK_SIZE};
use crate::mm::{MapPermission, PhysPageNum, VirtAddr, KERNEL_SPACE};
use crate::sync::UPSafeCell;
use alloc::{
sync::{Arc, Weak},
vec::Vec,
};
use lazy_static::*;
pub struct RecycleAllocator { pub struct RecycleAllocator {
current: usize, current: usize,
@ -28,21 +31,19 @@ impl RecycleAllocator {
pub fn dealloc(&mut self, id: usize) { pub fn dealloc(&mut self, id: usize) {
assert!(id < self.current); assert!(id < self.current);
assert!( assert!(
self.recycled.iter().find(|i| **i == id).is_none(), !self.recycled.iter().any(|i| *i == id),
"id {} has been deallocated!", id "id {} has been deallocated!",
id
); );
self.recycled.push(id); self.recycled.push(id);
} }
} }
lazy_static! { lazy_static! {
static ref PID_ALLOCATOR: UPSafeCell<RecycleAllocator> = unsafe { static ref PID_ALLOCATOR: UPSafeCell<RecycleAllocator> =
UPSafeCell::new(RecycleAllocator::new()) unsafe { UPSafeCell::new(RecycleAllocator::new()) };
}; static ref KSTACK_ALLOCATOR: UPSafeCell<RecycleAllocator> =
unsafe { UPSafeCell::new(RecycleAllocator::new()) };
static ref KSTACK_ALLOCATOR: UPSafeCell<RecycleAllocator> = unsafe {
UPSafeCell::new(RecycleAllocator::new())
};
} }
pub struct PidHandle(pub usize); pub struct PidHandle(pub usize);
@ -69,13 +70,11 @@ pub struct KernelStack(pub usize);
pub fn kstack_alloc() -> KernelStack { pub fn kstack_alloc() -> KernelStack {
let kstack_id = KSTACK_ALLOCATOR.exclusive_access().alloc(); let kstack_id = KSTACK_ALLOCATOR.exclusive_access().alloc();
let (kstack_bottom, kstack_top) = kernel_stack_position(kstack_id); let (kstack_bottom, kstack_top) = kernel_stack_position(kstack_id);
KERNEL_SPACE KERNEL_SPACE.exclusive_access().insert_framed_area(
.exclusive_access() kstack_bottom.into(),
.insert_framed_area( kstack_top.into(),
kstack_bottom.into(), MapPermission::R | MapPermission::W,
kstack_top.into(), );
MapPermission::R | MapPermission::W,
);
KernelStack(kstack_id) KernelStack(kstack_id)
} }
@ -91,11 +90,15 @@ impl Drop for KernelStack {
impl KernelStack { impl KernelStack {
#[allow(unused)] #[allow(unused)]
pub fn push_on_top<T>(&self, value: T) -> *mut T where pub fn push_on_top<T>(&self, value: T) -> *mut T
T: Sized, { where
T: Sized,
{
let kernel_stack_top = self.get_top(); let kernel_stack_top = self.get_top();
let ptr_mut = (kernel_stack_top - core::mem::size_of::<T>()) as *mut T; let ptr_mut = (kernel_stack_top - core::mem::size_of::<T>()) as *mut T;
unsafe { *ptr_mut = value; } unsafe {
*ptr_mut = value;
}
ptr_mut ptr_mut
} }
pub fn get_top(&self) -> usize { pub fn get_top(&self) -> usize {
@ -142,23 +145,19 @@ impl TaskUserRes {
// alloc user stack // alloc user stack
let ustack_bottom = ustack_bottom_from_tid(self.ustack_base, self.tid); let ustack_bottom = ustack_bottom_from_tid(self.ustack_base, self.tid);
let ustack_top = ustack_bottom + USER_STACK_SIZE; let ustack_top = ustack_bottom + USER_STACK_SIZE;
process_inner process_inner.memory_set.insert_framed_area(
.memory_set ustack_bottom.into(),
.insert_framed_area( ustack_top.into(),
ustack_bottom.into(), MapPermission::R | MapPermission::W | MapPermission::U,
ustack_top.into(), );
MapPermission::R | MapPermission::W | MapPermission::U,
);
// alloc trap_cx // alloc trap_cx
let trap_cx_bottom = trap_cx_bottom_from_tid(self.tid); let trap_cx_bottom = trap_cx_bottom_from_tid(self.tid);
let trap_cx_top = trap_cx_bottom + PAGE_SIZE; let trap_cx_top = trap_cx_bottom + PAGE_SIZE;
process_inner process_inner.memory_set.insert_framed_area(
.memory_set trap_cx_bottom.into(),
.insert_framed_area( trap_cx_top.into(),
trap_cx_bottom.into(), MapPermission::R | MapPermission::W,
trap_cx_top.into(), );
MapPermission::R | MapPermission::W,
);
} }
fn dealloc_user_res(&self) { fn dealloc_user_res(&self) {
@ -167,10 +166,14 @@ impl TaskUserRes {
let mut process_inner = process.inner_exclusive_access(); let mut process_inner = process.inner_exclusive_access();
// dealloc ustack manually // dealloc ustack manually
let ustack_bottom_va: VirtAddr = ustack_bottom_from_tid(self.ustack_base, self.tid).into(); let ustack_bottom_va: VirtAddr = ustack_bottom_from_tid(self.ustack_base, self.tid).into();
process_inner.memory_set.remove_area_with_start_vpn(ustack_bottom_va.into()); process_inner
.memory_set
.remove_area_with_start_vpn(ustack_bottom_va.into());
// dealloc trap_cx manually // dealloc trap_cx manually
let trap_cx_bottom_va: VirtAddr = trap_cx_bottom_from_tid(self.tid).into(); let trap_cx_bottom_va: VirtAddr = trap_cx_bottom_from_tid(self.tid).into();
process_inner.memory_set.remove_area_with_start_vpn(trap_cx_bottom_va.into()); process_inner
.memory_set
.remove_area_with_start_vpn(trap_cx_bottom_va.into());
} }
#[allow(unused)] #[allow(unused)]
@ -197,10 +200,16 @@ impl TaskUserRes {
let process = self.process.upgrade().unwrap(); let process = self.process.upgrade().unwrap();
let process_inner = process.inner_exclusive_access(); let process_inner = process.inner_exclusive_access();
let trap_cx_bottom_va: VirtAddr = trap_cx_bottom_from_tid(self.tid).into(); let trap_cx_bottom_va: VirtAddr = trap_cx_bottom_from_tid(self.tid).into();
process_inner.memory_set.translate(trap_cx_bottom_va.into()).unwrap().ppn() process_inner
.memory_set
.translate(trap_cx_bottom_va.into())
.unwrap()
.ppn()
} }
pub fn ustack_base(&self) -> usize { self.ustack_base } pub fn ustack_base(&self) -> usize {
self.ustack_base
}
pub fn ustack_top(&self) -> usize { pub fn ustack_top(&self) -> usize {
ustack_bottom_from_tid(self.ustack_base, self.tid) + USER_STACK_SIZE ustack_bottom_from_tid(self.ustack_base, self.tid) + USER_STACK_SIZE
} }
@ -212,4 +221,3 @@ impl Drop for TaskUserRes {
self.dealloc_user_res(); self.dealloc_user_res();
} }
} }

View file

@ -1,6 +1,6 @@
use super::{ProcessControlBlock, TaskControlBlock};
use crate::sync::UPSafeCell; use crate::sync::UPSafeCell;
use super::TaskControlBlock; use alloc::collections::{BTreeMap, VecDeque};
use alloc::collections::VecDeque;
use alloc::sync::Arc; use alloc::sync::Arc;
use lazy_static::*; use lazy_static::*;
@ -11,7 +11,9 @@ pub struct TaskManager {
/// A simple FIFO scheduler. /// A simple FIFO scheduler.
impl TaskManager { impl TaskManager {
pub fn new() -> Self { pub fn new() -> Self {
Self { ready_queue: VecDeque::new(), } Self {
ready_queue: VecDeque::new(),
}
} }
pub fn add(&mut self, task: Arc<TaskControlBlock>) { pub fn add(&mut self, task: Arc<TaskControlBlock>) {
self.ready_queue.push_back(task); self.ready_queue.push_back(task);
@ -22,9 +24,10 @@ impl TaskManager {
} }
lazy_static! { lazy_static! {
pub static ref TASK_MANAGER: UPSafeCell<TaskManager> = unsafe { pub static ref TASK_MANAGER: UPSafeCell<TaskManager> =
UPSafeCell::new(TaskManager::new()) unsafe { UPSafeCell::new(TaskManager::new()) };
}; pub static ref PID2PCB: UPSafeCell<BTreeMap<usize, Arc<ProcessControlBlock>>> =
unsafe { UPSafeCell::new(BTreeMap::new()) };
} }
pub fn add_task(task: Arc<TaskControlBlock>) { pub fn add_task(task: Arc<TaskControlBlock>) {
@ -34,3 +37,19 @@ pub fn add_task(task: Arc<TaskControlBlock>) {
pub fn fetch_task() -> Option<Arc<TaskControlBlock>> { pub fn fetch_task() -> Option<Arc<TaskControlBlock>> {
TASK_MANAGER.exclusive_access().fetch() TASK_MANAGER.exclusive_access().fetch()
} }
pub fn pid2process(pid: usize) -> Option<Arc<ProcessControlBlock>> {
let map = PID2PCB.exclusive_access();
map.get(&pid).map(Arc::clone)
}
pub fn insert_into_pid2process(pid: usize, process: Arc<ProcessControlBlock>) {
PID2PCB.exclusive_access().insert(pid, process);
}
pub fn remove_from_pid2process(pid: usize) {
let mut map = PID2PCB.exclusive_access();
if map.remove(&pid).is_none() {
panic!("cannot find pid {} in pid2task!", pid);
}
}

View file

@ -1,38 +1,29 @@
mod context; mod context;
mod switch;
mod task;
mod manager;
mod processor;
mod id; mod id;
mod manager;
mod process; mod process;
mod processor;
mod signal;
mod switch;
#[allow(clippy::module_inception)]
mod task;
use crate::fs::{open_file, OpenFlags}; use crate::fs::{open_file, OpenFlags};
use switch::__switch;
use alloc::sync::Arc; use alloc::sync::Arc;
use manager::fetch_task;
use lazy_static::*; use lazy_static::*;
use manager::fetch_task;
use process::ProcessControlBlock; use process::ProcessControlBlock;
use switch::__switch;
pub use context::TaskContext; pub use context::TaskContext;
pub use id::{kstack_alloc, pid_alloc, KernelStack, PidHandle};
pub use manager::{add_task, pid2process, remove_from_pid2process};
pub use processor::{ pub use processor::{
run_tasks, current_kstack_top, current_process, current_task, current_trap_cx, current_trap_cx_user_va,
current_task, current_user_token, run_tasks, schedule, take_current_task,
current_process,
current_user_token,
current_trap_cx_user_va,
current_trap_cx,
current_kstack_top,
take_current_task,
schedule,
}; };
pub use signal::SignalFlags;
pub use task::{TaskControlBlock, TaskStatus}; pub use task::{TaskControlBlock, TaskStatus};
pub use manager::add_task;
pub use id::{
PidHandle,
pid_alloc,
KernelStack,
kstack_alloc,
};
pub fn suspend_current_and_run_next() { pub fn suspend_current_and_run_next() {
// There must be an application running. // There must be an application running.
@ -76,6 +67,7 @@ pub fn exit_current_and_run_next(exit_code: i32) {
// however, if this is the main thread of current process // however, if this is the main thread of current process
// the process should terminate at once // the process should terminate at once
if tid == 0 { if tid == 0 {
remove_from_pid2process(process.getpid());
let mut process_inner = process.inner_exclusive_access(); let mut process_inner = process.inner_exclusive_access();
// mark this process as a zombie process // mark this process as a zombie process
process_inner.is_zombie = true; process_inner.is_zombie = true;
@ -103,6 +95,8 @@ pub fn exit_current_and_run_next(exit_code: i32) {
process_inner.children.clear(); process_inner.children.clear();
// deallocate other data in user space i.e. program code/data section // deallocate other data in user space i.e. program code/data section
process_inner.memory_set.recycle_data_pages(); process_inner.memory_set.recycle_data_pages();
// drop file descriptors
process_inner.fd_table.clear();
} }
drop(process); drop(process);
// we do not have to save task context // we do not have to save task context
@ -121,3 +115,15 @@ lazy_static! {
pub fn add_initproc() { pub fn add_initproc() {
let _initproc = INITPROC.clone(); let _initproc = INITPROC.clone();
} }
pub fn check_signals_of_current() -> Option<(i32, &'static str)> {
let process = current_process();
let process_inner = process.inner_exclusive_access();
process_inner.signals.check_error()
}
pub fn current_add_signal(signal: SignalFlags) {
let process = current_process();
let mut process_inner = process.inner_exclusive_access();
process_inner.signals |= signal;
}

View file

@ -1,20 +1,17 @@
use crate::mm::{
MemorySet,
KERNEL_SPACE,
translated_refmut,
};
use crate::trap::{TrapContext, trap_handler};
use crate::sync::{UPSafeCell, Mutex, Semaphore};
use core::cell::RefMut;
use super::id::RecycleAllocator; use super::id::RecycleAllocator;
use super::manager::insert_into_pid2process;
use super::TaskControlBlock; use super::TaskControlBlock;
use super::{PidHandle, pid_alloc}; use super::{add_task, SignalFlags};
use super::add_task; use super::{pid_alloc, PidHandle};
use alloc::sync::{Weak, Arc}; use crate::fs::{File, Stdin, Stdout};
use crate::mm::{translated_refmut, MemorySet, KERNEL_SPACE};
use crate::sync::{Condvar, Mutex, Semaphore, UPSafeCell};
use crate::trap::{trap_handler, TrapContext};
use alloc::string::String;
use alloc::sync::{Arc, Weak};
use alloc::vec; use alloc::vec;
use alloc::vec::Vec; use alloc::vec::Vec;
use alloc::string::String; use core::cell::RefMut;
use crate::fs::{File, Stdin, Stdout};
pub struct ProcessControlBlock { pub struct ProcessControlBlock {
// immutable // immutable
@ -30,10 +27,12 @@ pub struct ProcessControlBlockInner {
pub children: Vec<Arc<ProcessControlBlock>>, pub children: Vec<Arc<ProcessControlBlock>>,
pub exit_code: i32, pub exit_code: i32,
pub fd_table: Vec<Option<Arc<dyn File + Send + Sync>>>, pub fd_table: Vec<Option<Arc<dyn File + Send + Sync>>>,
pub signals: SignalFlags,
pub tasks: Vec<Option<Arc<TaskControlBlock>>>, pub tasks: Vec<Option<Arc<TaskControlBlock>>>,
pub task_res_allocator: RecycleAllocator, pub task_res_allocator: RecycleAllocator,
pub mutex_list: Vec<Option<Arc<dyn Mutex>>>, pub mutex_list: Vec<Option<Arc<dyn Mutex>>>,
pub semaphore_list: Vec<Option<Arc<Semaphore>>>, pub semaphore_list: Vec<Option<Arc<Semaphore>>>,
pub condvar_list: Vec<Option<Arc<Condvar>>>,
} }
impl ProcessControlBlockInner { impl ProcessControlBlockInner {
@ -43,8 +42,7 @@ impl ProcessControlBlockInner {
} }
pub fn alloc_fd(&mut self) -> usize { pub fn alloc_fd(&mut self) -> usize {
if let Some(fd) = (0..self.fd_table.len()) if let Some(fd) = (0..self.fd_table.len()).find(|fd| self.fd_table[*fd].is_none()) {
.find(|fd| self.fd_table[*fd].is_none()) {
fd fd
} else { } else {
self.fd_table.push(None); self.fd_table.push(None);
@ -56,7 +54,7 @@ impl ProcessControlBlockInner {
self.task_res_allocator.alloc() self.task_res_allocator.alloc()
} }
pub fn dealloc_tid(&mut self, tid: usize){ pub fn dealloc_tid(&mut self, tid: usize) {
self.task_res_allocator.dealloc(tid) self.task_res_allocator.dealloc(tid)
} }
@ -81,25 +79,29 @@ impl ProcessControlBlock {
let pid_handle = pid_alloc(); let pid_handle = pid_alloc();
let process = Arc::new(Self { let process = Arc::new(Self {
pid: pid_handle, pid: pid_handle,
inner: unsafe { UPSafeCell::new(ProcessControlBlockInner { inner: unsafe {
is_zombie: false, UPSafeCell::new(ProcessControlBlockInner {
memory_set, is_zombie: false,
parent: None, memory_set,
children: Vec::new(), parent: None,
exit_code: 0, children: Vec::new(),
fd_table: vec![ exit_code: 0,
// 0 -> stdin fd_table: vec![
Some(Arc::new(Stdin)), // 0 -> stdin
// 1 -> stdout Some(Arc::new(Stdin)),
Some(Arc::new(Stdout)), // 1 -> stdout
// 2 -> stderr Some(Arc::new(Stdout)),
Some(Arc::new(Stdout)), // 2 -> stderr
], Some(Arc::new(Stdout)),
tasks: Vec::new(), ],
task_res_allocator: RecycleAllocator::new(), signals: SignalFlags::empty(),
mutex_list: Vec::new(), tasks: Vec::new(),
semaphore_list: Vec::new(), task_res_allocator: RecycleAllocator::new(),
})} mutex_list: Vec::new(),
semaphore_list: Vec::new(),
condvar_list: Vec::new(),
})
},
}); });
// create a main thread, we should allocate ustack and trap_cx here // create a main thread, we should allocate ustack and trap_cx here
let task = Arc::new(TaskControlBlock::new( let task = Arc::new(TaskControlBlock::new(
@ -124,6 +126,7 @@ impl ProcessControlBlock {
let mut process_inner = process.inner_exclusive_access(); let mut process_inner = process.inner_exclusive_access();
process_inner.tasks.push(Some(Arc::clone(&task))); process_inner.tasks.push(Some(Arc::clone(&task)));
drop(process_inner); drop(process_inner);
insert_into_pid2process(process.getpid(), Arc::clone(&process));
// add main thread to scheduler // add main thread to scheduler
add_task(task); add_task(task);
process process
@ -152,7 +155,7 @@ impl ProcessControlBlock {
.map(|arg| { .map(|arg| {
translated_refmut( translated_refmut(
new_token, new_token,
(argv_base + arg * core::mem::size_of::<usize>()) as *mut usize (argv_base + arg * core::mem::size_of::<usize>()) as *mut usize,
) )
}) })
.collect(); .collect();
@ -202,25 +205,35 @@ impl ProcessControlBlock {
// create child process pcb // create child process pcb
let child = Arc::new(Self { let child = Arc::new(Self {
pid, pid,
inner: unsafe { UPSafeCell::new(ProcessControlBlockInner { inner: unsafe {
is_zombie: false, UPSafeCell::new(ProcessControlBlockInner {
memory_set, is_zombie: false,
parent: Some(Arc::downgrade(self)), memory_set,
children: Vec::new(), parent: Some(Arc::downgrade(self)),
exit_code: 0, children: Vec::new(),
fd_table: new_fd_table, exit_code: 0,
tasks: Vec::new(), fd_table: new_fd_table,
task_res_allocator: RecycleAllocator::new(), signals: SignalFlags::empty(),
mutex_list: Vec::new(), tasks: Vec::new(),
semaphore_list: Vec::new(), task_res_allocator: RecycleAllocator::new(),
})} mutex_list: Vec::new(),
semaphore_list: Vec::new(),
condvar_list: Vec::new(),
})
},
}); });
// add child // add child
parent.children.push(Arc::clone(&child)); parent.children.push(Arc::clone(&child));
// create main thread of child process // create main thread of child process
let task = Arc::new(TaskControlBlock::new( let task = Arc::new(TaskControlBlock::new(
Arc::clone(&child), Arc::clone(&child),
parent.get_task(0).inner_exclusive_access().res.as_ref().unwrap().ustack_base(), parent
.get_task(0)
.inner_exclusive_access()
.res
.as_ref()
.unwrap()
.ustack_base(),
// here we do not allocate trap_cx or ustack again // here we do not allocate trap_cx or ustack again
// but mention that we allocate a new kstack here // but mention that we allocate a new kstack here
false, false,
@ -234,6 +247,7 @@ impl ProcessControlBlock {
let trap_cx = task_inner.get_trap_cx(); let trap_cx = task_inner.get_trap_cx();
trap_cx.kernel_sp = task.kstack.get_top(); trap_cx.kernel_sp = task.kstack.get_top();
drop(task_inner); drop(task_inner);
insert_into_pid2process(child.getpid(), Arc::clone(&child));
// add this thread to scheduler // add this thread to scheduler
add_task(task); add_task(task);
child child
@ -243,4 +257,3 @@ impl ProcessControlBlock {
self.pid.0 self.pid.0
} }
} }

View file

@ -1,10 +1,10 @@
use super::{TaskContext, TaskControlBlock, ProcessControlBlock}; use super::__switch;
use super::{fetch_task, TaskStatus};
use super::{ProcessControlBlock, TaskContext, TaskControlBlock};
use crate::sync::UPSafeCell;
use crate::trap::TrapContext;
use alloc::sync::Arc; use alloc::sync::Arc;
use lazy_static::*; use lazy_static::*;
use super::{fetch_task, TaskStatus};
use super::__switch;
use crate::trap::TrapContext;
use crate::sync::UPSafeCell;
pub struct Processor { pub struct Processor {
current: Option<Arc<TaskControlBlock>>, current: Option<Arc<TaskControlBlock>>,
@ -25,14 +25,12 @@ impl Processor {
self.current.take() self.current.take()
} }
pub fn current(&self) -> Option<Arc<TaskControlBlock>> { pub fn current(&self) -> Option<Arc<TaskControlBlock>> {
self.current.as_ref().map(|task| Arc::clone(task)) self.current.as_ref().map(Arc::clone)
} }
} }
lazy_static! { lazy_static! {
pub static ref PROCESSOR: UPSafeCell<Processor> = unsafe { pub static ref PROCESSOR: UPSafeCell<Processor> = unsafe { UPSafeCell::new(Processor::new()) };
UPSafeCell::new(Processor::new())
};
} }
pub fn run_tasks() { pub fn run_tasks() {
@ -50,10 +48,7 @@ pub fn run_tasks() {
// release processor manually // release processor manually
drop(processor); drop(processor);
unsafe { unsafe {
__switch( __switch(idle_task_cx_ptr, next_task_cx_ptr);
idle_task_cx_ptr,
next_task_cx_ptr,
);
} }
} else { } else {
println!("no tasks available in run_tasks"); println!("no tasks available in run_tasks");
@ -75,12 +70,14 @@ pub fn current_process() -> Arc<ProcessControlBlock> {
pub fn current_user_token() -> usize { pub fn current_user_token() -> usize {
let task = current_task().unwrap(); let task = current_task().unwrap();
let token = task.get_user_token(); task.get_user_token()
token
} }
pub fn current_trap_cx() -> &'static mut TrapContext { pub fn current_trap_cx() -> &'static mut TrapContext {
current_task().unwrap().inner_exclusive_access().get_trap_cx() current_task()
.unwrap()
.inner_exclusive_access()
.get_trap_cx()
} }
pub fn current_trap_cx_user_va() -> usize { pub fn current_trap_cx_user_va() -> usize {
@ -94,10 +91,7 @@ pub fn current_trap_cx_user_va() -> usize {
} }
pub fn current_kstack_top() -> usize { pub fn current_kstack_top() -> usize {
current_task() current_task().unwrap().kstack.get_top()
.unwrap()
.kstack
.get_top()
} }
pub fn schedule(switched_task_cx_ptr: *mut TaskContext) { pub fn schedule(switched_task_cx_ptr: *mut TaskContext) {
@ -105,9 +99,6 @@ pub fn schedule(switched_task_cx_ptr: *mut TaskContext) {
let idle_task_cx_ptr = processor.get_idle_task_cx_ptr(); let idle_task_cx_ptr = processor.get_idle_task_cx_ptr();
drop(processor); drop(processor);
unsafe { unsafe {
__switch( __switch(switched_task_cx_ptr, idle_task_cx_ptr);
switched_task_cx_ptr,
idle_task_cx_ptr,
);
} }
} }

29
os/src/task/signal.rs Normal file
View file

@ -0,0 +1,29 @@
use bitflags::*;
bitflags! {
pub struct SignalFlags: u32 {
const SIGINT = 1 << 2;
const SIGILL = 1 << 4;
const SIGABRT = 1 << 6;
const SIGFPE = 1 << 8;
const SIGSEGV = 1 << 11;
}
}
impl SignalFlags {
pub fn check_error(&self) -> Option<(i32, &'static str)> {
if self.contains(Self::SIGINT) {
Some((-2, "Killed, SIGINT=2"))
} else if self.contains(Self::SIGILL) {
Some((-4, "Illegal Instruction, SIGILL=4"))
} else if self.contains(Self::SIGABRT) {
Some((-6, "Aborted, SIGABRT=6"))
} else if self.contains(Self::SIGFPE) {
Some((-8, "Erroneous Arithmetic Operation, SIGFPE=8"))
} else if self.contains(Self::SIGSEGV) {
Some((-11, "Segmentation Fault, SIGSEGV=11"))
} else {
None
}
}
}

View file

@ -1,12 +1,8 @@
use super::TaskContext;
use core::arch::global_asm; use core::arch::global_asm;
global_asm!(include_str!("switch.S")); global_asm!(include_str!("switch.S"));
use super::TaskContext;
extern "C" { extern "C" {
pub fn __switch( pub fn __switch(current_task_cx_ptr: *mut TaskContext, next_task_cx_ptr: *const TaskContext);
current_task_cx_ptr: *mut TaskContext,
next_task_cx_ptr: *const TaskContext
);
} }

View file

@ -1,8 +1,8 @@
use alloc::sync::{Arc, Weak};
use crate::{mm::PhysPageNum, sync::UPSafeCell};
use crate::trap::TrapContext;
use super::id::TaskUserRes; use super::id::TaskUserRes;
use super::{KernelStack, ProcessControlBlock, TaskContext, kstack_alloc}; use super::{kstack_alloc, KernelStack, ProcessControlBlock, TaskContext};
use crate::trap::TrapContext;
use crate::{mm::PhysPageNum, sync::UPSafeCell};
use alloc::sync::{Arc, Weak};
use core::cell::RefMut; use core::cell::RefMut;
pub struct TaskControlBlock { pub struct TaskControlBlock {
@ -48,7 +48,7 @@ impl TaskControlBlock {
pub fn new( pub fn new(
process: Arc<ProcessControlBlock>, process: Arc<ProcessControlBlock>,
ustack_base: usize, ustack_base: usize,
alloc_user_res: bool alloc_user_res: bool,
) -> Self { ) -> Self {
let res = TaskUserRes::new(Arc::clone(&process), ustack_base, alloc_user_res); let res = TaskUserRes::new(Arc::clone(&process), ustack_base, alloc_user_res);
let trap_cx_ppn = res.trap_cx_ppn(); let trap_cx_ppn = res.trap_cx_ppn();
@ -57,15 +57,15 @@ impl TaskControlBlock {
Self { Self {
process: Arc::downgrade(&process), process: Arc::downgrade(&process),
kstack, kstack,
inner: unsafe { UPSafeCell::new( inner: unsafe {
TaskControlBlockInner { UPSafeCell::new(TaskControlBlockInner {
res: Some(res), res: Some(res),
trap_cx_ppn, trap_cx_ppn,
task_cx: TaskContext::goto_trap_return(kstack_top), task_cx: TaskContext::goto_trap_return(kstack_top),
task_status: TaskStatus::Ready, task_status: TaskStatus::Ready,
exit_code: None, exit_code: None,
} })
)}, },
} }
} }
} }

View file

@ -1,13 +1,13 @@
use core::cmp::Ordering; use core::cmp::Ordering;
use riscv::register::time;
use crate::sbi::set_timer;
use crate::config::CLOCK_FREQ; use crate::config::CLOCK_FREQ;
use crate::task::{TaskControlBlock, add_task}; use crate::sbi::set_timer;
use crate::sync::UPSafeCell; use crate::sync::UPSafeCell;
use crate::task::{add_task, TaskControlBlock};
use alloc::collections::BinaryHeap; use alloc::collections::BinaryHeap;
use alloc::sync::Arc; use alloc::sync::Arc;
use lazy_static::*; use lazy_static::*;
use riscv::register::time;
const TICKS_PER_SEC: usize = 100; const TICKS_PER_SEC: usize = 100;
const MSEC_PER_SEC: usize = 1000; const MSEC_PER_SEC: usize = 1000;
@ -50,17 +50,13 @@ impl Ord for TimerCondVar {
} }
lazy_static! { lazy_static! {
static ref TIMERS: UPSafeCell<BinaryHeap<TimerCondVar>> = unsafe { UPSafeCell::new( static ref TIMERS: UPSafeCell<BinaryHeap<TimerCondVar>> =
BinaryHeap::<TimerCondVar>::new() unsafe { UPSafeCell::new(BinaryHeap::<TimerCondVar>::new()) };
)};
} }
pub fn add_timer(expire_ms: usize, task: Arc<TaskControlBlock>) { pub fn add_timer(expire_ms: usize, task: Arc<TaskControlBlock>) {
let mut timers = TIMERS.exclusive_access(); let mut timers = TIMERS.exclusive_access();
timers.push(TimerCondVar { timers.push(TimerCondVar { expire_ms, task });
expire_ms,
task,
});
} }
pub fn check_timer() { pub fn check_timer() {
@ -69,8 +65,9 @@ pub fn check_timer() {
while let Some(timer) = timers.peek() { while let Some(timer) = timers.peek() {
if timer.expire_ms <= current_ms { if timer.expire_ms <= current_ms {
add_task(Arc::clone(&timer.task)); add_task(Arc::clone(&timer.task));
drop(timer);
timers.pop(); timers.pop();
} else { break; } } else {
break;
}
} }
} }

View file

@ -1,4 +1,4 @@
use riscv::register::sstatus::{Sstatus, self, SPP}; use riscv::register::sstatus::{self, Sstatus, SPP};
#[repr(C)] #[repr(C)]
#[derive(Debug)] #[derive(Debug)]
@ -12,7 +12,9 @@ pub struct TrapContext {
} }
impl TrapContext { impl TrapContext {
pub fn set_sp(&mut self, sp: usize) { self.x[2] = sp; } pub fn set_sp(&mut self, sp: usize) {
self.x[2] = sp;
}
pub fn app_init_context( pub fn app_init_context(
entry: usize, entry: usize,
sp: usize, sp: usize,

View file

@ -1,28 +1,18 @@
mod context; mod context;
use crate::config::TRAMPOLINE;
use crate::syscall::syscall;
use crate::task::{
check_signals_of_current, current_add_signal, current_trap_cx, current_trap_cx_user_va,
current_user_token, exit_current_and_run_next, suspend_current_and_run_next, SignalFlags,
};
use crate::timer::{check_timer, set_next_trigger};
use core::arch::{asm, global_asm}; use core::arch::{asm, global_asm};
use riscv::register::{ use riscv::register::{
mtvec::TrapMode, mtvec::TrapMode,
stvec, scause::{self, Exception, Interrupt, Trap},
scause::{ sie, stval, stvec,
self,
Trap,
Exception,
Interrupt,
},
stval,
sie,
}; };
use crate::syscall::syscall;
use crate::task::{
exit_current_and_run_next,
suspend_current_and_run_next,
current_user_token,
current_trap_cx,
current_trap_cx_user_va,
};
use crate::timer::{set_next_trigger, check_timer};
use crate::config::TRAMPOLINE;
global_asm!(include_str!("trap.S")); global_asm!(include_str!("trap.S"));
@ -43,7 +33,9 @@ fn set_user_trap_entry() {
} }
pub fn enable_timer_interrupt() { pub fn enable_timer_interrupt() {
unsafe { sie::set_stimer(); } unsafe {
sie::set_stimer();
}
} }
#[no_mangle] #[no_mangle]
@ -62,25 +54,24 @@ pub fn trap_handler() -> ! {
cx = current_trap_cx(); cx = current_trap_cx();
cx.x[10] = result as usize; cx.x[10] = result as usize;
} }
Trap::Exception(Exception::StoreFault) | Trap::Exception(Exception::StoreFault)
Trap::Exception(Exception::StorePageFault) | | Trap::Exception(Exception::StorePageFault)
Trap::Exception(Exception::InstructionFault) | | Trap::Exception(Exception::InstructionFault)
Trap::Exception(Exception::InstructionPageFault) | | Trap::Exception(Exception::InstructionPageFault)
Trap::Exception(Exception::LoadFault) | | Trap::Exception(Exception::LoadFault)
Trap::Exception(Exception::LoadPageFault) => { | Trap::Exception(Exception::LoadPageFault) => {
/*
println!( println!(
"[kernel] {:?} in application, bad addr = {:#x}, bad instruction = {:#x}, core dumped.", "[kernel] {:?} in application, bad addr = {:#x}, bad instruction = {:#x}, kernel killed it.",
scause.cause(), scause.cause(),
stval, stval,
current_trap_cx().sepc, current_trap_cx().sepc,
); );
// page fault exit code */
exit_current_and_run_next(-2); current_add_signal(SignalFlags::SIGSEGV);
} }
Trap::Exception(Exception::IllegalInstruction) => { Trap::Exception(Exception::IllegalInstruction) => {
println!("[kernel] IllegalInstruction in application, core dumped."); current_add_signal(SignalFlags::SIGILL);
// illegal instruction exit code
exit_current_and_run_next(-3);
} }
Trap::Interrupt(Interrupt::SupervisorTimer) => { Trap::Interrupt(Interrupt::SupervisorTimer) => {
set_next_trigger(); set_next_trigger();
@ -88,10 +79,18 @@ pub fn trap_handler() -> ! {
suspend_current_and_run_next(); suspend_current_and_run_next();
} }
_ => { _ => {
panic!("Unsupported trap {:?}, stval = {:#x}!", scause.cause(), stval); panic!(
"Unsupported trap {:?}, stval = {:#x}!",
scause.cause(),
stval
);
} }
} }
//println!("before trap_return"); // check signals
if let Some((errno, msg)) = check_signals_of_current() {
println!("[kernel] {}", msg);
exit_current_and_run_next(errno);
}
trap_return(); trap_return();
} }

View file

@ -9,3 +9,4 @@ edition = "2018"
[dependencies] [dependencies]
buddy_system_allocator = "0.6" buddy_system_allocator = "0.6"
bitflags = "1.2.1" bitflags = "1.2.1"
riscv = { git = "https://github.com/rcore-os/riscv", features = ["inline-asm"] }

View file

@ -5,13 +5,7 @@
extern crate user_lib; extern crate user_lib;
extern crate alloc; extern crate alloc;
use user_lib::{ use user_lib::{close, open, read, OpenFlags};
open,
OpenFlags,
close,
read,
};
use alloc::string::String;
#[no_mangle] #[no_mangle]
pub fn main(argc: usize, argv: &[&str]) -> i32 { pub fn main(argc: usize, argv: &[&str]) -> i32 {
@ -21,14 +15,14 @@ pub fn main(argc: usize, argv: &[&str]) -> i32 {
panic!("Error occured when opening file"); panic!("Error occured when opening file");
} }
let fd = fd as usize; let fd = fd as usize;
let mut buf = [0u8; 16]; let mut buf = [0u8; 256];
let mut s = String::new();
loop { loop {
let size = read(fd, &mut buf) as usize; let size = read(fd, &mut buf) as usize;
if size == 0 { break; } if size == 0 {
s.push_str(core::str::from_utf8(&buf[..size]).unwrap()); break;
}
print!("{}", core::str::from_utf8(&buf[..size]).unwrap());
} }
println!("{}", s);
close(fd); close(fd);
0 0
} }

View file

@ -9,8 +9,8 @@ extern crate user_lib;
#[no_mangle] #[no_mangle]
pub fn main(argc: usize, argv: &[&str]) -> i32 { pub fn main(argc: usize, argv: &[&str]) -> i32 {
println!("argc = {}", argc); println!("argc = {}", argc);
for i in 0..argc { for (i, arg) in argv.iter().enumerate() {
println!("argv[{}] = {}", i, argv[i]); println!("argv[{}] = {}", i, arg);
} }
0 0
} }

View file

@ -0,0 +1,30 @@
#![no_std]
#![no_main]
#[macro_use]
extern crate user_lib;
use user_lib::read;
#[no_mangle]
pub fn main(_argc: usize, _argv: &[&str]) -> i32 {
let mut buf = [0u8; 256];
let mut lines = 0usize;
let mut total_size = 0usize;
loop {
let len = read(0, &mut buf) as usize;
if len == 0 {
break;
}
total_size += len;
let string = core::str::from_utf8(&buf[..len]).unwrap();
lines += string
.chars()
.fold(0, |acc, c| acc + if c == '\n' { 1 } else { 0 });
}
if total_size > 0 {
lines += 1;
}
println!("{}", lines);
0
}

View file

@ -3,7 +3,7 @@
#[macro_use] #[macro_use]
extern crate user_lib; extern crate user_lib;
use user_lib::{fork, yield_, waitpid, exit, wait}; use user_lib::{exit, fork, wait, waitpid, yield_};
const MAGIC: i32 = -0x10384; const MAGIC: i32 = -0x10384;
@ -13,7 +13,9 @@ pub fn main() -> i32 {
let pid = fork(); let pid = fork();
if pid == 0 { if pid == 0 {
println!("I am the child."); println!("I am the child.");
for _ in 0..7 { yield_(); } for _ in 0..7 {
yield_();
}
exit(MAGIC); exit(MAGIC);
} else { } else {
println!("I am parent, fork a child pid {}", pid); println!("I am parent, fork a child pid {}", pid);
@ -26,4 +28,3 @@ pub fn main() -> i32 {
println!("exit pass."); println!("exit pass.");
0 0
} }

View file

@ -4,13 +4,7 @@
#[macro_use] #[macro_use]
extern crate user_lib; extern crate user_lib;
use user_lib::{ use user_lib::{close, open, read, write, OpenFlags};
open,
close,
read,
write,
OpenFlags,
};
#[no_mangle] #[no_mangle]
pub fn main() -> i32 { pub fn main() -> i32 {
@ -29,10 +23,7 @@ pub fn main() -> i32 {
let read_len = read(fd, &mut buffer) as usize; let read_len = read(fd, &mut buffer) as usize;
close(fd); close(fd);
assert_eq!( assert_eq!(test_str, core::str::from_utf8(&buffer[..read_len]).unwrap(),);
test_str,
core::str::from_utf8(&buffer[..read_len]).unwrap(),
);
println!("file_test passed!"); println!("file_test passed!");
0 0
} }

View file

@ -4,9 +4,9 @@
#[macro_use] #[macro_use]
extern crate user_lib; extern crate user_lib;
use user_lib::{fork, wait, exit}; use user_lib::{exit, fork, wait};
const MAX_CHILD: usize = 40; const MAX_CHILD: usize = 30;
#[no_mangle] #[no_mangle]
pub fn main() -> i32 { pub fn main() -> i32 {

View file

@ -4,7 +4,7 @@
#[macro_use] #[macro_use]
extern crate user_lib; extern crate user_lib;
use user_lib::{fork, wait, getpid, exit, sleep, get_time}; use user_lib::{exit, fork, get_time, getpid, sleep, wait};
static NUM: usize = 30; static NUM: usize = 30;
@ -14,7 +14,8 @@ pub fn main() -> i32 {
let pid = fork(); let pid = fork();
if pid == 0 { if pid == 0 {
let current_time = get_time(); let current_time = get_time();
let sleep_length = (current_time as i32 as isize) * (current_time as i32 as isize) % 1000 + 1000; let sleep_length =
(current_time as i32 as isize) * (current_time as i32 as isize) % 1000 + 1000;
println!("pid {} sleep for {} ms", getpid(), sleep_length); println!("pid {} sleep for {} ms", getpid(), sleep_length);
sleep(sleep_length as usize); sleep(sleep_length as usize);
println!("pid {} OK!", getpid()); println!("pid {} OK!", getpid());

View file

@ -4,7 +4,7 @@
#[macro_use] #[macro_use]
extern crate user_lib; extern crate user_lib;
use user_lib::{sleep, getpid, fork, exit, yield_}; use user_lib::{exit, fork, getpid, sleep, yield_};
const DEPTH: usize = 4; const DEPTH: usize = 4;

View file

@ -4,33 +4,30 @@
#[macro_use] #[macro_use]
extern crate user_lib; extern crate user_lib;
use user_lib::{ use user_lib::{close, get_time, open, write, OpenFlags};
OpenFlags,
open,
close,
write,
get_time,
};
#[no_mangle] #[no_mangle]
pub fn main() -> i32 { pub fn main() -> i32 {
let mut buffer = [0u8; 1024]; // 1KiB let mut buffer = [0u8; 1024]; // 1KiB
for i in 0..buffer.len() { for (i, ch) in buffer.iter_mut().enumerate() {
buffer[i] = i as u8; *ch = i as u8;
} }
let f = open("testf", OpenFlags::CREATE | OpenFlags::WRONLY); let f = open("testf\0", OpenFlags::CREATE | OpenFlags::WRONLY);
if f < 0 { if f < 0 {
panic!("Open test file failed!"); panic!("Open test file failed!");
} }
let f = f as usize; let f = f as usize;
let start = get_time(); let start = get_time();
let size_mb = 1usize; let size_mb = 1usize;
for _ in 0..1024*size_mb { for _ in 0..1024 * size_mb {
write(f, &buffer); write(f, &buffer);
} }
close(f); close(f);
let time_ms = (get_time() - start) as usize; let time_ms = (get_time() - start) as usize;
let speed_kbs = size_mb * 1000000 / time_ms; let speed_kbs = size_mb * 1000000 / time_ms;
println!("{}MiB written, time cost = {}ms, write speed = {}KiB/s", size_mb, time_ms, speed_kbs); println!(
"{}MiB written, time cost = {}ms, write speed = {}KiB/s",
size_mb, time_ms, speed_kbs
);
0 0
} }

10
user/src/bin/infloop.rs Normal file
View file

@ -0,0 +1,10 @@
#![no_std]
#![no_main]
#![allow(clippy::empty_loop)]
extern crate user_lib;
#[no_mangle]
pub fn main(_argc: usize, _argv: &[&str]) -> ! {
loop {}
}

View file

@ -1,20 +1,14 @@
#![no_std] #![no_std]
#![no_main] #![no_main]
#[macro_use]
extern crate user_lib; extern crate user_lib;
use user_lib::{ use user_lib::{exec, fork, wait, yield_};
fork,
wait,
exec,
yield_,
};
#[no_mangle] #[no_mangle]
fn main() -> i32 { fn main() -> i32 {
if fork() == 0 { if fork() == 0 {
exec("user_shell\0", &[0 as *const u8]); exec("user_shell\0", &[core::ptr::null::<u8>()]);
} else { } else {
loop { loop {
let mut exit_code: i32 = 0; let mut exit_code: i32 = 0;
@ -23,11 +17,13 @@ fn main() -> i32 {
yield_(); yield_();
continue; continue;
} }
/*
println!( println!(
"[initproc] Released a zombie process, pid={}, exit_code={}", "[initproc] Released a zombie process, pid={}, exit_code={}",
pid, pid,
exit_code, exit_code,
); );
*/
} }
} }
0 0

View file

@ -1,12 +1,13 @@
#![no_std] #![no_std]
#![no_main] #![no_main]
#![allow(clippy::needless_range_loop)]
#[macro_use] #[macro_use]
extern crate user_lib; extern crate user_lib;
use user_lib::{fork, wait, yield_, exit, getpid, get_time}; use user_lib::{exit, fork, get_time, getpid, wait, yield_};
static NUM: usize = 35; static NUM: usize = 30;
const N: usize = 10; const N: usize = 10;
static P: i32 = 10007; static P: i32 = 10007;
type Arr = [[i32; N]; N]; type Arr = [[i32; N]; N];

View file

@ -1,15 +1,16 @@
#![no_std] #![no_std]
#![no_main] #![no_main]
#![allow(clippy::println_empty_string)]
#[macro_use] #[macro_use]
extern crate user_lib; extern crate user_lib;
extern crate alloc; extern crate alloc;
use user_lib::{semaphore_create, semaphore_up, semaphore_down};
use user_lib::{thread_create, waittid};
use user_lib::exit;
use alloc::vec::Vec; use alloc::vec::Vec;
use user_lib::exit;
use user_lib::{semaphore_create, semaphore_down, semaphore_up};
use user_lib::{thread_create, waittid};
const SEM_MUTEX: usize = 0; const SEM_MUTEX: usize = 0;
const SEM_EMPTY: usize = 1; const SEM_EMPTY: usize = 1;
@ -57,7 +58,10 @@ pub fn main() -> i32 {
let ids: Vec<_> = (0..PRODUCER_COUNT).collect(); let ids: Vec<_> = (0..PRODUCER_COUNT).collect();
let mut threads = Vec::new(); let mut threads = Vec::new();
for i in 0..PRODUCER_COUNT { for i in 0..PRODUCER_COUNT {
threads.push(thread_create(producer as usize, &ids.as_slice()[i] as *const _ as usize)); threads.push(thread_create(
producer as usize,
&ids.as_slice()[i] as *const _ as usize,
));
} }
threads.push(thread_create(consumer as usize, 0)); threads.push(thread_create(consumer as usize, 0));
// wait for all threads to complete // wait for all threads to complete

View file

@ -1,14 +1,15 @@
#![no_std] #![no_std]
#![no_main] #![no_main]
#![allow(clippy::println_empty_string)]
#[macro_use] #[macro_use]
extern crate user_lib; extern crate user_lib;
extern crate alloc; extern crate alloc;
use alloc::vec::Vec;
use user_lib::{exit, get_time, sleep};
use user_lib::{mutex_blocking_create, mutex_lock, mutex_unlock}; use user_lib::{mutex_blocking_create, mutex_lock, mutex_unlock};
use user_lib::{thread_create, waittid}; use user_lib::{thread_create, waittid};
use user_lib::{sleep, exit, get_time};
use alloc::vec::Vec;
const N: usize = 5; const N: usize = 5;
const ROUND: usize = 4; const ROUND: usize = 4;
@ -38,16 +39,24 @@ fn philosopher_dining_problem(id: *const usize) {
let max = left + right - min; let max = left + right - min;
for round in 0..ROUND { for round in 0..ROUND {
// thinking // thinking
unsafe { THINK[id][2 * round] = get_time_u(); } unsafe {
THINK[id][2 * round] = get_time_u();
}
sleep(ARR[id][2 * round]); sleep(ARR[id][2 * round]);
unsafe { THINK[id][2 * round + 1] = get_time_u(); } unsafe {
THINK[id][2 * round + 1] = get_time_u();
}
// wait for forks // wait for forks
mutex_lock(min); mutex_lock(min);
mutex_lock(max); mutex_lock(max);
// eating // eating
unsafe { EAT[id][2 * round] = get_time_u(); } unsafe {
EAT[id][2 * round] = get_time_u();
}
sleep(ARR[id][2 * round + 1]); sleep(ARR[id][2 * round + 1]);
unsafe { EAT[id][2 * round + 1] = get_time_u(); } unsafe {
EAT[id][2 * round + 1] = get_time_u();
}
mutex_unlock(max); mutex_unlock(max);
mutex_unlock(min); mutex_unlock(min);
} }
@ -61,7 +70,10 @@ pub fn main() -> i32 {
let start = get_time_u(); let start = get_time_u();
for i in 0..N { for i in 0..N {
assert_eq!(mutex_blocking_create(), i as isize); assert_eq!(mutex_blocking_create(), i as isize);
v.push(thread_create(philosopher_dining_problem as usize, &ids.as_slice()[i] as *const _ as usize)); v.push(thread_create(
philosopher_dining_problem as usize,
&ids.as_slice()[i] as *const _ as usize,
));
} }
for tid in v.iter() { for tid in v.iter() {
waittid(*tid as usize); waittid(*tid as usize);
@ -71,19 +83,19 @@ pub fn main() -> i32 {
println!("'-' -> THINKING; 'x' -> EATING; ' ' -> WAITING "); println!("'-' -> THINKING; 'x' -> EATING; ' ' -> WAITING ");
for id in (0..N).into_iter().chain(0..=0) { for id in (0..N).into_iter().chain(0..=0) {
print!("#{}:", id); print!("#{}:", id);
for j in 0..time_cost/GRAPH_SCALE { for j in 0..time_cost / GRAPH_SCALE {
let current_time = j * GRAPH_SCALE + start; let current_time = j * GRAPH_SCALE + start;
if (0..ROUND).find(|round| unsafe { if (0..ROUND).any(|round| unsafe {
let start_thinking = THINK[id][2 * round]; let start_thinking = THINK[id][2 * round];
let end_thinking = THINK[id][2 * round + 1]; let end_thinking = THINK[id][2 * round + 1];
start_thinking <= current_time && current_time <= end_thinking start_thinking <= current_time && current_time <= end_thinking
}).is_some() { }) {
print!("-"); print!("-");
} else if (0..ROUND).find(|round| unsafe { } else if (0..ROUND).any(|round| unsafe {
let start_eating = EAT[id][2 * round]; let start_eating = EAT[id][2 * round];
let end_eating = EAT[id][2 * round + 1]; let end_eating = EAT[id][2 * round + 1];
start_eating <= current_time && current_time <= end_eating start_eating <= current_time && current_time <= end_eating
}).is_some() { }) {
print!("x"); print!("x");
} else { } else {
print!(" "); print!(" ");

View file

@ -6,8 +6,8 @@ extern crate user_lib;
extern crate alloc; extern crate alloc;
use user_lib::{fork, close, pipe, read, write, wait, get_time};
use alloc::format; use alloc::format;
use user_lib::{close, fork, get_time, pipe, read, wait, write};
const LENGTH: usize = 3000; const LENGTH: usize = 3000;
#[no_mangle] #[no_mangle]
@ -40,11 +40,14 @@ pub fn main() -> i32 {
// close write end of up pipe // close write end of up pipe
close(up_pipe_fd[1]); close(up_pipe_fd[1]);
// generate a long random string // generate a long random string
for i in 0..LENGTH { for ch in random_str.iter_mut() {
random_str[i] = get_time() as u8; *ch = get_time() as u8;
} }
// send it // send it
assert_eq!(write(down_pipe_fd[1], &random_str) as usize, random_str.len()); assert_eq!(
write(down_pipe_fd[1], &random_str) as usize,
random_str.len()
);
// close write end of down pipe // close write end of down pipe
close(down_pipe_fd[1]); close(down_pipe_fd[1]);
// calculate sum(parent) // calculate sum(parent)
@ -57,9 +60,8 @@ pub fn main() -> i32 {
// check // check
assert_eq!( assert_eq!(
sum, sum,
str::parse::<usize>( str::parse::<usize>(core::str::from_utf8(&child_result[..result_len]).unwrap())
core::str::from_utf8(&child_result[..result_len]).unwrap() .unwrap()
).unwrap()
); );
let mut _unused: i32 = 0; let mut _unused: i32 = 0;
wait(&mut _unused); wait(&mut _unused);

View file

@ -4,7 +4,7 @@
#[macro_use] #[macro_use]
extern crate user_lib; extern crate user_lib;
use user_lib::{fork, close, pipe, read, write, wait}; use user_lib::{close, fork, pipe, read, wait, write};
static STR: &str = "Hello, world!"; static STR: &str = "Hello, world!";

17
user/src/bin/priv_csr.rs Normal file
View file

@ -0,0 +1,17 @@
#![no_std]
#![no_main]
#[macro_use]
extern crate user_lib;
use riscv::register::sstatus::{self, SPP};
#[no_mangle]
fn main() -> i32 {
println!("Try to access privileged CSR in U Mode");
println!("Kernel should kill this application!");
unsafe {
sstatus::set_spp(SPP::User);
}
0
}

17
user/src/bin/priv_inst.rs Normal file
View file

@ -0,0 +1,17 @@
#![no_std]
#![no_main]
#[macro_use]
extern crate user_lib;
use core::arch::asm;
#[no_mangle]
fn main() -> i32 {
println!("Try to execute privileged instruction in U Mode");
println!("Kernel should kill this application!");
unsafe {
asm!("sret");
}
0
}

View file

@ -5,8 +5,8 @@
extern crate user_lib; extern crate user_lib;
extern crate alloc; extern crate alloc;
use user_lib::{exit, thread_create, waittid, get_time};
use alloc::vec::Vec; use alloc::vec::Vec;
use user_lib::{exit, get_time, thread_create, waittid};
static mut A: usize = 0; static mut A: usize = 0;
const PER_THREAD: usize = 1000; const PER_THREAD: usize = 1000;
@ -17,7 +17,9 @@ unsafe fn f() -> ! {
for _ in 0..PER_THREAD { for _ in 0..PER_THREAD {
let a = &mut A as *mut usize; let a = &mut A as *mut usize;
let cur = a.read_volatile(); let cur = a.read_volatile();
for _ in 0..500 { t = t * t % 10007; } for _ in 0..500 {
t = t * t % 10007;
}
a.write_volatile(cur + 1); a.write_volatile(cur + 1);
} }
exit(t as i32) exit(t as i32)

View file

@ -0,0 +1,53 @@
#![no_std]
#![no_main]
#[macro_use]
extern crate user_lib;
extern crate alloc;
use crate::alloc::string::ToString;
use alloc::vec::Vec;
use user_lib::{exit, get_time, thread_create, waittid};
static mut A: usize = 0;
const PER_THREAD: usize = 1000;
const THREAD_COUNT: usize = 16;
unsafe fn f(count: usize) -> ! {
let mut t = 2usize;
for _ in 0..PER_THREAD {
let a = &mut A as *mut usize;
let cur = a.read_volatile();
for _ in 0..count {
t = t * t % 10007;
}
a.write_volatile(cur + 1);
}
exit(t as i32)
}
#[no_mangle]
pub fn main(argc: usize, argv: &[&str]) -> i32 {
let count: usize;
if argc == 1 {
count = THREAD_COUNT;
} else if argc == 2 {
count = argv[1].to_string().parse::<usize>().unwrap();
} else {
println!("ERROR in argv");
exit(-1);
}
let start = get_time();
let mut v = Vec::new();
for _ in 0..THREAD_COUNT {
v.push(thread_create(f as usize, count) as usize);
}
let mut time_cost = Vec::new();
for tid in v.iter() {
time_cost.push(waittid(*tid));
}
println!("time cost is {}ms", get_time() - start);
assert_eq!(unsafe { A }, PER_THREAD * THREAD_COUNT);
0
}

View file

@ -5,9 +5,9 @@
extern crate user_lib; extern crate user_lib;
extern crate alloc; extern crate alloc;
use user_lib::{exit, thread_create, waittid, get_time, yield_};
use alloc::vec::Vec; use alloc::vec::Vec;
use core::sync::atomic::{AtomicBool, Ordering}; use core::sync::atomic::{AtomicBool, Ordering};
use user_lib::{exit, get_time, thread_create, waittid, yield_};
static mut A: usize = 0; static mut A: usize = 0;
static OCCUPIED: AtomicBool = AtomicBool::new(false); static OCCUPIED: AtomicBool = AtomicBool::new(false);
@ -17,12 +17,17 @@ const THREAD_COUNT: usize = 16;
unsafe fn f() -> ! { unsafe fn f() -> ! {
let mut t = 2usize; let mut t = 2usize;
for _ in 0..PER_THREAD { for _ in 0..PER_THREAD {
while OCCUPIED.compare_exchange(false, true, Ordering::Relaxed, Ordering::Relaxed).is_err() { while OCCUPIED
.compare_exchange(false, true, Ordering::Relaxed, Ordering::Relaxed)
.is_err()
{
yield_(); yield_();
} }
let a = &mut A as *mut usize; let a = &mut A as *mut usize;
let cur = a.read_volatile(); let cur = a.read_volatile();
for _ in 0..500 { t = t * t % 10007; } for _ in 0..500 {
t = t * t % 10007;
}
a.write_volatile(cur + 1); a.write_volatile(cur + 1);
OCCUPIED.store(false, Ordering::Relaxed); OCCUPIED.store(false, Ordering::Relaxed);
} }

View file

@ -5,8 +5,8 @@
extern crate user_lib; extern crate user_lib;
extern crate alloc; extern crate alloc;
use user_lib::{exit, thread_create, waittid, get_time, yield_};
use alloc::vec::Vec; use alloc::vec::Vec;
use user_lib::{exit, get_time, thread_create, waittid, yield_};
static mut A: usize = 0; static mut A: usize = 0;
static mut OCCUPIED: bool = false; static mut OCCUPIED: bool = false;
@ -16,12 +16,16 @@ const THREAD_COUNT: usize = 16;
unsafe fn f() -> ! { unsafe fn f() -> ! {
let mut t = 2usize; let mut t = 2usize;
for _ in 0..PER_THREAD { for _ in 0..PER_THREAD {
while OCCUPIED { yield_(); } while OCCUPIED {
yield_();
}
OCCUPIED = true; OCCUPIED = true;
// enter critical section // enter critical section
let a = &mut A as *mut usize; let a = &mut A as *mut usize;
let cur = a.read_volatile(); let cur = a.read_volatile();
for _ in 0..500 { t = t * t % 10007; } for _ in 0..500 {
t = t * t % 10007;
}
a.write_volatile(cur + 1); a.write_volatile(cur + 1);
// exit critical section // exit critical section
OCCUPIED = false; OCCUPIED = false;

View file

@ -5,9 +5,9 @@
extern crate user_lib; extern crate user_lib;
extern crate alloc; extern crate alloc;
use user_lib::{exit, thread_create, waittid, get_time};
use user_lib::{mutex_blocking_create, mutex_lock, mutex_unlock};
use alloc::vec::Vec; use alloc::vec::Vec;
use user_lib::{exit, get_time, thread_create, waittid};
use user_lib::{mutex_blocking_create, mutex_lock, mutex_unlock};
static mut A: usize = 0; static mut A: usize = 0;
const PER_THREAD: usize = 1000; const PER_THREAD: usize = 1000;
@ -19,7 +19,9 @@ unsafe fn f() -> ! {
mutex_lock(0); mutex_lock(0);
let a = &mut A as *mut usize; let a = &mut A as *mut usize;
let cur = a.read_volatile(); let cur = a.read_volatile();
for _ in 0..500 { t = t * t % 10007; } for _ in 0..500 {
t = t * t % 10007;
}
a.write_volatile(cur + 1); a.write_volatile(cur + 1);
mutex_unlock(0); mutex_unlock(0);
} }

View file

@ -5,9 +5,9 @@
extern crate user_lib; extern crate user_lib;
extern crate alloc; extern crate alloc;
use user_lib::{exit, thread_create, waittid, get_time};
use user_lib::{mutex_create, mutex_lock, mutex_unlock};
use alloc::vec::Vec; use alloc::vec::Vec;
use user_lib::{exit, get_time, thread_create, waittid};
use user_lib::{mutex_create, mutex_lock, mutex_unlock};
static mut A: usize = 0; static mut A: usize = 0;
const PER_THREAD: usize = 1000; const PER_THREAD: usize = 1000;
@ -19,7 +19,9 @@ unsafe fn f() -> ! {
mutex_lock(0); mutex_lock(0);
let a = &mut A as *mut usize; let a = &mut A as *mut usize;
let cur = a.read_volatile(); let cur = a.read_volatile();
for _ in 0..500 { t = t * t % 10007; } for _ in 0..500 {
t = t * t % 10007;
}
a.write_volatile(cur + 1); a.write_volatile(cur + 1);
mutex_unlock(0); mutex_unlock(0);
} }

View file

@ -4,13 +4,13 @@
#[macro_use] #[macro_use]
extern crate user_lib; extern crate user_lib;
use user_lib::{fork, exec, wait}; use user_lib::{exec, fork, wait};
#[no_mangle] #[no_mangle]
pub fn main() -> i32 { pub fn main() -> i32 {
for i in 0..1000 { for i in 0..1000 {
if fork() == 0 { if fork() == 0 {
exec("pipe_large_test\0", &[0 as *const u8]); exec("pipe_large_test\0", &[core::ptr::null::<u8>()]);
} else { } else {
let mut _unused: i32 = 0; let mut _unused: i32 = 0;
wait(&mut _unused); wait(&mut _unused);

View file

@ -4,10 +4,10 @@
#[macro_use] #[macro_use]
extern crate user_lib; extern crate user_lib;
use user_lib::{sleep, exit, get_time, fork, waitpid}; use user_lib::{exit, fork, get_time, sleep, waitpid};
fn sleepy() { fn sleepy() {
let time: usize = 1000; let time: usize = 100;
for i in 0..5 { for i in 0..5 {
sleep(time); sleep(time);
println!("sleep {} x {} msecs.", i + 1, time); println!("sleep {} x {} msecs.", i + 1, time);

View file

@ -13,7 +13,11 @@ pub fn main() -> i32 {
println!("current time_msec = {}", start); println!("current time_msec = {}", start);
sleep(100); sleep(100);
let end = get_time(); let end = get_time();
println!("time_msec = {} after sleeping 100 ticks, delta = {}ms!", end, end - start); println!(
"time_msec = {} after sleeping 100 ticks, delta = {}ms!",
end,
end - start
);
println!("r_sleep passed!"); println!("r_sleep passed!");
0 0
} }

View file

@ -5,7 +5,7 @@
extern crate user_lib; extern crate user_lib;
fn f(d: usize) { fn f(d: usize) {
println!("d = {}",d); println!("d = {}", d);
f(d + 1); f(d + 1);
} }

View file

@ -0,0 +1,15 @@
#![no_std]
#![no_main]
#[macro_use]
extern crate user_lib;
#[no_mangle]
fn main() -> i32 {
println!("Into Test store_fault, we will insert an invalid store operation...");
println!("Kernel should kill this application!");
unsafe {
core::ptr::null_mut::<u8>().write_volatile(0);
}
0
}

45
user/src/bin/sync_sem.rs Normal file
View file

@ -0,0 +1,45 @@
#![no_std]
#![no_main]
#[macro_use]
extern crate user_lib;
extern crate alloc;
use alloc::vec;
use user_lib::exit;
use user_lib::{semaphore_create, semaphore_down, semaphore_up};
use user_lib::{sleep, thread_create, waittid};
const SEM_SYNC: usize = 0;
unsafe fn first() -> ! {
sleep(10);
println!("First work and wakeup Second");
semaphore_up(SEM_SYNC);
exit(0)
}
unsafe fn second() -> ! {
println!("Second want to continue,but need to wait first");
semaphore_down(SEM_SYNC);
println!("Second can work now");
exit(0)
}
#[no_mangle]
pub fn main() -> i32 {
// create semaphores
assert_eq!(semaphore_create(0) as usize, SEM_SYNC);
// create threads
let threads = vec![
thread_create(first as usize, 0),
thread_create(second as usize, 0),
];
// wait for all threads to complete
for thread in threads.iter() {
waittid(*thread as usize);
}
println!("sync_sem passed!");
0
}

View file

@ -0,0 +1,59 @@
#![no_std]
#![no_main]
#[macro_use]
extern crate user_lib;
extern crate alloc;
use alloc::vec;
use user_lib::exit;
use user_lib::{
condvar_create, condvar_signal, condvar_wait, mutex_blocking_create, mutex_lock, mutex_unlock,
};
use user_lib::{sleep, thread_create, waittid};
static mut A: usize = 0;
const CONDVAR_ID: usize = 0;
const MUTEX_ID: usize = 0;
unsafe fn first() -> ! {
sleep(10);
println!("First work, Change A --> 1 and wakeup Second");
mutex_lock(MUTEX_ID);
A = 1;
condvar_signal(CONDVAR_ID);
mutex_unlock(MUTEX_ID);
exit(0)
}
unsafe fn second() -> ! {
println!("Second want to continue,but need to wait A=1");
mutex_lock(MUTEX_ID);
while A == 0 {
println!("Second: A is {}", A);
condvar_wait(CONDVAR_ID, MUTEX_ID);
}
mutex_unlock(MUTEX_ID);
println!("A is {}, Second can work now", A);
exit(0)
}
#[no_mangle]
pub fn main() -> i32 {
// create condvar & mutex
assert_eq!(condvar_create() as usize, CONDVAR_ID);
assert_eq!(mutex_blocking_create() as usize, MUTEX_ID);
// create threads
let threads = vec![
thread_create(first as usize, 0),
thread_create(second as usize, 0),
];
// wait for all threads to complete
for thread in threads.iter() {
waittid(*thread as usize);
}
println!("test_condvar passed!");
0
}

View file

@ -5,30 +5,37 @@
extern crate user_lib; extern crate user_lib;
extern crate alloc; extern crate alloc;
use user_lib::{thread_create, waittid, exit}; use alloc::vec;
use alloc::vec::Vec; use user_lib::{exit, thread_create, waittid};
pub fn thread_a() -> ! { pub fn thread_a() -> ! {
for _ in 0..1000 { print!("a"); } for _ in 0..1000 {
print!("a");
}
exit(1) exit(1)
} }
pub fn thread_b() -> ! { pub fn thread_b() -> ! {
for _ in 0..1000 { print!("b"); } for _ in 0..1000 {
print!("b");
}
exit(2) exit(2)
} }
pub fn thread_c() -> ! { pub fn thread_c() -> ! {
for _ in 0..1000 { print!("c"); } for _ in 0..1000 {
print!("c");
}
exit(3) exit(3)
} }
#[no_mangle] #[no_mangle]
pub fn main() -> i32 { pub fn main() -> i32 {
let mut v = Vec::new(); let v = vec![
v.push(thread_create(thread_a as usize, 0)); thread_create(thread_a as usize, 0),
v.push(thread_create(thread_b as usize, 0)); thread_create(thread_b as usize, 0),
v.push(thread_create(thread_c as usize, 0)); thread_create(thread_c as usize, 0),
];
for tid in v.iter() { for tid in v.iter() {
let exit_code = waittid(*tid as usize); let exit_code = waittid(*tid as usize);
println!("thread#{} exited with code {}", tid, exit_code); println!("thread#{} exited with code {}", tid, exit_code);

View file

@ -5,8 +5,8 @@
extern crate user_lib; extern crate user_lib;
extern crate alloc; extern crate alloc;
use user_lib::{thread_create, waittid, exit};
use alloc::vec::Vec; use alloc::vec::Vec;
use user_lib::{exit, thread_create, waittid};
struct Argument { struct Argument {
pub ch: char, pub ch: char,
@ -15,7 +15,9 @@ struct Argument {
fn thread_print(arg: *const Argument) -> ! { fn thread_print(arg: *const Argument) -> ! {
let arg = unsafe { &*arg }; let arg = unsafe { &*arg };
for _ in 0..1000 { print!("{}", arg.ch); } for _ in 0..1000 {
print!("{}", arg.ch);
}
exit(arg.rc) exit(arg.rc)
} }
@ -23,12 +25,15 @@ fn thread_print(arg: *const Argument) -> ! {
pub fn main() -> i32 { pub fn main() -> i32 {
let mut v = Vec::new(); let mut v = Vec::new();
let args = [ let args = [
Argument { ch: 'a', rc: 1, }, Argument { ch: 'a', rc: 1 },
Argument { ch: 'b', rc: 2, }, Argument { ch: 'b', rc: 2 },
Argument { ch: 'c', rc: 3, }, Argument { ch: 'c', rc: 3 },
]; ];
for i in 0..3 { for arg in args.iter() {
v.push(thread_create(thread_print as usize, &args[i] as *const _ as usize)); v.push(thread_create(
thread_print as usize,
arg as *const _ as usize,
));
} }
for tid in v.iter() { for tid in v.iter() {
let exit_code = waittid(*tid as usize); let exit_code = waittid(*tid as usize);

View file

@ -0,0 +1,46 @@
#![no_std]
#![no_main]
#[macro_use]
extern crate user_lib;
use user_lib::{exec, fork, get_time, kill, waitpid, waitpid_nb, SignalFlags};
#[no_mangle]
pub fn main(argc: usize, argv: &[&str]) -> i32 {
assert_eq!(argc, 3, "argc must be 3!");
let timeout_ms = argv[2]
.parse::<isize>()
.expect("Error when parsing timeout!");
let pid = fork() as usize;
if pid == 0 {
if exec(argv[1], &[core::ptr::null::<u8>()]) != 0 {
println!("Error when executing '{}'", argv[1]);
return -4;
}
} else {
let start_time = get_time();
let mut child_exited = false;
let mut exit_code: i32 = 0;
loop {
if get_time() - start_time > timeout_ms {
break;
}
if waitpid_nb(pid, &mut exit_code) as usize == pid {
child_exited = true;
println!(
"child exited in {}ms, exit_code = {}",
get_time() - start_time,
exit_code,
);
}
}
if !child_exited {
println!("child has run for {}ms, kill it!", timeout_ms);
kill(pid, SignalFlags::SIGINT.bits());
assert_eq!(waitpid(pid, &mut exit_code) as usize, pid);
println!("exit code of the child is {}", exit_code);
}
}
0
}

View file

@ -1,5 +1,6 @@
#![no_std] #![no_std]
#![no_main] #![no_main]
#![allow(clippy::println_empty_string)]
extern crate alloc; extern crate alloc;
@ -10,116 +11,191 @@ const LF: u8 = 0x0au8;
const CR: u8 = 0x0du8; const CR: u8 = 0x0du8;
const DL: u8 = 0x7fu8; const DL: u8 = 0x7fu8;
const BS: u8 = 0x08u8; const BS: u8 = 0x08u8;
const LINE_START: &str = ">> ";
use alloc::string::String; use alloc::string::String;
use alloc::vec::Vec; use alloc::vec::Vec;
use user_lib::{
fork,
exec,
waitpid,
open,
OpenFlags,
close,
dup,
};
use user_lib::console::getchar; use user_lib::console::getchar;
use user_lib::{close, dup, exec, fork, open, pipe, waitpid, OpenFlags};
#[derive(Debug)]
struct ProcessArguments {
input: String,
output: String,
args_copy: Vec<String>,
args_addr: Vec<*const u8>,
}
impl ProcessArguments {
pub fn new(command: &str) -> Self {
let args: Vec<_> = command.split(' ').collect();
let mut args_copy: Vec<String> = args
.iter()
.filter(|&arg| !arg.is_empty())
.map(|&arg| {
let mut string = String::new();
string.push_str(arg);
string.push('\0');
string
})
.collect();
// redirect input
let mut input = String::new();
if let Some((idx, _)) = args_copy
.iter()
.enumerate()
.find(|(_, arg)| arg.as_str() == "<\0")
{
input = args_copy[idx + 1].clone();
args_copy.drain(idx..=idx + 1);
}
// redirect output
let mut output = String::new();
if let Some((idx, _)) = args_copy
.iter()
.enumerate()
.find(|(_, arg)| arg.as_str() == ">\0")
{
output = args_copy[idx + 1].clone();
args_copy.drain(idx..=idx + 1);
}
let mut args_addr: Vec<*const u8> = args_copy.iter().map(|arg| arg.as_ptr()).collect();
args_addr.push(core::ptr::null::<u8>());
Self {
input,
output,
args_copy,
args_addr,
}
}
}
#[no_mangle] #[no_mangle]
pub fn main() -> i32 { pub fn main() -> i32 {
println!("Rust user shell"); println!("Rust user shell");
let mut line: String = String::new(); let mut line: String = String::new();
print!(">> "); print!("{}", LINE_START);
loop { loop {
let c = getchar(); let c = getchar();
match c { match c {
LF | CR => { LF | CR => {
println!(""); println!("");
if !line.is_empty() { if !line.is_empty() {
let args: Vec<_> = line.as_str().split(' ').collect(); let splited: Vec<_> = line.as_str().split('|').collect();
let mut args_copy: Vec<String> = args let process_arguments_list: Vec<_> = splited
.iter()
.map(|&arg| {
let mut string = String::new();
string.push_str(arg);
string
})
.collect();
args_copy
.iter_mut()
.for_each(|string| {
string.push('\0');
});
// redirect input
let mut input = String::new();
if let Some((idx, _)) = args_copy
.iter()
.enumerate()
.find(|(_, arg)| arg.as_str() == "<\0") {
input = args_copy[idx + 1].clone();
args_copy.drain(idx..=idx + 1);
}
// redirect output
let mut output = String::new();
if let Some((idx, _)) = args_copy
.iter()
.enumerate()
.find(|(_, arg)| arg.as_str() == ">\0") {
output = args_copy[idx + 1].clone();
args_copy.drain(idx..=idx + 1);
}
let mut args_addr: Vec<*const u8> = args_copy
.iter() .iter()
.map(|arg| arg.as_ptr()) .map(|&cmd| ProcessArguments::new(cmd))
.collect(); .collect();
args_addr.push(0 as *const u8); let mut valid = true;
let pid = fork(); for (i, process_args) in process_arguments_list.iter().enumerate() {
if pid == 0 { if i == 0 {
// input redirection if !process_args.output.is_empty() {
if !input.is_empty() { valid = false;
let input_fd = open(input.as_str(), OpenFlags::RDONLY);
if input_fd == -1 {
println!("Error when opening file {}", input);
return -4;
} }
let input_fd = input_fd as usize; } else if i == process_arguments_list.len() - 1 {
close(0); if !process_args.input.is_empty() {
assert_eq!(dup(input_fd), 0); valid = false;
close(input_fd);
}
// output redirection
if !output.is_empty() {
let output_fd = open(
output.as_str(),
OpenFlags::CREATE | OpenFlags::WRONLY
);
if output_fd == -1 {
println!("Error when opening file {}", output);
return -4;
} }
let output_fd = output_fd as usize; } else if !process_args.output.is_empty() || !process_args.input.is_empty()
close(1); {
assert_eq!(dup(output_fd), 1); valid = false;
close(output_fd);
} }
// child process }
if exec(args_copy[0].as_str(), args_addr.as_slice()) == -1 { if process_arguments_list.len() == 1 {
println!("Error when executing!"); valid = true;
return -4; }
} if !valid {
unreachable!(); println!("Invalid command: Inputs/Outputs cannot be correctly binded!");
} else { } else {
// create pipes
let mut pipes_fd: Vec<[usize; 2]> = Vec::new();
if !process_arguments_list.is_empty() {
for _ in 0..process_arguments_list.len() - 1 {
let mut pipe_fd = [0usize; 2];
pipe(&mut pipe_fd);
pipes_fd.push(pipe_fd);
}
}
let mut children: Vec<_> = Vec::new();
for (i, process_argument) in process_arguments_list.iter().enumerate() {
let pid = fork();
if pid == 0 {
let input = &process_argument.input;
let output = &process_argument.output;
let args_copy = &process_argument.args_copy;
let args_addr = &process_argument.args_addr;
// redirect input
if !input.is_empty() {
let input_fd = open(input.as_str(), OpenFlags::RDONLY);
if input_fd == -1 {
println!("Error when opening file {}", input);
return -4;
}
let input_fd = input_fd as usize;
close(0);
assert_eq!(dup(input_fd), 0);
close(input_fd);
}
// redirect output
if !output.is_empty() {
let output_fd = open(
output.as_str(),
OpenFlags::CREATE | OpenFlags::WRONLY,
);
if output_fd == -1 {
println!("Error when opening file {}", output);
return -4;
}
let output_fd = output_fd as usize;
close(1);
assert_eq!(dup(output_fd), 1);
close(output_fd);
}
// receive input from the previous process
if i > 0 {
close(0);
let read_end = pipes_fd.get(i - 1).unwrap()[0];
assert_eq!(dup(read_end), 0);
}
// send output to the next process
if i < process_arguments_list.len() - 1 {
close(1);
let write_end = pipes_fd.get(i).unwrap()[1];
assert_eq!(dup(write_end), 1);
}
// close all pipe ends inherited from the parent process
for pipe_fd in pipes_fd.iter() {
close(pipe_fd[0]);
close(pipe_fd[1]);
}
// execute new application
if exec(args_copy[0].as_str(), args_addr.as_slice()) == -1 {
println!("Error when executing!");
return -4;
}
unreachable!();
} else {
children.push(pid);
}
}
for pipe_fd in pipes_fd.iter() {
close(pipe_fd[0]);
close(pipe_fd[1]);
}
let mut exit_code: i32 = 0; let mut exit_code: i32 = 0;
let exit_pid = waitpid(pid as usize, &mut exit_code); for pid in children.into_iter() {
assert_eq!(pid, exit_pid); let exit_pid = waitpid(pid as usize, &mut exit_code);
println!("Shell: Process {} exited with code {}", pid, exit_code); assert_eq!(pid, exit_pid);
//println!("Shell: Process {} exited with code {}", pid, exit_code);
}
} }
line.clear(); line.clear();
} }
print!(">> "); print!("{}", LINE_START);
} }
BS | DL => { BS | DL => {
if !line.is_empty() { if !line.is_empty() {

View file

@ -26,13 +26,16 @@ pub fn main() -> i32 {
println!("Usertests: Running {}", test); println!("Usertests: Running {}", test);
let pid = fork(); let pid = fork();
if pid == 0 { if pid == 0 {
exec(*test, &[0 as *const u8]); exec(*test, &[core::ptr::null::<u8>()]);
panic!("unreachable!"); panic!("unreachable!");
} else { } else {
let mut exit_code: i32 = Default::default(); let mut exit_code: i32 = Default::default();
let wait_pid = waitpid(pid as usize, &mut exit_code); let wait_pid = waitpid(pid as usize, &mut exit_code);
assert_eq!(pid, wait_pid); assert_eq!(pid, wait_pid);
println!("\x1b[32mUsertests: Test {} in Process {} exited with code {}\x1b[0m", test, pid, exit_code); println!(
"\x1b[32mUsertests: Test {} in Process {} exited with code {}\x1b[0m",
test, pid, exit_code
);
} }
} }
println!("Usertests passed!"); println!("Usertests passed!");

View file

@ -1,12 +1,18 @@
use super::exit; use super::{getpid, kill, SignalFlags};
#[panic_handler] #[panic_handler]
fn panic_handler(panic_info: &core::panic::PanicInfo) -> ! { fn panic_handler(panic_info: &core::panic::PanicInfo) -> ! {
let err = panic_info.message().unwrap(); let err = panic_info.message().unwrap();
if let Some(location) = panic_info.location() { if let Some(location) = panic_info.location() {
println!("Panicked at {}:{}, {}", location.file(), location.line(), err); println!(
"Panicked at {}:{}, {}",
location.file(),
location.line(),
err
);
} else { } else {
println!("Panicked: {}", err); println!("Panicked: {}", err);
} }
exit(-1); kill(getpid() as usize, SignalFlags::SIGABRT.bits());
unreachable!()
} }

View file

@ -1,21 +1,20 @@
#![no_std] #![no_std]
#![feature(asm)]
#![feature(linkage)] #![feature(linkage)]
#![feature(panic_info_message)] #![feature(panic_info_message)]
#![feature(alloc_error_handler)] #![feature(alloc_error_handler)]
#[macro_use] #[macro_use]
pub mod console; pub mod console;
mod syscall;
mod lang_items; mod lang_items;
mod syscall;
extern crate alloc; extern crate alloc;
#[macro_use] #[macro_use]
extern crate bitflags; extern crate bitflags;
use syscall::*;
use buddy_system_allocator::LockedHeap;
use alloc::vec::Vec; use alloc::vec::Vec;
use buddy_system_allocator::LockedHeap;
use syscall::*;
const USER_HEAP_SIZE: usize = 32768; const USER_HEAP_SIZE: usize = 32768;
@ -38,16 +37,16 @@ pub extern "C" fn _start(argc: usize, argv: usize) -> ! {
} }
let mut v: Vec<&'static str> = Vec::new(); let mut v: Vec<&'static str> = Vec::new();
for i in 0..argc { for i in 0..argc {
let str_start = unsafe { let str_start =
((argv + i * core::mem::size_of::<usize>()) as *const usize).read_volatile() unsafe { ((argv + i * core::mem::size_of::<usize>()) as *const usize).read_volatile() };
}; let len = (0usize..)
let len = (0usize..).find(|i| unsafe { .find(|i| unsafe { ((str_start + *i) as *const u8).read_volatile() == 0 })
((str_start + *i) as *const u8).read_volatile() == 0 .unwrap();
}).unwrap();
v.push( v.push(
core::str::from_utf8(unsafe { core::str::from_utf8(unsafe {
core::slice::from_raw_parts(str_start as *const u8, len) core::slice::from_raw_parts(str_start as *const u8, len)
}).unwrap() })
.unwrap(),
); );
} }
exit(main(argc, v.as_slice())); exit(main(argc, v.as_slice()));
@ -69,22 +68,48 @@ bitflags! {
} }
} }
pub fn dup(fd: usize) -> isize { sys_dup(fd) } pub fn dup(fd: usize) -> isize {
pub fn open(path: &str, flags: OpenFlags) -> isize { sys_open(path, flags.bits) } sys_dup(fd)
pub fn close(fd: usize) -> isize { sys_close(fd) } }
pub fn pipe(pipe_fd: &mut [usize]) -> isize { sys_pipe(pipe_fd) } pub fn open(path: &str, flags: OpenFlags) -> isize {
pub fn read(fd: usize, buf: &mut [u8]) -> isize { sys_read(fd, buf) } sys_open(path, flags.bits)
pub fn write(fd: usize, buf: &[u8]) -> isize { sys_write(fd, buf) } }
pub fn exit(exit_code: i32) -> ! { sys_exit(exit_code); } pub fn close(fd: usize) -> isize {
pub fn yield_() -> isize { sys_yield() } sys_close(fd)
pub fn get_time() -> isize { sys_get_time() } }
pub fn getpid() -> isize { sys_getpid() } pub fn pipe(pipe_fd: &mut [usize]) -> isize {
pub fn fork() -> isize { sys_fork() } sys_pipe(pipe_fd)
pub fn exec(path: &str, args: &[*const u8]) -> isize { sys_exec(path, args) } }
pub fn read(fd: usize, buf: &mut [u8]) -> isize {
sys_read(fd, buf)
}
pub fn write(fd: usize, buf: &[u8]) -> isize {
sys_write(fd, buf)
}
pub fn exit(exit_code: i32) -> ! {
sys_exit(exit_code);
}
pub fn yield_() -> isize {
sys_yield()
}
pub fn get_time() -> isize {
sys_get_time()
}
pub fn getpid() -> isize {
sys_getpid()
}
pub fn fork() -> isize {
sys_fork()
}
pub fn exec(path: &str, args: &[*const u8]) -> isize {
sys_exec(path, args)
}
pub fn wait(exit_code: &mut i32) -> isize { pub fn wait(exit_code: &mut i32) -> isize {
loop { loop {
match sys_waitpid(-1, exit_code as *mut _) { match sys_waitpid(-1, exit_code as *mut _) {
-2 => { yield_(); } -2 => {
yield_();
}
// -1 or a real pid // -1 or a real pid
exit_pid => return exit_pid, exit_pid => return exit_pid,
} }
@ -94,31 +119,66 @@ pub fn wait(exit_code: &mut i32) -> isize {
pub fn waitpid(pid: usize, exit_code: &mut i32) -> isize { pub fn waitpid(pid: usize, exit_code: &mut i32) -> isize {
loop { loop {
match sys_waitpid(pid as isize, exit_code as *mut _) { match sys_waitpid(pid as isize, exit_code as *mut _) {
-2 => { yield_(); } -2 => {
yield_();
}
// -1 or a real pid // -1 or a real pid
exit_pid => return exit_pid, exit_pid => return exit_pid,
} }
} }
} }
pub fn waitpid_nb(pid: usize, exit_code: &mut i32) -> isize {
sys_waitpid(pid as isize, exit_code as *mut _)
}
bitflags! {
pub struct SignalFlags: i32 {
const SIGINT = 1 << 2;
const SIGILL = 1 << 4;
const SIGABRT = 1 << 6;
const SIGFPE = 1 << 8;
const SIGSEGV = 1 << 11;
}
}
pub fn kill(pid: usize, signal: i32) -> isize {
sys_kill(pid, signal)
}
pub fn sleep(sleep_ms: usize) { pub fn sleep(sleep_ms: usize) {
sys_sleep(sleep_ms); sys_sleep(sleep_ms);
} }
pub fn thread_create(entry: usize, arg: usize) -> isize { sys_thread_create(entry, arg) } pub fn thread_create(entry: usize, arg: usize) -> isize {
pub fn gettid() -> isize { sys_gettid() } sys_thread_create(entry, arg)
}
pub fn gettid() -> isize {
sys_gettid()
}
pub fn waittid(tid: usize) -> isize { pub fn waittid(tid: usize) -> isize {
loop { loop {
match sys_waittid(tid) { match sys_waittid(tid) {
-2 => { yield_(); } -2 => {
yield_();
}
exit_code => return exit_code, exit_code => return exit_code,
} }
} }
} }
pub fn mutex_create() -> isize { sys_mutex_create(false) } pub fn mutex_create() -> isize {
pub fn mutex_blocking_create() -> isize { sys_mutex_create(true) } sys_mutex_create(false)
pub fn mutex_lock(mutex_id: usize) { sys_mutex_lock(mutex_id); } }
pub fn mutex_unlock(mutex_id: usize) { sys_mutex_unlock(mutex_id); } pub fn mutex_blocking_create() -> isize {
sys_mutex_create(true)
}
pub fn mutex_lock(mutex_id: usize) {
sys_mutex_lock(mutex_id);
}
pub fn mutex_unlock(mutex_id: usize) {
sys_mutex_unlock(mutex_id);
}
pub fn semaphore_create(res_count: usize) -> isize { pub fn semaphore_create(res_count: usize) -> isize {
sys_semaphore_create(res_count) sys_semaphore_create(res_count)
} }
@ -128,4 +188,12 @@ pub fn semaphore_up(sem_id: usize) {
pub fn semaphore_down(sem_id: usize) { pub fn semaphore_down(sem_id: usize) {
sys_semaphore_down(sem_id); sys_semaphore_down(sem_id);
} }
pub fn condvar_create() -> isize {
sys_condvar_create(0)
}
pub fn condvar_signal(condvar_id: usize) {
sys_condvar_signal(condvar_id);
}
pub fn condvar_wait(condvar_id: usize, mutex_id: usize) {
sys_condvar_wait(condvar_id, mutex_id);
}

View file

@ -7,6 +7,7 @@ const SYSCALL_WRITE: usize = 64;
const SYSCALL_EXIT: usize = 93; const SYSCALL_EXIT: usize = 93;
const SYSCALL_SLEEP: usize = 101; const SYSCALL_SLEEP: usize = 101;
const SYSCALL_YIELD: usize = 124; const SYSCALL_YIELD: usize = 124;
const SYSCALL_KILL: usize = 129;
const SYSCALL_GET_TIME: usize = 169; const SYSCALL_GET_TIME: usize = 169;
const SYSCALL_GETPID: usize = 172; const SYSCALL_GETPID: usize = 172;
const SYSCALL_FORK: usize = 220; const SYSCALL_FORK: usize = 220;
@ -21,6 +22,9 @@ const SYSCALL_MUTEX_UNLOCK: usize = 1012;
const SYSCALL_SEMAPHORE_CREATE: usize = 1020; const SYSCALL_SEMAPHORE_CREATE: usize = 1020;
const SYSCALL_SEMAPHORE_UP: usize = 1021; const SYSCALL_SEMAPHORE_UP: usize = 1021;
const SYSCALL_SEMAPHORE_DOWN: usize = 1022; const SYSCALL_SEMAPHORE_DOWN: usize = 1022;
const SYSCALL_CONDVAR_CREATE: usize = 1030;
const SYSCALL_CONDVAR_SIGNAL: usize = 1031;
const SYSCALL_CONDVAR_WAIT: usize = 1032;
fn syscall(id: usize, args: [usize; 3]) -> isize { fn syscall(id: usize, args: [usize; 3]) -> isize {
let mut ret: isize; let mut ret: isize;
@ -53,7 +57,10 @@ pub fn sys_pipe(pipe: &mut [usize]) -> isize {
} }
pub fn sys_read(fd: usize, buffer: &mut [u8]) -> isize { pub fn sys_read(fd: usize, buffer: &mut [u8]) -> isize {
syscall(SYSCALL_READ, [fd, buffer.as_mut_ptr() as usize, buffer.len()]) syscall(
SYSCALL_READ,
[fd, buffer.as_mut_ptr() as usize, buffer.len()],
)
} }
pub fn sys_write(fd: usize, buffer: &[u8]) -> isize { pub fn sys_write(fd: usize, buffer: &[u8]) -> isize {
@ -73,6 +80,10 @@ pub fn sys_yield() -> isize {
syscall(SYSCALL_YIELD, [0, 0, 0]) syscall(SYSCALL_YIELD, [0, 0, 0])
} }
pub fn sys_kill(pid: usize, signal: i32) -> isize {
syscall(SYSCALL_KILL, [pid, signal as usize, 0])
}
pub fn sys_get_time() -> isize { pub fn sys_get_time() -> isize {
syscall(SYSCALL_GET_TIME, [0, 0, 0]) syscall(SYSCALL_GET_TIME, [0, 0, 0])
} }
@ -86,7 +97,10 @@ pub fn sys_fork() -> isize {
} }
pub fn sys_exec(path: &str, args: &[*const u8]) -> isize { pub fn sys_exec(path: &str, args: &[*const u8]) -> isize {
syscall(SYSCALL_EXEC, [path.as_ptr() as usize, args.as_ptr() as usize, 0]) syscall(
SYSCALL_EXEC,
[path.as_ptr() as usize, args.as_ptr() as usize, 0],
)
} }
pub fn sys_waitpid(pid: isize, exit_code: *mut i32) -> isize { pub fn sys_waitpid(pid: isize, exit_code: *mut i32) -> isize {
@ -128,3 +142,15 @@ pub fn sys_semaphore_up(sem_id: usize) -> isize {
pub fn sys_semaphore_down(sem_id: usize) -> isize { pub fn sys_semaphore_down(sem_id: usize) -> isize {
syscall(SYSCALL_SEMAPHORE_DOWN, [sem_id, 0, 0]) syscall(SYSCALL_SEMAPHORE_DOWN, [sem_id, 0, 0])
} }
pub fn sys_condvar_create(_arg: usize) -> isize {
syscall(SYSCALL_CONDVAR_CREATE, [_arg, 0, 0])
}
pub fn sys_condvar_signal(condvar_id: usize) -> isize {
syscall(SYSCALL_CONDVAR_SIGNAL, [condvar_id, 0, 0])
}
pub fn sys_condvar_wait(condvar_id: usize, mutex_id: usize) -> isize {
syscall(SYSCALL_CONDVAR_WAIT, [condvar_id, mutex_id, 0])
}