1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
|
//! Fake HAL implementation for tests.
#![deny(unsafe_op_in_unsafe_fn)]
use crate::{BufferDirection, Hal, PhysAddr, PAGE_SIZE};
use alloc::alloc::{alloc_zeroed, dealloc, handle_alloc_error};
use core::{
alloc::Layout,
ptr::{self, NonNull},
};
use zerocopy::FromZeroes;
#[derive(Debug)]
pub struct FakeHal;
/// Fake HAL implementation for use in unit tests.
unsafe impl Hal for FakeHal {
fn dma_alloc(pages: usize, _direction: BufferDirection) -> (PhysAddr, NonNull<u8>) {
assert_ne!(pages, 0);
let layout = Layout::from_size_align(pages * PAGE_SIZE, PAGE_SIZE).unwrap();
// Safe because the size and alignment of the layout are non-zero.
let ptr = unsafe { alloc_zeroed(layout) };
if let Some(ptr) = NonNull::new(ptr) {
(ptr.as_ptr() as PhysAddr, ptr)
} else {
handle_alloc_error(layout);
}
}
unsafe fn dma_dealloc(_paddr: PhysAddr, vaddr: NonNull<u8>, pages: usize) -> i32 {
assert_ne!(pages, 0);
let layout = Layout::from_size_align(pages * PAGE_SIZE, PAGE_SIZE).unwrap();
// Safe because the layout is the same as was used when the memory was allocated by
// `dma_alloc` above.
unsafe {
dealloc(vaddr.as_ptr(), layout);
}
0
}
unsafe fn mmio_phys_to_virt(paddr: PhysAddr, _size: usize) -> NonNull<u8> {
NonNull::new(paddr as _).unwrap()
}
unsafe fn share(buffer: NonNull<[u8]>, direction: BufferDirection) -> PhysAddr {
assert_ne!(buffer.len(), 0);
// To ensure that the driver is handling and unsharing buffers properly, allocate a new
// buffer and copy to it if appropriate.
let mut shared_buffer = u8::new_box_slice_zeroed(buffer.len());
if let BufferDirection::DriverToDevice | BufferDirection::Both = direction {
unsafe {
buffer
.as_ptr()
.cast::<u8>()
.copy_to(shared_buffer.as_mut_ptr(), buffer.len());
}
}
let vaddr = Box::into_raw(shared_buffer) as *mut u8 as usize;
// Nothing to do, as the host already has access to all memory.
virt_to_phys(vaddr)
}
unsafe fn unshare(paddr: PhysAddr, buffer: NonNull<[u8]>, direction: BufferDirection) {
assert_ne!(buffer.len(), 0);
assert_ne!(paddr, 0);
let vaddr = phys_to_virt(paddr);
let shared_buffer = unsafe {
Box::from_raw(ptr::slice_from_raw_parts_mut(
vaddr as *mut u8,
buffer.len(),
))
};
if let BufferDirection::DeviceToDriver | BufferDirection::Both = direction {
unsafe {
buffer
.as_ptr()
.cast::<u8>()
.copy_from(shared_buffer.as_ptr(), buffer.len());
}
}
}
}
fn virt_to_phys(vaddr: usize) -> PhysAddr {
vaddr
}
fn phys_to_virt(paddr: PhysAddr) -> usize {
paddr
}
|