virtio_drivers/device/net/
dev_raw.rs

1use super::{Config, EthernetAddress, Features, VirtioNetHdr};
2use super::{MIN_BUFFER_LEN, NET_HDR_SIZE, QUEUE_RECEIVE, QUEUE_TRANSMIT, SUPPORTED_FEATURES};
3use crate::hal::Hal;
4use crate::queue::VirtQueue;
5use crate::transport::Transport;
6use crate::volatile::volread;
7use crate::{Error, Result};
8use log::{debug, info, warn};
9use zerocopy::AsBytes;
10
11/// Raw driver for a VirtIO network device.
12///
13/// This is a raw version of the VirtIONet driver. It provides non-blocking
14/// methods for transmitting and receiving raw slices, without the buffer
15/// management. For more higher-level functions such as receive buffer backing,
16/// see [`VirtIONet`].
17///
18/// [`VirtIONet`]: super::VirtIONet
19pub struct VirtIONetRaw<H: Hal, T: Transport, const QUEUE_SIZE: usize> {
20    transport: T,
21    mac: EthernetAddress,
22    recv_queue: VirtQueue<H, QUEUE_SIZE>,
23    send_queue: VirtQueue<H, QUEUE_SIZE>,
24}
25
26impl<H: Hal, T: Transport, const QUEUE_SIZE: usize> VirtIONetRaw<H, T, QUEUE_SIZE> {
27    /// Create a new VirtIO-Net driver.
28    pub fn new(mut transport: T) -> Result<Self> {
29        let negotiated_features = transport.begin_init(SUPPORTED_FEATURES);
30        info!("negotiated_features {:?}", negotiated_features);
31        // read configuration space
32        let config = transport.config_space::<Config>()?;
33        let mac;
34        // Safe because config points to a valid MMIO region for the config space.
35        unsafe {
36            mac = volread!(config, mac);
37            debug!(
38                "Got MAC={:02x?}, status={:?}",
39                mac,
40                volread!(config, status)
41            );
42        }
43        let send_queue = VirtQueue::new(
44            &mut transport,
45            QUEUE_TRANSMIT,
46            negotiated_features.contains(Features::RING_INDIRECT_DESC),
47            negotiated_features.contains(Features::RING_EVENT_IDX),
48        )?;
49        let recv_queue = VirtQueue::new(
50            &mut transport,
51            QUEUE_RECEIVE,
52            negotiated_features.contains(Features::RING_INDIRECT_DESC),
53            negotiated_features.contains(Features::RING_EVENT_IDX),
54        )?;
55
56        transport.finish_init();
57
58        Ok(VirtIONetRaw {
59            transport,
60            mac,
61            recv_queue,
62            send_queue,
63        })
64    }
65
66    /// Acknowledge interrupt.
67    pub fn ack_interrupt(&mut self) -> bool {
68        self.transport.ack_interrupt()
69    }
70
71    /// Disable interrupts.
72    pub fn disable_interrupts(&mut self) {
73        self.send_queue.set_dev_notify(false);
74        self.recv_queue.set_dev_notify(false);
75    }
76
77    /// Enable interrupts.
78    pub fn enable_interrupts(&mut self) {
79        self.send_queue.set_dev_notify(true);
80        self.recv_queue.set_dev_notify(true);
81    }
82
83    /// Get MAC address.
84    pub fn mac_address(&self) -> EthernetAddress {
85        self.mac
86    }
87
88    /// Whether can send packet.
89    pub fn can_send(&self) -> bool {
90        self.send_queue.available_desc() >= 2
91    }
92
93    /// Whether the length of the receive buffer is valid.
94    fn check_rx_buf_len(rx_buf: &[u8]) -> Result<()> {
95        if rx_buf.len() < MIN_BUFFER_LEN {
96            warn!("Receive buffer len {} is too small", rx_buf.len());
97            Err(Error::InvalidParam)
98        } else {
99            Ok(())
100        }
101    }
102
103    /// Whether the length of the transmit buffer is valid.
104    fn check_tx_buf_len(tx_buf: &[u8]) -> Result<()> {
105        if tx_buf.len() < NET_HDR_SIZE {
106            warn!("Transmit buffer len {} is too small", tx_buf.len());
107            Err(Error::InvalidParam)
108        } else {
109            Ok(())
110        }
111    }
112
113    /// Fill the header of the `buffer` with [`VirtioNetHdr`].
114    ///
115    /// If the `buffer` is not large enough, it returns [`Error::InvalidParam`].
116    pub fn fill_buffer_header(&self, buffer: &mut [u8]) -> Result<usize> {
117        if buffer.len() < NET_HDR_SIZE {
118            return Err(Error::InvalidParam);
119        }
120        let header = VirtioNetHdr::default();
121        buffer[..NET_HDR_SIZE].copy_from_slice(header.as_bytes());
122        Ok(NET_HDR_SIZE)
123    }
124
125    /// Submits a request to transmit a buffer immediately without waiting for
126    /// the transmission to complete.
127    ///
128    /// It will submit request to the VirtIO net device and return a token
129    /// identifying the position of the first descriptor in the chain. If there
130    /// are not enough descriptors to allocate, then it returns
131    /// [`Error::QueueFull`].
132    ///
133    /// The caller needs to fill the `tx_buf` with a header by calling
134    /// [`fill_buffer_header`] before transmission. Then it calls [`poll_transmit`]
135    /// with the returned token to check whether the device has finished handling
136    /// the request. Once it has, the caller must call [`transmit_complete`] with
137    /// the same buffer before reading the result (transmitted length).
138    ///
139    /// # Safety
140    ///
141    /// `tx_buf` is still borrowed by the underlying VirtIO net device even after
142    /// this method returns. Thus, it is the caller's responsibility to guarantee
143    /// that they are not accessed before the request is completed in order to
144    /// avoid data races.
145    ///
146    /// [`fill_buffer_header`]: Self::fill_buffer_header
147    /// [`poll_transmit`]: Self::poll_transmit
148    /// [`transmit_complete`]: Self::transmit_complete
149    pub unsafe fn transmit_begin(&mut self, tx_buf: &[u8]) -> Result<u16> {
150        Self::check_tx_buf_len(tx_buf)?;
151        let token = self.send_queue.add(&[tx_buf], &mut [])?;
152        if self.send_queue.should_notify() {
153            self.transport.notify(QUEUE_TRANSMIT);
154        }
155        Ok(token)
156    }
157
158    /// Fetches the token of the next completed transmission request from the
159    /// used ring and returns it, without removing it from the used ring. If
160    /// there are no pending completed requests it returns [`None`].
161    pub fn poll_transmit(&mut self) -> Option<u16> {
162        self.send_queue.peek_used()
163    }
164
165    /// Completes a transmission operation which was started by [`transmit_begin`].
166    /// Returns number of bytes transmitted.
167    ///
168    /// # Safety
169    ///
170    /// The same buffer must be passed in again as was passed to
171    /// [`transmit_begin`] when it returned the token.
172    ///
173    /// [`transmit_begin`]: Self::transmit_begin
174    pub unsafe fn transmit_complete(&mut self, token: u16, tx_buf: &[u8]) -> Result<usize> {
175        let len = self.send_queue.pop_used(token, &[tx_buf], &mut [])?;
176        Ok(len as usize)
177    }
178
179    /// Submits a request to receive a buffer immediately without waiting for
180    /// the reception to complete.
181    ///
182    /// It will submit request to the VirtIO net device and return a token
183    /// identifying the position of the first descriptor in the chain. If there
184    /// are not enough descriptors to allocate, then it returns
185    /// [`Error::QueueFull`].
186    ///
187    /// The caller can then call [`poll_receive`] with the returned token to
188    /// check whether the device has finished handling the request. Once it has,
189    /// the caller must call [`receive_complete`] with the same buffer before
190    /// reading the response.
191    ///
192    /// # Safety
193    ///
194    /// `rx_buf` is still borrowed by the underlying VirtIO net device even after
195    /// this method returns. Thus, it is the caller's responsibility to guarantee
196    /// that they are not accessed before the request is completed in order to
197    /// avoid data races.
198    ///
199    /// [`poll_receive`]: Self::poll_receive
200    /// [`receive_complete`]: Self::receive_complete
201    pub unsafe fn receive_begin(&mut self, rx_buf: &mut [u8]) -> Result<u16> {
202        Self::check_rx_buf_len(rx_buf)?;
203        let token = self.recv_queue.add(&[], &mut [rx_buf])?;
204        if self.recv_queue.should_notify() {
205            self.transport.notify(QUEUE_RECEIVE);
206        }
207        Ok(token)
208    }
209
210    /// Fetches the token of the next completed reception request from the
211    /// used ring and returns it, without removing it from the used ring. If
212    /// there are no pending completed requests it returns [`None`].
213    pub fn poll_receive(&self) -> Option<u16> {
214        self.recv_queue.peek_used()
215    }
216
217    /// Completes a transmission operation which was started by [`receive_begin`].
218    ///
219    /// After completion, the `rx_buf` will contain a header followed by the
220    /// received packet. It returns the length of the header and the length of
221    /// the packet.
222    ///
223    /// # Safety
224    ///
225    /// The same buffer must be passed in again as was passed to
226    /// [`receive_begin`] when it returned the token.
227    ///
228    /// [`receive_begin`]: Self::receive_begin
229    pub unsafe fn receive_complete(
230        &mut self,
231        token: u16,
232        rx_buf: &mut [u8],
233    ) -> Result<(usize, usize)> {
234        let len = self.recv_queue.pop_used(token, &[], &mut [rx_buf])? as usize;
235        let packet_len = len.checked_sub(NET_HDR_SIZE).ok_or(Error::IoError)?;
236        Ok((NET_HDR_SIZE, packet_len))
237    }
238
239    /// Sends a packet to the network, and blocks until the request completed.
240    pub fn send(&mut self, tx_buf: &[u8]) -> Result {
241        let header = VirtioNetHdr::default();
242        if tx_buf.is_empty() {
243            // Special case sending an empty packet, to avoid adding an empty buffer to the
244            // virtqueue.
245            self.send_queue.add_notify_wait_pop(
246                &[header.as_bytes()],
247                &mut [],
248                &mut self.transport,
249            )?;
250        } else {
251            self.send_queue.add_notify_wait_pop(
252                &[header.as_bytes(), tx_buf],
253                &mut [],
254                &mut self.transport,
255            )?;
256        }
257        Ok(())
258    }
259
260    /// Blocks and waits for a packet to be received.
261    ///
262    /// After completion, the `rx_buf` will contain a header followed by the
263    /// received packet. It returns the length of the header and the length of
264    /// the packet.
265    pub fn receive_wait(&mut self, rx_buf: &mut [u8]) -> Result<(usize, usize)> {
266        let token = unsafe { self.receive_begin(rx_buf)? };
267        while self.poll_receive().is_none() {
268            core::hint::spin_loop();
269        }
270        unsafe { self.receive_complete(token, rx_buf) }
271    }
272}
273
274impl<H: Hal, T: Transport, const QUEUE_SIZE: usize> Drop for VirtIONetRaw<H, T, QUEUE_SIZE> {
275    fn drop(&mut self) {
276        // Clear any pointers pointing to DMA regions, so the device doesn't try to access them
277        // after they have been freed.
278        self.transport.queue_unset(QUEUE_RECEIVE);
279        self.transport.queue_unset(QUEUE_TRANSMIT);
280    }
281}