stm32/eth_v2: update to new embassy-net trait, remove PeripheralMutex.
This commit is contained in:
@ -1,19 +1,10 @@
|
||||
use core::sync::atomic::{fence, Ordering};
|
||||
|
||||
use embassy_net::{Packet, PacketBox, PacketBoxExt, PacketBuf};
|
||||
use vcell::VolatileCell;
|
||||
|
||||
use crate::eth::{Packet, RX_BUFFER_SIZE, TX_BUFFER_SIZE};
|
||||
use crate::pac::ETH;
|
||||
|
||||
#[non_exhaustive]
|
||||
#[derive(Debug, Copy, Clone)]
|
||||
#[cfg_attr(feature = "defmt", derive(defmt::Format))]
|
||||
pub enum Error {
|
||||
NoBufferAvailable,
|
||||
// TODO: Break down this error into several others
|
||||
TransmissionError,
|
||||
}
|
||||
|
||||
/// Transmit and Receive Descriptor fields
|
||||
#[allow(dead_code)]
|
||||
mod emac_consts {
|
||||
@ -41,7 +32,7 @@ use emac_consts::*;
|
||||
/// * tdes2: buffer lengths
|
||||
/// * tdes3: control and payload/frame length
|
||||
#[repr(C)]
|
||||
struct TDes {
|
||||
pub(crate) struct TDes {
|
||||
tdes0: VolatileCell<u32>,
|
||||
tdes1: VolatileCell<u32>,
|
||||
tdes2: VolatileCell<u32>,
|
||||
@ -59,41 +50,26 @@ impl TDes {
|
||||
}
|
||||
|
||||
/// Return true if this TDes is not currently owned by the DMA
|
||||
pub fn available(&self) -> bool {
|
||||
fn available(&self) -> bool {
|
||||
self.tdes3.get() & EMAC_DES3_OWN == 0
|
||||
}
|
||||
}
|
||||
|
||||
pub(crate) struct TDesRing<const N: usize> {
|
||||
td: [TDes; N],
|
||||
buffers: [Option<PacketBuf>; N],
|
||||
tdidx: usize,
|
||||
pub(crate) struct TDesRing<'a> {
|
||||
descriptors: &'a mut [TDes],
|
||||
buffers: &'a mut [Packet<TX_BUFFER_SIZE>],
|
||||
index: usize,
|
||||
}
|
||||
|
||||
impl<const N: usize> TDesRing<N> {
|
||||
pub const fn new() -> Self {
|
||||
const TDES: TDes = TDes::new();
|
||||
const BUFFERS: Option<PacketBuf> = None;
|
||||
impl<'a> TDesRing<'a> {
|
||||
/// Initialise this TDesRing. Assume TDesRing is corrupt.
|
||||
pub fn new(descriptors: &'a mut [TDes], buffers: &'a mut [Packet<TX_BUFFER_SIZE>]) -> Self {
|
||||
assert!(descriptors.len() > 0);
|
||||
assert!(descriptors.len() == buffers.len());
|
||||
|
||||
Self {
|
||||
td: [TDES; N],
|
||||
buffers: [BUFFERS; N],
|
||||
tdidx: 0,
|
||||
}
|
||||
}
|
||||
|
||||
/// Initialise this TDesRing. Assume TDesRing is corrupt
|
||||
///
|
||||
/// The current memory address of the buffers inside this TDesRing
|
||||
/// will be stored in the descriptors, so ensure the TDesRing is
|
||||
/// not moved after initialisation.
|
||||
pub(crate) fn init(&mut self) {
|
||||
assert!(N > 0);
|
||||
|
||||
for td in self.td.iter_mut() {
|
||||
for td in descriptors.iter_mut() {
|
||||
*td = TDes::new();
|
||||
}
|
||||
self.tdidx = 0;
|
||||
|
||||
// Initialize the pointers in the DMA engine. (There will be a memory barrier later
|
||||
// before the DMA engine is enabled.)
|
||||
@ -101,80 +77,60 @@ impl<const N: usize> TDesRing<N> {
|
||||
unsafe {
|
||||
let dma = ETH.ethernet_dma();
|
||||
|
||||
dma.dmactx_dlar().write(|w| w.0 = &self.td as *const _ as u32);
|
||||
dma.dmactx_rlr().write(|w| w.set_tdrl((N as u16) - 1));
|
||||
dma.dmactx_dtpr().write(|w| w.0 = &self.td[0] as *const _ as u32);
|
||||
dma.dmactx_dlar().write(|w| w.0 = descriptors.as_mut_ptr() as u32);
|
||||
dma.dmactx_rlr().write(|w| w.set_tdrl((descriptors.len() as u16) - 1));
|
||||
dma.dmactx_dtpr().write(|w| w.0 = 0);
|
||||
}
|
||||
|
||||
Self {
|
||||
descriptors,
|
||||
buffers,
|
||||
index: 0,
|
||||
}
|
||||
}
|
||||
|
||||
/// Return true if a TDes is available for use
|
||||
pub(crate) fn available(&self) -> bool {
|
||||
self.td[self.tdidx].available()
|
||||
pub(crate) fn len(&self) -> usize {
|
||||
self.descriptors.len()
|
||||
}
|
||||
|
||||
pub(crate) fn transmit(&mut self, pkt: PacketBuf) -> Result<(), Error> {
|
||||
if !self.available() {
|
||||
return Err(Error::NoBufferAvailable);
|
||||
/// Return the next available packet buffer for transmitting, or None
|
||||
pub(crate) fn available(&mut self) -> Option<&mut [u8]> {
|
||||
let d = &mut self.descriptors[self.index];
|
||||
if d.available() {
|
||||
Some(&mut self.buffers[self.index].0)
|
||||
} else {
|
||||
None
|
||||
}
|
||||
let x = self.tdidx;
|
||||
let td = &mut self.td[x];
|
||||
}
|
||||
|
||||
let pkt_len = pkt.len();
|
||||
assert!(pkt_len as u32 <= EMAC_TDES2_B1L);
|
||||
let address = pkt.as_ptr() as u32;
|
||||
/// Transmit the packet written in a buffer returned by `available`.
|
||||
pub(crate) fn transmit(&mut self, len: usize) {
|
||||
let td = &mut self.descriptors[self.index];
|
||||
assert!(td.available());
|
||||
assert!(len as u32 <= EMAC_TDES2_B1L);
|
||||
|
||||
// Read format
|
||||
td.tdes0.set(address);
|
||||
td.tdes2.set(pkt_len as u32 & EMAC_TDES2_B1L | EMAC_TDES2_IOC);
|
||||
td.tdes0.set(self.buffers[self.index].0.as_ptr() as u32);
|
||||
td.tdes2.set(len as u32 & EMAC_TDES2_B1L | EMAC_TDES2_IOC);
|
||||
|
||||
// FD: Contains first buffer of packet
|
||||
// LD: Contains last buffer of packet
|
||||
// Give the DMA engine ownership
|
||||
td.tdes3.set(EMAC_DES3_FD | EMAC_DES3_LD | EMAC_DES3_OWN);
|
||||
|
||||
self.buffers[x].replace(pkt);
|
||||
|
||||
// Ensure changes to the descriptor are committed before DMA engine sees tail pointer store.
|
||||
// This will generate an DMB instruction.
|
||||
// "Preceding reads and writes cannot be moved past subsequent writes."
|
||||
fence(Ordering::Release);
|
||||
|
||||
// Move the tail pointer (TPR) to the next descriptor
|
||||
let x = (x + 1) % N;
|
||||
self.index = self.index + 1;
|
||||
if self.index == self.descriptors.len() {
|
||||
self.index = 0;
|
||||
}
|
||||
|
||||
// signal DMA it can try again.
|
||||
// NOTE(unsafe) Atomic write
|
||||
unsafe {
|
||||
ETH.ethernet_dma()
|
||||
.dmactx_dtpr()
|
||||
.write(|w| w.0 = &self.td[x] as *const _ as u32);
|
||||
}
|
||||
self.tdidx = x;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
pub(crate) fn on_interrupt(&mut self) -> Result<(), Error> {
|
||||
let previous = (self.tdidx + N - 1) % N;
|
||||
let td = &self.td[previous];
|
||||
|
||||
// DMB to ensure that we are reading an updated value, probably not needed at the hardware
|
||||
// level, but this is also a hint to the compiler that we're syncing on the buffer.
|
||||
fence(Ordering::SeqCst);
|
||||
|
||||
let tdes3 = td.tdes3.get();
|
||||
|
||||
if tdes3 & EMAC_DES3_OWN != 0 {
|
||||
// Transmission isn't done yet, probably a receive interrupt that fired this
|
||||
return Ok(());
|
||||
}
|
||||
assert!(tdes3 & EMAC_DES3_CTXT == 0);
|
||||
|
||||
// Release the buffer
|
||||
self.buffers[previous].take();
|
||||
|
||||
if tdes3 & EMAC_DES3_ES != 0 {
|
||||
Err(Error::TransmissionError)
|
||||
} else {
|
||||
Ok(())
|
||||
}
|
||||
unsafe { ETH.ethernet_dma().dmactx_dtpr().write(|w| w.0 = 0) }
|
||||
}
|
||||
}
|
||||
|
||||
@ -185,7 +141,7 @@ impl<const N: usize> TDesRing<N> {
|
||||
/// * rdes2:
|
||||
/// * rdes3: OWN and Status
|
||||
#[repr(C)]
|
||||
struct RDes {
|
||||
pub(crate) struct RDes {
|
||||
rdes0: VolatileCell<u32>,
|
||||
rdes1: VolatileCell<u32>,
|
||||
rdes2: VolatileCell<u32>,
|
||||
@ -204,7 +160,7 @@ impl RDes {
|
||||
|
||||
/// Return true if this RDes is acceptable to us
|
||||
#[inline(always)]
|
||||
pub fn valid(&self) -> bool {
|
||||
fn valid(&self) -> bool {
|
||||
// Write-back descriptor is valid if:
|
||||
//
|
||||
// Contains first buffer of packet AND contains last buf of
|
||||
@ -215,177 +171,96 @@ impl RDes {
|
||||
|
||||
/// Return true if this RDes is not currently owned by the DMA
|
||||
#[inline(always)]
|
||||
pub fn available(&self) -> bool {
|
||||
fn available(&self) -> bool {
|
||||
self.rdes3.get() & EMAC_DES3_OWN == 0 // Owned by us
|
||||
}
|
||||
|
||||
#[inline(always)]
|
||||
pub fn set_ready(&mut self, buf_addr: u32) {
|
||||
self.rdes0.set(buf_addr);
|
||||
fn set_ready(&mut self, buf: *mut u8) {
|
||||
self.rdes0.set(buf as u32);
|
||||
self.rdes3.set(EMAC_RDES3_BUF1V | EMAC_RDES3_IOC | EMAC_DES3_OWN);
|
||||
}
|
||||
}
|
||||
|
||||
/// Rx ring of descriptors and packets
|
||||
///
|
||||
/// This ring has three major locations that work in lock-step. The DMA will never write to the tail
|
||||
/// index, so the `read_index` must never pass the tail index. The `next_tail_index` is always 1
|
||||
/// slot ahead of the real tail index, and it must never pass the `read_index` or it could overwrite
|
||||
/// a packet still to be passed to the application.
|
||||
///
|
||||
/// nt can't pass r (no alloc)
|
||||
/// +---+---+---+---+ Read ok +---+---+---+---+ No Read +---+---+---+---+
|
||||
/// | | | | | ------------> | | | | | ------------> | | | | |
|
||||
/// +---+---+---+---+ Allocation ok +---+---+---+---+ +---+---+---+---+
|
||||
/// ^ ^t ^t ^ ^t ^
|
||||
/// |r |r |r
|
||||
/// |nt |nt |nt
|
||||
///
|
||||
///
|
||||
/// +---+---+---+---+ Read ok +---+---+---+---+ Can't read +---+---+---+---+
|
||||
/// | | | | | ------------> | | | | | ------------> | | | | |
|
||||
/// +---+---+---+---+ Allocation fail +---+---+---+---+ Allocation ok +---+---+---+---+
|
||||
/// ^ ^t ^ ^t ^ ^ ^ ^t
|
||||
/// |r | |r | | |r
|
||||
/// |nt |nt |nt
|
||||
///
|
||||
pub(crate) struct RDesRing<const N: usize> {
|
||||
rd: [RDes; N],
|
||||
buffers: [Option<PacketBox>; N],
|
||||
read_idx: usize,
|
||||
next_tail_idx: usize,
|
||||
pub(crate) struct RDesRing<'a> {
|
||||
descriptors: &'a mut [RDes],
|
||||
buffers: &'a mut [Packet<RX_BUFFER_SIZE>],
|
||||
index: usize,
|
||||
}
|
||||
|
||||
impl<const N: usize> RDesRing<N> {
|
||||
pub const fn new() -> Self {
|
||||
const RDES: RDes = RDes::new();
|
||||
const BUFFERS: Option<PacketBox> = None;
|
||||
impl<'a> RDesRing<'a> {
|
||||
pub(crate) fn new(descriptors: &'a mut [RDes], buffers: &'a mut [Packet<RX_BUFFER_SIZE>]) -> Self {
|
||||
assert!(descriptors.len() > 1);
|
||||
assert!(descriptors.len() == buffers.len());
|
||||
|
||||
Self {
|
||||
rd: [RDES; N],
|
||||
buffers: [BUFFERS; N],
|
||||
read_idx: 0,
|
||||
next_tail_idx: 0,
|
||||
}
|
||||
}
|
||||
|
||||
pub(crate) fn init(&mut self) {
|
||||
assert!(N > 1);
|
||||
|
||||
for desc in self.rd.iter_mut() {
|
||||
for (i, desc) in descriptors.iter_mut().enumerate() {
|
||||
*desc = RDes::new();
|
||||
desc.set_ready(buffers[i].0.as_mut_ptr());
|
||||
}
|
||||
|
||||
let mut last_index = 0;
|
||||
for (index, buf) in self.buffers.iter_mut().enumerate() {
|
||||
let pkt = match PacketBox::new(Packet::new()) {
|
||||
Some(p) => p,
|
||||
None => {
|
||||
if index == 0 {
|
||||
panic!("Could not allocate at least one buffer for Ethernet receiving");
|
||||
} else {
|
||||
break;
|
||||
}
|
||||
}
|
||||
};
|
||||
let addr = pkt.as_ptr() as u32;
|
||||
*buf = Some(pkt);
|
||||
self.rd[index].set_ready(addr);
|
||||
last_index = index;
|
||||
}
|
||||
self.next_tail_idx = (last_index + 1) % N;
|
||||
|
||||
unsafe {
|
||||
let dma = ETH.ethernet_dma();
|
||||
|
||||
dma.dmacrx_dlar().write(|w| w.0 = self.rd.as_ptr() as u32);
|
||||
dma.dmacrx_rlr().write(|w| w.set_rdrl((N as u16) - 1));
|
||||
dma.dmacrx_dlar().write(|w| w.0 = descriptors.as_mut_ptr() as u32);
|
||||
dma.dmacrx_rlr().write(|w| w.set_rdrl((descriptors.len() as u16) - 1));
|
||||
dma.dmacrx_dtpr().write(|w| w.0 = 0);
|
||||
}
|
||||
|
||||
// We manage to allocate all buffers, set the index to the last one, that means
|
||||
// that the DMA won't consider the last one as ready, because it (unfortunately)
|
||||
// stops at the tail ptr and wraps at the end of the ring, which means that we
|
||||
// can't tell it to stop after the last buffer.
|
||||
let tail_ptr = &self.rd[last_index] as *const _ as u32;
|
||||
fence(Ordering::Release);
|
||||
|
||||
dma.dmacrx_dtpr().write(|w| w.0 = tail_ptr);
|
||||
Self {
|
||||
descriptors,
|
||||
buffers,
|
||||
index: 0,
|
||||
}
|
||||
}
|
||||
|
||||
pub(crate) fn on_interrupt(&mut self) {
|
||||
// XXX: Do we need to do anything here ? Maybe we should try to advance the tail ptr, but it
|
||||
// would soon hit the read ptr anyway, and we will wake smoltcp's stack on the interrupt
|
||||
// which should try to pop a packet...
|
||||
}
|
||||
|
||||
pub(crate) fn pop_packet(&mut self) -> Option<PacketBuf> {
|
||||
/// Get a received packet if any, or None.
|
||||
pub(crate) fn available(&mut self) -> Option<&mut [u8]> {
|
||||
// Not sure if the contents of the write buffer on the M7 can affects reads, so we are using
|
||||
// a DMB here just in case, it also serves as a hint to the compiler that we're syncing the
|
||||
// buffer (I think .-.)
|
||||
fence(Ordering::SeqCst);
|
||||
|
||||
let read_available = self.rd[self.read_idx].available();
|
||||
let tail_index = (self.next_tail_idx + N - 1) % N;
|
||||
|
||||
let pkt = if read_available && self.read_idx != tail_index {
|
||||
let pkt = self.buffers[self.read_idx].take();
|
||||
let len = (self.rd[self.read_idx].rdes3.get() & EMAC_RDES3_PKTLEN) as usize;
|
||||
|
||||
assert!(pkt.is_some());
|
||||
let valid = self.rd[self.read_idx].valid();
|
||||
|
||||
self.read_idx = (self.read_idx + 1) % N;
|
||||
if valid {
|
||||
pkt.map(|p| p.slice(0..len))
|
||||
} else {
|
||||
None
|
||||
// We might have to process many packets, in case some have been rx'd but are invalid.
|
||||
loop {
|
||||
let descriptor = &mut self.descriptors[self.index];
|
||||
if !descriptor.available() {
|
||||
return None;
|
||||
}
|
||||
} else {
|
||||
None
|
||||
};
|
||||
|
||||
// Try to advance the tail_idx
|
||||
if self.next_tail_idx != self.read_idx {
|
||||
match PacketBox::new(Packet::new()) {
|
||||
Some(b) => {
|
||||
let addr = b.as_ptr() as u32;
|
||||
self.buffers[self.next_tail_idx].replace(b);
|
||||
self.rd[self.next_tail_idx].set_ready(addr);
|
||||
|
||||
// "Preceding reads and writes cannot be moved past subsequent writes."
|
||||
fence(Ordering::Release);
|
||||
|
||||
// NOTE(unsafe) atomic write
|
||||
unsafe {
|
||||
ETH.ethernet_dma()
|
||||
.dmacrx_dtpr()
|
||||
.write(|w| w.0 = &self.rd[self.next_tail_idx] as *const _ as u32);
|
||||
}
|
||||
|
||||
self.next_tail_idx = (self.next_tail_idx + 1) % N;
|
||||
}
|
||||
None => {}
|
||||
// If packet is invalid, pop it and try again.
|
||||
if !descriptor.valid() {
|
||||
warn!("invalid packet: {:08x}", descriptor.rdes0.get());
|
||||
self.pop_packet();
|
||||
continue;
|
||||
}
|
||||
|
||||
break;
|
||||
}
|
||||
pkt
|
||||
|
||||
let descriptor = &mut self.descriptors[self.index];
|
||||
let len = (descriptor.rdes3.get() & EMAC_RDES3_PKTLEN) as usize;
|
||||
return Some(&mut self.buffers[self.index].0[..len]);
|
||||
}
|
||||
}
|
||||
|
||||
pub struct DescriptorRing<const T: usize, const R: usize> {
|
||||
pub(crate) tx: TDesRing<T>,
|
||||
pub(crate) rx: RDesRing<R>,
|
||||
}
|
||||
/// Pop the packet previously returned by `available`.
|
||||
pub(crate) fn pop_packet(&mut self) {
|
||||
let descriptor = &mut self.descriptors[self.index];
|
||||
assert!(descriptor.available());
|
||||
|
||||
impl<const T: usize, const R: usize> DescriptorRing<T, R> {
|
||||
pub const fn new() -> Self {
|
||||
Self {
|
||||
tx: TDesRing::new(),
|
||||
rx: RDesRing::new(),
|
||||
self.descriptors[self.index].set_ready(self.buffers[self.index].0.as_mut_ptr());
|
||||
|
||||
// "Preceding reads and writes cannot be moved past subsequent writes."
|
||||
fence(Ordering::Release);
|
||||
|
||||
// signal DMA it can try again.
|
||||
// NOTE(unsafe) Atomic write
|
||||
unsafe { ETH.ethernet_dma().dmacrx_dtpr().write(|w| w.0 = 0) }
|
||||
|
||||
// Increment index.
|
||||
self.index += 1;
|
||||
if self.index == self.descriptors.len() {
|
||||
self.index = 0
|
||||
}
|
||||
}
|
||||
|
||||
pub fn init(&mut self) {
|
||||
self.tx.init();
|
||||
self.rx.init();
|
||||
}
|
||||
}
|
||||
|
Reference in New Issue
Block a user