mirror of https://gitee.com/openkylin/qemu.git
652 lines
20 KiB
C
652 lines
20 KiB
C
/*
|
|
* QEMU Freescale eTSEC Emulator
|
|
*
|
|
* Copyright (c) 2011-2013 AdaCore
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
* of this software and associated documentation files (the "Software"), to deal
|
|
* in the Software without restriction, including without limitation the rights
|
|
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
* copies of the Software, and to permit persons to whom the Software is
|
|
* furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
|
* THE SOFTWARE.
|
|
*/
|
|
#include "qemu/osdep.h"
|
|
#include "net/checksum.h"
|
|
#include "qemu/log.h"
|
|
#include "etsec.h"
|
|
#include "registers.h"
|
|
|
|
/* #define ETSEC_RING_DEBUG */
|
|
/* #define HEX_DUMP */
|
|
/* #define DEBUG_BD */
|
|
|
|
#ifdef ETSEC_RING_DEBUG
|
|
static const int debug_etsec = 1;
|
|
#else
|
|
static const int debug_etsec;
|
|
#endif
|
|
|
|
#define RING_DEBUG(fmt, ...) do { \
|
|
if (debug_etsec) { \
|
|
qemu_log(fmt , ## __VA_ARGS__); \
|
|
} \
|
|
} while (0)
|
|
|
|
#ifdef DEBUG_BD
|
|
|
|
static void print_tx_bd_flags(uint16_t flags)
|
|
{
|
|
qemu_log(" Ready: %d\n", !!(flags & BD_TX_READY));
|
|
qemu_log(" PAD/CRC: %d\n", !!(flags & BD_TX_PADCRC));
|
|
qemu_log(" Wrap: %d\n", !!(flags & BD_WRAP));
|
|
qemu_log(" Interrupt: %d\n", !!(flags & BD_INTERRUPT));
|
|
qemu_log(" Last in frame: %d\n", !!(flags & BD_LAST));
|
|
qemu_log(" Tx CRC: %d\n", !!(flags & BD_TX_TC));
|
|
qemu_log(" User-defined preamble / defer: %d\n",
|
|
!!(flags & BD_TX_PREDEF));
|
|
qemu_log(" Huge frame enable / Late collision: %d\n",
|
|
!!(flags & BD_TX_HFELC));
|
|
qemu_log(" Control frame / Retransmission Limit: %d\n",
|
|
!!(flags & BD_TX_CFRL));
|
|
qemu_log(" Retry count: %d\n",
|
|
(flags >> BD_TX_RC_OFFSET) & BD_TX_RC_MASK);
|
|
qemu_log(" Underrun / TCP/IP off-load enable: %d\n",
|
|
!!(flags & BD_TX_TOEUN));
|
|
qemu_log(" Truncation: %d\n", !!(flags & BD_TX_TR));
|
|
}
|
|
|
|
static void print_rx_bd_flags(uint16_t flags)
|
|
{
|
|
qemu_log(" Empty: %d\n", !!(flags & BD_RX_EMPTY));
|
|
qemu_log(" Receive software ownership: %d\n", !!(flags & BD_RX_RO1));
|
|
qemu_log(" Wrap: %d\n", !!(flags & BD_WRAP));
|
|
qemu_log(" Interrupt: %d\n", !!(flags & BD_INTERRUPT));
|
|
qemu_log(" Last in frame: %d\n", !!(flags & BD_LAST));
|
|
qemu_log(" First in frame: %d\n", !!(flags & BD_RX_FIRST));
|
|
qemu_log(" Miss: %d\n", !!(flags & BD_RX_MISS));
|
|
qemu_log(" Broadcast: %d\n", !!(flags & BD_RX_BROADCAST));
|
|
qemu_log(" Multicast: %d\n", !!(flags & BD_RX_MULTICAST));
|
|
qemu_log(" Rx frame length violation: %d\n", !!(flags & BD_RX_LG));
|
|
qemu_log(" Rx non-octet aligned frame: %d\n", !!(flags & BD_RX_NO));
|
|
qemu_log(" Short frame: %d\n", !!(flags & BD_RX_SH));
|
|
qemu_log(" Rx CRC Error: %d\n", !!(flags & BD_RX_CR));
|
|
qemu_log(" Overrun: %d\n", !!(flags & BD_RX_OV));
|
|
qemu_log(" Truncation: %d\n", !!(flags & BD_RX_TR));
|
|
}
|
|
|
|
|
|
static void print_bd(eTSEC_rxtx_bd bd, int mode, uint32_t index)
|
|
{
|
|
qemu_log("eTSEC %s Data Buffer Descriptor (%u)\n",
|
|
mode == eTSEC_TRANSMIT ? "Transmit" : "Receive",
|
|
index);
|
|
qemu_log(" Flags : 0x%04x\n", bd.flags);
|
|
if (mode == eTSEC_TRANSMIT) {
|
|
print_tx_bd_flags(bd.flags);
|
|
} else {
|
|
print_rx_bd_flags(bd.flags);
|
|
}
|
|
qemu_log(" Length : 0x%04x\n", bd.length);
|
|
qemu_log(" Pointer : 0x%08x\n", bd.bufptr);
|
|
}
|
|
|
|
#endif /* DEBUG_BD */
|
|
|
|
static void read_buffer_descriptor(eTSEC *etsec,
|
|
hwaddr addr,
|
|
eTSEC_rxtx_bd *bd)
|
|
{
|
|
assert(bd != NULL);
|
|
|
|
RING_DEBUG("READ Buffer Descriptor @ 0x" TARGET_FMT_plx"\n", addr);
|
|
cpu_physical_memory_read(addr,
|
|
bd,
|
|
sizeof(eTSEC_rxtx_bd));
|
|
|
|
if (etsec->regs[DMACTRL].value & DMACTRL_LE) {
|
|
bd->flags = lduw_le_p(&bd->flags);
|
|
bd->length = lduw_le_p(&bd->length);
|
|
bd->bufptr = ldl_le_p(&bd->bufptr);
|
|
} else {
|
|
bd->flags = lduw_be_p(&bd->flags);
|
|
bd->length = lduw_be_p(&bd->length);
|
|
bd->bufptr = ldl_be_p(&bd->bufptr);
|
|
}
|
|
}
|
|
|
|
static void write_buffer_descriptor(eTSEC *etsec,
|
|
hwaddr addr,
|
|
eTSEC_rxtx_bd *bd)
|
|
{
|
|
assert(bd != NULL);
|
|
|
|
if (etsec->regs[DMACTRL].value & DMACTRL_LE) {
|
|
stw_le_p(&bd->flags, bd->flags);
|
|
stw_le_p(&bd->length, bd->length);
|
|
stl_le_p(&bd->bufptr, bd->bufptr);
|
|
} else {
|
|
stw_be_p(&bd->flags, bd->flags);
|
|
stw_be_p(&bd->length, bd->length);
|
|
stl_be_p(&bd->bufptr, bd->bufptr);
|
|
}
|
|
|
|
RING_DEBUG("Write Buffer Descriptor @ 0x" TARGET_FMT_plx"\n", addr);
|
|
cpu_physical_memory_write(addr,
|
|
bd,
|
|
sizeof(eTSEC_rxtx_bd));
|
|
}
|
|
|
|
static void ievent_set(eTSEC *etsec,
|
|
uint32_t flags)
|
|
{
|
|
etsec->regs[IEVENT].value |= flags;
|
|
|
|
etsec_update_irq(etsec);
|
|
}
|
|
|
|
static void tx_padding_and_crc(eTSEC *etsec, uint32_t min_frame_len)
|
|
{
|
|
int add = min_frame_len - etsec->tx_buffer_len;
|
|
|
|
/* Padding */
|
|
if (add > 0) {
|
|
RING_DEBUG("pad:%u\n", add);
|
|
etsec->tx_buffer = g_realloc(etsec->tx_buffer,
|
|
etsec->tx_buffer_len + add);
|
|
|
|
memset(etsec->tx_buffer + etsec->tx_buffer_len, 0x0, add);
|
|
etsec->tx_buffer_len += add;
|
|
}
|
|
|
|
/* Never add CRC in QEMU */
|
|
}
|
|
|
|
static void process_tx_fcb(eTSEC *etsec)
|
|
{
|
|
uint8_t flags = (uint8_t)(*etsec->tx_buffer);
|
|
/* L3 header offset from start of frame */
|
|
uint8_t l3_header_offset = (uint8_t)*(etsec->tx_buffer + 3);
|
|
/* L4 header offset from start of L3 header */
|
|
uint8_t l4_header_offset = (uint8_t)*(etsec->tx_buffer + 2);
|
|
/* L3 header */
|
|
uint8_t *l3_header = etsec->tx_buffer + 8 + l3_header_offset;
|
|
/* L4 header */
|
|
uint8_t *l4_header = l3_header + l4_header_offset;
|
|
|
|
/* if packet is IP4 and IP checksum is requested */
|
|
if (flags & FCB_TX_IP && flags & FCB_TX_CIP) {
|
|
/* do IP4 checksum (TODO This function does TCP/UDP checksum
|
|
* but not sure if it also does IP4 checksum.) */
|
|
net_checksum_calculate(etsec->tx_buffer + 8,
|
|
etsec->tx_buffer_len - 8);
|
|
}
|
|
/* TODO Check the correct usage of the PHCS field of the FCB in case the NPH
|
|
* flag is on */
|
|
|
|
/* if packet is IP4 and TCP or UDP */
|
|
if (flags & FCB_TX_IP && flags & FCB_TX_TUP) {
|
|
/* if UDP */
|
|
if (flags & FCB_TX_UDP) {
|
|
/* if checksum is requested */
|
|
if (flags & FCB_TX_CTU) {
|
|
/* do UDP checksum */
|
|
|
|
net_checksum_calculate(etsec->tx_buffer + 8,
|
|
etsec->tx_buffer_len - 8);
|
|
} else {
|
|
/* set checksum field to 0 */
|
|
l4_header[6] = 0;
|
|
l4_header[7] = 0;
|
|
}
|
|
} else if (flags & FCB_TX_CTU) { /* if TCP and checksum is requested */
|
|
/* do TCP checksum */
|
|
net_checksum_calculate(etsec->tx_buffer + 8,
|
|
etsec->tx_buffer_len - 8);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void process_tx_bd(eTSEC *etsec,
|
|
eTSEC_rxtx_bd *bd)
|
|
{
|
|
uint8_t *tmp_buff = NULL;
|
|
hwaddr tbdbth = (hwaddr)(etsec->regs[TBDBPH].value & 0xF) << 32;
|
|
|
|
if (bd->length == 0) {
|
|
/* ERROR */
|
|
return;
|
|
}
|
|
|
|
if (etsec->tx_buffer_len == 0) {
|
|
/* It's the first BD */
|
|
etsec->first_bd = *bd;
|
|
}
|
|
|
|
/* TODO: if TxBD[TOE/UN] skip the Tx Frame Control Block*/
|
|
|
|
/* Load this Data Buffer */
|
|
etsec->tx_buffer = g_realloc(etsec->tx_buffer,
|
|
etsec->tx_buffer_len + bd->length);
|
|
tmp_buff = etsec->tx_buffer + etsec->tx_buffer_len;
|
|
cpu_physical_memory_read(bd->bufptr + tbdbth, tmp_buff, bd->length);
|
|
|
|
/* Update buffer length */
|
|
etsec->tx_buffer_len += bd->length;
|
|
|
|
|
|
if (etsec->tx_buffer_len != 0 && (bd->flags & BD_LAST)) {
|
|
if (etsec->regs[MACCFG1].value & MACCFG1_TX_EN) {
|
|
/* MAC Transmit enabled */
|
|
|
|
/* Process offload Tx FCB */
|
|
if (etsec->first_bd.flags & BD_TX_TOEUN) {
|
|
process_tx_fcb(etsec);
|
|
}
|
|
|
|
if (etsec->first_bd.flags & BD_TX_PADCRC
|
|
|| etsec->regs[MACCFG2].value & MACCFG2_PADCRC) {
|
|
|
|
/* Padding and CRC (Padding implies CRC) */
|
|
tx_padding_and_crc(etsec, 64);
|
|
|
|
} else if (etsec->first_bd.flags & BD_TX_TC
|
|
|| etsec->regs[MACCFG2].value & MACCFG2_CRC_EN) {
|
|
|
|
/* Only CRC */
|
|
/* Never add CRC in QEMU */
|
|
}
|
|
|
|
#if defined(HEX_DUMP)
|
|
qemu_log("eTSEC Send packet size:%d\n", etsec->tx_buffer_len);
|
|
qemu_hexdump(etsec->tx_buffer, stderr, "", etsec->tx_buffer_len);
|
|
#endif /* ETSEC_RING_DEBUG */
|
|
|
|
if (etsec->first_bd.flags & BD_TX_TOEUN) {
|
|
qemu_send_packet(qemu_get_queue(etsec->nic),
|
|
etsec->tx_buffer + 8,
|
|
etsec->tx_buffer_len - 8);
|
|
} else {
|
|
qemu_send_packet(qemu_get_queue(etsec->nic),
|
|
etsec->tx_buffer,
|
|
etsec->tx_buffer_len);
|
|
}
|
|
|
|
}
|
|
|
|
etsec->tx_buffer_len = 0;
|
|
|
|
if (bd->flags & BD_INTERRUPT) {
|
|
ievent_set(etsec, IEVENT_TXF);
|
|
}
|
|
} else {
|
|
if (bd->flags & BD_INTERRUPT) {
|
|
ievent_set(etsec, IEVENT_TXB);
|
|
}
|
|
}
|
|
|
|
/* Update DB flags */
|
|
|
|
/* Clear Ready */
|
|
bd->flags &= ~BD_TX_READY;
|
|
|
|
/* Clear Defer */
|
|
bd->flags &= ~BD_TX_PREDEF;
|
|
|
|
/* Clear Late Collision */
|
|
bd->flags &= ~BD_TX_HFELC;
|
|
|
|
/* Clear Retransmission Limit */
|
|
bd->flags &= ~BD_TX_CFRL;
|
|
|
|
/* Clear Retry Count */
|
|
bd->flags &= ~(BD_TX_RC_MASK << BD_TX_RC_OFFSET);
|
|
|
|
/* Clear Underrun */
|
|
bd->flags &= ~BD_TX_TOEUN;
|
|
|
|
/* Clear Truncation */
|
|
bd->flags &= ~BD_TX_TR;
|
|
}
|
|
|
|
void etsec_walk_tx_ring(eTSEC *etsec, int ring_nbr)
|
|
{
|
|
hwaddr ring_base = 0;
|
|
hwaddr bd_addr = 0;
|
|
eTSEC_rxtx_bd bd;
|
|
uint16_t bd_flags;
|
|
|
|
if (!(etsec->regs[MACCFG1].value & MACCFG1_TX_EN)) {
|
|
RING_DEBUG("%s: MAC Transmit not enabled\n", __func__);
|
|
return;
|
|
}
|
|
|
|
ring_base = (hwaddr)(etsec->regs[TBASEH].value & 0xF) << 32;
|
|
ring_base += etsec->regs[TBASE0 + ring_nbr].value & ~0x7;
|
|
bd_addr = etsec->regs[TBPTR0 + ring_nbr].value & ~0x7;
|
|
|
|
do {
|
|
read_buffer_descriptor(etsec, bd_addr, &bd);
|
|
|
|
#ifdef DEBUG_BD
|
|
print_bd(bd,
|
|
eTSEC_TRANSMIT,
|
|
(bd_addr - ring_base) / sizeof(eTSEC_rxtx_bd));
|
|
|
|
#endif /* DEBUG_BD */
|
|
|
|
/* Save flags before BD update */
|
|
bd_flags = bd.flags;
|
|
|
|
if (!(bd_flags & BD_TX_READY)) {
|
|
break;
|
|
}
|
|
|
|
process_tx_bd(etsec, &bd);
|
|
/* Write back BD after update */
|
|
write_buffer_descriptor(etsec, bd_addr, &bd);
|
|
|
|
/* Wrap or next BD */
|
|
if (bd_flags & BD_WRAP) {
|
|
bd_addr = ring_base;
|
|
} else {
|
|
bd_addr += sizeof(eTSEC_rxtx_bd);
|
|
}
|
|
} while (TRUE);
|
|
|
|
/* Save the Buffer Descriptor Pointers to last bd that was not
|
|
* succesfully closed */
|
|
etsec->regs[TBPTR0 + ring_nbr].value = bd_addr;
|
|
|
|
/* Set transmit halt THLTx */
|
|
etsec->regs[TSTAT].value |= 1 << (31 - ring_nbr);
|
|
}
|
|
|
|
static void fill_rx_bd(eTSEC *etsec,
|
|
eTSEC_rxtx_bd *bd,
|
|
const uint8_t **buf,
|
|
size_t *size)
|
|
{
|
|
uint16_t to_write;
|
|
hwaddr bufptr = bd->bufptr +
|
|
((hwaddr)(etsec->regs[TBDBPH].value & 0xF) << 32);
|
|
uint8_t padd[etsec->rx_padding];
|
|
uint8_t rem;
|
|
|
|
RING_DEBUG("eTSEC fill Rx buffer @ 0x%016" HWADDR_PRIx
|
|
" size:%zu(padding + crc:%u) + fcb:%u\n",
|
|
bufptr, *size, etsec->rx_padding, etsec->rx_fcb_size);
|
|
|
|
bd->length = 0;
|
|
|
|
/* This operation will only write FCB */
|
|
if (etsec->rx_fcb_size != 0) {
|
|
|
|
cpu_physical_memory_write(bufptr, etsec->rx_fcb, etsec->rx_fcb_size);
|
|
|
|
bufptr += etsec->rx_fcb_size;
|
|
bd->length += etsec->rx_fcb_size;
|
|
etsec->rx_fcb_size = 0;
|
|
|
|
}
|
|
|
|
/* We remove padding from the computation of to_write because it is not
|
|
* allocated in the buffer.
|
|
*/
|
|
to_write = MIN(*size - etsec->rx_padding,
|
|
etsec->regs[MRBLR].value - etsec->rx_fcb_size);
|
|
|
|
/* This operation can only write packet data and no padding */
|
|
if (to_write > 0) {
|
|
cpu_physical_memory_write(bufptr, *buf, to_write);
|
|
|
|
*buf += to_write;
|
|
bufptr += to_write;
|
|
*size -= to_write;
|
|
|
|
bd->flags &= ~BD_RX_EMPTY;
|
|
bd->length += to_write;
|
|
}
|
|
|
|
if (*size == etsec->rx_padding) {
|
|
/* The remaining bytes are only for padding which is not actually
|
|
* allocated in the data buffer.
|
|
*/
|
|
|
|
rem = MIN(etsec->regs[MRBLR].value - bd->length, etsec->rx_padding);
|
|
|
|
if (rem > 0) {
|
|
memset(padd, 0x0, sizeof(padd));
|
|
etsec->rx_padding -= rem;
|
|
*size -= rem;
|
|
bd->length += rem;
|
|
cpu_physical_memory_write(bufptr, padd, rem);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void rx_init_frame(eTSEC *etsec, const uint8_t *buf, size_t size)
|
|
{
|
|
uint32_t fcb_size = 0;
|
|
uint8_t prsdep = (etsec->regs[RCTRL].value >> RCTRL_PRSDEP_OFFSET)
|
|
& RCTRL_PRSDEP_MASK;
|
|
|
|
if (prsdep != 0) {
|
|
/* Prepend FCB (FCB size + RCTRL[PAL]) */
|
|
fcb_size = 8 + ((etsec->regs[RCTRL].value >> 16) & 0x1F);
|
|
|
|
etsec->rx_fcb_size = fcb_size;
|
|
|
|
/* TODO: fill_FCB(etsec); */
|
|
memset(etsec->rx_fcb, 0x0, sizeof(etsec->rx_fcb));
|
|
|
|
} else {
|
|
etsec->rx_fcb_size = 0;
|
|
}
|
|
|
|
g_free(etsec->rx_buffer);
|
|
|
|
/* Do not copy the frame for now */
|
|
etsec->rx_buffer = (uint8_t *)buf;
|
|
etsec->rx_buffer_len = size;
|
|
|
|
/* CRC padding (We don't have to compute the CRC) */
|
|
etsec->rx_padding = 4;
|
|
|
|
/*
|
|
* Ensure that payload length + CRC length is at least 802.3
|
|
* minimum MTU size bytes long (64)
|
|
*/
|
|
if (etsec->rx_buffer_len < 60) {
|
|
etsec->rx_padding += 60 - etsec->rx_buffer_len;
|
|
}
|
|
|
|
etsec->rx_first_in_frame = 1;
|
|
etsec->rx_remaining_data = etsec->rx_buffer_len;
|
|
RING_DEBUG("%s: rx_buffer_len:%u rx_padding+crc:%u\n", __func__,
|
|
etsec->rx_buffer_len, etsec->rx_padding);
|
|
}
|
|
|
|
ssize_t etsec_rx_ring_write(eTSEC *etsec, const uint8_t *buf, size_t size)
|
|
{
|
|
int ring_nbr = 0; /* Always use ring0 (no filer) */
|
|
|
|
if (etsec->rx_buffer_len != 0) {
|
|
RING_DEBUG("%s: We can't receive now,"
|
|
" a buffer is already in the pipe\n", __func__);
|
|
return 0;
|
|
}
|
|
|
|
if (etsec->regs[RSTAT].value & 1 << (23 - ring_nbr)) {
|
|
RING_DEBUG("%s: The ring is halted\n", __func__);
|
|
return -1;
|
|
}
|
|
|
|
if (etsec->regs[DMACTRL].value & DMACTRL_GRS) {
|
|
RING_DEBUG("%s: Graceful receive stop\n", __func__);
|
|
return -1;
|
|
}
|
|
|
|
if (!(etsec->regs[MACCFG1].value & MACCFG1_RX_EN)) {
|
|
RING_DEBUG("%s: MAC Receive not enabled\n", __func__);
|
|
return -1;
|
|
}
|
|
|
|
if ((etsec->regs[RCTRL].value & RCTRL_RSF) && (size < 60)) {
|
|
/* CRC is not in the packet yet, so short frame is below 60 bytes */
|
|
RING_DEBUG("%s: Drop short frame\n", __func__);
|
|
return -1;
|
|
}
|
|
|
|
rx_init_frame(etsec, buf, size);
|
|
|
|
etsec_walk_rx_ring(etsec, ring_nbr);
|
|
|
|
return size;
|
|
}
|
|
|
|
void etsec_walk_rx_ring(eTSEC *etsec, int ring_nbr)
|
|
{
|
|
hwaddr ring_base = 0;
|
|
hwaddr bd_addr = 0;
|
|
hwaddr start_bd_addr = 0;
|
|
eTSEC_rxtx_bd bd;
|
|
uint16_t bd_flags;
|
|
size_t remaining_data;
|
|
const uint8_t *buf;
|
|
uint8_t *tmp_buf;
|
|
size_t size;
|
|
|
|
if (etsec->rx_buffer_len == 0) {
|
|
/* No frame to send */
|
|
RING_DEBUG("No frame to send\n");
|
|
return;
|
|
}
|
|
|
|
remaining_data = etsec->rx_remaining_data + etsec->rx_padding;
|
|
buf = etsec->rx_buffer
|
|
+ (etsec->rx_buffer_len - etsec->rx_remaining_data);
|
|
size = etsec->rx_buffer_len + etsec->rx_padding;
|
|
|
|
ring_base = (hwaddr)(etsec->regs[RBASEH].value & 0xF) << 32;
|
|
ring_base += etsec->regs[RBASE0 + ring_nbr].value & ~0x7;
|
|
start_bd_addr = bd_addr = etsec->regs[RBPTR0 + ring_nbr].value & ~0x7;
|
|
|
|
do {
|
|
read_buffer_descriptor(etsec, bd_addr, &bd);
|
|
|
|
#ifdef DEBUG_BD
|
|
print_bd(bd,
|
|
eTSEC_RECEIVE,
|
|
(bd_addr - ring_base) / sizeof(eTSEC_rxtx_bd));
|
|
|
|
#endif /* DEBUG_BD */
|
|
|
|
/* Save flags before BD update */
|
|
bd_flags = bd.flags;
|
|
|
|
if (bd_flags & BD_RX_EMPTY) {
|
|
fill_rx_bd(etsec, &bd, &buf, &remaining_data);
|
|
|
|
if (etsec->rx_first_in_frame) {
|
|
bd.flags |= BD_RX_FIRST;
|
|
etsec->rx_first_in_frame = 0;
|
|
etsec->rx_first_bd = bd;
|
|
}
|
|
|
|
/* Last in frame */
|
|
if (remaining_data == 0) {
|
|
|
|
/* Clear flags */
|
|
|
|
bd.flags &= ~0x7ff;
|
|
|
|
bd.flags |= BD_LAST;
|
|
|
|
/* NOTE: non-octet aligned frame is impossible in qemu */
|
|
|
|
if (size >= etsec->regs[MAXFRM].value) {
|
|
/* frame length violation */
|
|
qemu_log("%s frame length violation: size:%zu MAXFRM:%d\n",
|
|
__func__, size, etsec->regs[MAXFRM].value);
|
|
|
|
bd.flags |= BD_RX_LG;
|
|
}
|
|
|
|
if (size < 64) {
|
|
/* Short frame */
|
|
bd.flags |= BD_RX_SH;
|
|
}
|
|
|
|
/* TODO: Broadcast and Multicast */
|
|
|
|
if (bd.flags & BD_INTERRUPT) {
|
|
/* Set RXFx */
|
|
etsec->regs[RSTAT].value |= 1 << (7 - ring_nbr);
|
|
|
|
/* Set IEVENT */
|
|
ievent_set(etsec, IEVENT_RXF);
|
|
}
|
|
|
|
} else {
|
|
if (bd.flags & BD_INTERRUPT) {
|
|
/* Set IEVENT */
|
|
ievent_set(etsec, IEVENT_RXB);
|
|
}
|
|
}
|
|
|
|
/* Write back BD after update */
|
|
write_buffer_descriptor(etsec, bd_addr, &bd);
|
|
}
|
|
|
|
/* Wrap or next BD */
|
|
if (bd_flags & BD_WRAP) {
|
|
bd_addr = ring_base;
|
|
} else {
|
|
bd_addr += sizeof(eTSEC_rxtx_bd);
|
|
}
|
|
} while (remaining_data != 0
|
|
&& (bd_flags & BD_RX_EMPTY)
|
|
&& bd_addr != start_bd_addr);
|
|
|
|
/* Reset ring ptr */
|
|
etsec->regs[RBPTR0 + ring_nbr].value = bd_addr;
|
|
|
|
/* The frame is too large to fit in the Rx ring */
|
|
if (remaining_data > 0) {
|
|
|
|
/* Set RSTAT[QHLTx] */
|
|
etsec->regs[RSTAT].value |= 1 << (23 - ring_nbr);
|
|
|
|
/* Save remaining data to send the end of the frame when the ring will
|
|
* be restarted
|
|
*/
|
|
etsec->rx_remaining_data = remaining_data;
|
|
|
|
/* Copy the frame */
|
|
tmp_buf = g_malloc(size);
|
|
memcpy(tmp_buf, etsec->rx_buffer, size);
|
|
etsec->rx_buffer = tmp_buf;
|
|
|
|
RING_DEBUG("no empty RxBD available any more\n");
|
|
} else {
|
|
etsec->rx_buffer_len = 0;
|
|
etsec->rx_buffer = NULL;
|
|
if (etsec->need_flush) {
|
|
qemu_flush_queued_packets(qemu_get_queue(etsec->nic));
|
|
}
|
|
}
|
|
|
|
RING_DEBUG("eTSEC End of ring_write: remaining_data:%zu\n", remaining_data);
|
|
}
|