1
0
mirror of https://git.FreeBSD.org/src.git synced 2024-12-21 11:13:30 +00:00
freebsd/sys/dev/hme/if_hmevar.h
2010-03-03 17:55:51 +00:00

159 lines
5.3 KiB
C

/*-
* Copyright (c) 1999 The NetBSD Foundation, Inc.
* All rights reserved.
*
* This code is derived from software contributed to The NetBSD Foundation
* by Paul Kranenburg.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*
* from: NetBSD: hmevar.h,v 1.5 2000/06/25 01:10:04 eeh Exp
*
* $FreeBSD$
*/
#include <sys/callout.h>
/*
* Number of receive and transmit descriptors. For each receive descriptor,
* an mbuf cluster is allocated and set up to receive a packet, and a dma map
* is created. Therefore, this number should not be too high to not waste
* memory.
* TX descriptors have no static cost, except for the memory directly allocated
* for them. TX queue elements (the number of which is fixed by HME_NTXQ) hold
* the software state for a transmit job; each has a dmamap allocated for it.
* There may be multiple descriptors allocated to a single queue element.
* HME_NTXQ and HME_NTXSEGS are completely arbitrary.
*/
#define HME_NRXDESC 128
#define HME_NTXDESC 256
#define HME_NTXQ 64
#define HME_NTXSEGS 16
/* Maximum size of a mapped RX buffer. */
#define HME_BUFSZ 1600
/*
* RX DMA descriptor. The descriptors are preallocated; the dma map is
* reused.
*/
struct hme_rxdesc {
struct mbuf *hrx_m;
bus_dmamap_t hrx_dmamap;
};
/* Lazily leave at least one burst size grace space. */
#define HME_DESC_RXLEN(sc, d) \
ulmin(HME_BUFSZ, (d)->hrx_m->m_len - (sc)->sc_burst)
struct hme_txdesc {
struct mbuf *htx_m;
bus_dmamap_t htx_dmamap;
int htx_lastdesc;
STAILQ_ENTRY(hme_txdesc) htx_q;
};
STAILQ_HEAD(hme_txdq, hme_txdesc);
struct hme_ring {
/* Ring Descriptors */
caddr_t rb_membase; /* Packet buffer: CPU address */
bus_addr_t rb_dmabase; /* Packet buffer: DMA address */
caddr_t rb_txd; /* Transmit descriptors */
bus_addr_t rb_txddma; /* DMA address of same */
caddr_t rb_rxd; /* Receive descriptors */
bus_addr_t rb_rxddma; /* DMA address of same */
/* Ring Descriptor state */
int rb_tdhead, rb_tdtail;
int rb_rdtail;
int rb_td_nbusy;
/* Descriptors */
struct hme_rxdesc rb_rxdesc[HME_NRXDESC];
struct hme_txdesc rb_txdesc[HME_NTXQ];
struct hme_txdq rb_txfreeq;
struct hme_txdq rb_txbusyq;
bus_dmamap_t rb_spare_dmamap;
};
struct hme_softc {
struct ifnet *sc_ifp;
struct ifmedia sc_ifmedia;
device_t sc_dev;
device_t sc_miibus;
struct mii_data *sc_mii; /* MII media control */
u_char sc_enaddr[ETHER_ADDR_LEN];
struct callout sc_tick_ch; /* tick callout */
int sc_wdog_timer; /* watchdog timer */
/* The following bus handles are to be provided by the bus front-end */
bus_dma_tag_t sc_pdmatag; /* bus dma parent tag */
bus_dma_tag_t sc_cdmatag; /* control bus dma tag */
bus_dmamap_t sc_cdmamap; /* control bus dma handle */
bus_dma_tag_t sc_rdmatag; /* RX bus dma tag */
bus_dma_tag_t sc_tdmatag; /* RX bus dma tag */
bus_space_handle_t sc_sebh; /* HME Global registers */
bus_space_handle_t sc_erxh; /* HME ERX registers */
bus_space_handle_t sc_etxh; /* HME ETX registers */
bus_space_handle_t sc_mach; /* HME MAC registers */
bus_space_handle_t sc_mifh; /* HME MIF registers */
bus_space_tag_t sc_sebt; /* HME Global registers */
bus_space_tag_t sc_erxt; /* HME ERX registers */
bus_space_tag_t sc_etxt; /* HME ETX registers */
bus_space_tag_t sc_mact; /* HME MAC registers */
bus_space_tag_t sc_mift; /* HME MIF registers */
int sc_burst; /* DVMA burst size in effect */
int sc_phys[2]; /* MII instance -> PHY map */
u_int sc_flags;
#define HME_LINK (1 << 0) /* link is up */
#define HME_PCI (1 << 1) /* PCI busses are little-endian */
int sc_ifflags;
int sc_csum_features;
/* Ring descriptor */
struct hme_ring sc_rb;
struct mtx sc_lock;
};
#define HME_LOCK(_sc) mtx_lock(&(_sc)->sc_lock)
#define HME_UNLOCK(_sc) mtx_unlock(&(_sc)->sc_lock)
#define HME_LOCK_ASSERT(_sc, _what) mtx_assert(&(_sc)->sc_lock, (_what))
extern devclass_t hme_devclass;
int hme_config(struct hme_softc *);
void hme_detach(struct hme_softc *);
void hme_suspend(struct hme_softc *);
void hme_resume(struct hme_softc *);
void hme_intr(void *);
/* MII methods & callbacks */
int hme_mii_readreg(device_t, int, int);
int hme_mii_writereg(device_t, int, int, int);
void hme_mii_statchg(device_t);