1566 lines
		
	
	
		
			41 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1566 lines
		
	
	
		
			41 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
// SPDX-License-Identifier: GPL-2.0+
 | 
						|
/*	FDDI network adapter driver for DEC FDDIcontroller 700/700-C devices.
 | 
						|
 *
 | 
						|
 *	Copyright (c) 2018  Maciej W. Rozycki
 | 
						|
 *
 | 
						|
 *	This program is free software; you can redistribute it and/or
 | 
						|
 *	modify it under the terms of the GNU General Public License
 | 
						|
 *	as published by the Free Software Foundation; either version
 | 
						|
 *	2 of the License, or (at your option) any later version.
 | 
						|
 *
 | 
						|
 *	References:
 | 
						|
 *
 | 
						|
 *	Dave Sawyer & Phil Weeks & Frank Itkowsky,
 | 
						|
 *	"DEC FDDIcontroller 700 Port Specification",
 | 
						|
 *	Revision 1.1, Digital Equipment Corporation
 | 
						|
 */
 | 
						|
 | 
						|
/* ------------------------------------------------------------------------- */
 | 
						|
/* FZA configurable parameters.                                              */
 | 
						|
 | 
						|
/* The number of transmit ring descriptors; either 0 for 512 or 1 for 1024.  */
 | 
						|
#define FZA_RING_TX_MODE 0
 | 
						|
 | 
						|
/* The number of receive ring descriptors; from 2 up to 256.  */
 | 
						|
#define FZA_RING_RX_SIZE 256
 | 
						|
 | 
						|
/* End of FZA configurable parameters.  No need to change anything below.    */
 | 
						|
/* ------------------------------------------------------------------------- */
 | 
						|
 | 
						|
#include <linux/delay.h>
 | 
						|
#include <linux/device.h>
 | 
						|
#include <linux/dma-mapping.h>
 | 
						|
#include <linux/init.h>
 | 
						|
#include <linux/interrupt.h>
 | 
						|
#include <linux/io.h>
 | 
						|
#include <linux/io-64-nonatomic-lo-hi.h>
 | 
						|
#include <linux/ioport.h>
 | 
						|
#include <linux/kernel.h>
 | 
						|
#include <linux/list.h>
 | 
						|
#include <linux/module.h>
 | 
						|
#include <linux/netdevice.h>
 | 
						|
#include <linux/fddidevice.h>
 | 
						|
#include <linux/sched.h>
 | 
						|
#include <linux/skbuff.h>
 | 
						|
#include <linux/spinlock.h>
 | 
						|
#include <linux/stat.h>
 | 
						|
#include <linux/tc.h>
 | 
						|
#include <linux/timer.h>
 | 
						|
#include <linux/types.h>
 | 
						|
#include <linux/wait.h>
 | 
						|
 | 
						|
#include <asm/barrier.h>
 | 
						|
 | 
						|
#include "defza.h"
 | 
						|
 | 
						|
#define DRV_NAME "defza"
 | 
						|
#define DRV_VERSION "v.1.1.4"
 | 
						|
#define DRV_RELDATE "Oct  6 2018"
 | 
						|
 | 
						|
static const char version[] =
 | 
						|
	DRV_NAME ": " DRV_VERSION "  " DRV_RELDATE "  Maciej W. Rozycki\n";
 | 
						|
 | 
						|
MODULE_AUTHOR("Maciej W. Rozycki <macro@orcam.me.uk>");
 | 
						|
MODULE_DESCRIPTION("DEC FDDIcontroller 700 (DEFZA-xx) driver");
 | 
						|
MODULE_LICENSE("GPL");
 | 
						|
 | 
						|
static int loopback;
 | 
						|
module_param(loopback, int, 0644);
 | 
						|
 | 
						|
/* Ring Purger Multicast */
 | 
						|
static u8 hw_addr_purger[8] = { 0x09, 0x00, 0x2b, 0x02, 0x01, 0x05 };
 | 
						|
/* Directed Beacon Multicast */
 | 
						|
static u8 hw_addr_beacon[8] = { 0x01, 0x80, 0xc2, 0x00, 0x01, 0x00 };
 | 
						|
 | 
						|
/* Shorthands for MMIO accesses that we require to be strongly ordered
 | 
						|
 * WRT preceding MMIO accesses.
 | 
						|
 */
 | 
						|
#define readw_o readw_relaxed
 | 
						|
#define readl_o readl_relaxed
 | 
						|
 | 
						|
#define writew_o writew_relaxed
 | 
						|
#define writel_o writel_relaxed
 | 
						|
 | 
						|
/* Shorthands for MMIO accesses that we are happy with being weakly ordered
 | 
						|
 * WRT preceding MMIO accesses.
 | 
						|
 */
 | 
						|
#define readw_u readw_relaxed
 | 
						|
#define readl_u readl_relaxed
 | 
						|
#define readq_u readq_relaxed
 | 
						|
 | 
						|
#define writew_u writew_relaxed
 | 
						|
#define writel_u writel_relaxed
 | 
						|
#define writeq_u writeq_relaxed
 | 
						|
 | 
						|
static inline struct sk_buff *fza_alloc_skb_irq(struct net_device *dev,
 | 
						|
						unsigned int length)
 | 
						|
{
 | 
						|
	return __netdev_alloc_skb(dev, length, GFP_ATOMIC);
 | 
						|
}
 | 
						|
 | 
						|
static inline struct sk_buff *fza_alloc_skb(struct net_device *dev,
 | 
						|
					    unsigned int length)
 | 
						|
{
 | 
						|
	return __netdev_alloc_skb(dev, length, GFP_KERNEL);
 | 
						|
}
 | 
						|
 | 
						|
static inline void fza_skb_align(struct sk_buff *skb, unsigned int v)
 | 
						|
{
 | 
						|
	unsigned long x, y;
 | 
						|
 | 
						|
	x = (unsigned long)skb->data;
 | 
						|
	y = ALIGN(x, v);
 | 
						|
 | 
						|
	skb_reserve(skb, y - x);
 | 
						|
}
 | 
						|
 | 
						|
static inline void fza_reads(const void __iomem *from, void *to,
 | 
						|
			     unsigned long size)
 | 
						|
{
 | 
						|
	if (sizeof(unsigned long) == 8) {
 | 
						|
		const u64 __iomem *src = from;
 | 
						|
		const u32 __iomem *src_trail;
 | 
						|
		u64 *dst = to;
 | 
						|
		u32 *dst_trail;
 | 
						|
 | 
						|
		for (size = (size + 3) / 4; size > 1; size -= 2)
 | 
						|
			*dst++ = readq_u(src++);
 | 
						|
		if (size) {
 | 
						|
			src_trail = (u32 __iomem *)src;
 | 
						|
			dst_trail = (u32 *)dst;
 | 
						|
			*dst_trail = readl_u(src_trail);
 | 
						|
		}
 | 
						|
	} else {
 | 
						|
		const u32 __iomem *src = from;
 | 
						|
		u32 *dst = to;
 | 
						|
 | 
						|
		for (size = (size + 3) / 4; size; size--)
 | 
						|
			*dst++ = readl_u(src++);
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static inline void fza_writes(const void *from, void __iomem *to,
 | 
						|
			      unsigned long size)
 | 
						|
{
 | 
						|
	if (sizeof(unsigned long) == 8) {
 | 
						|
		const u64 *src = from;
 | 
						|
		const u32 *src_trail;
 | 
						|
		u64 __iomem *dst = to;
 | 
						|
		u32 __iomem *dst_trail;
 | 
						|
 | 
						|
		for (size = (size + 3) / 4; size > 1; size -= 2)
 | 
						|
			writeq_u(*src++, dst++);
 | 
						|
		if (size) {
 | 
						|
			src_trail = (u32 *)src;
 | 
						|
			dst_trail = (u32 __iomem *)dst;
 | 
						|
			writel_u(*src_trail, dst_trail);
 | 
						|
		}
 | 
						|
	} else {
 | 
						|
		const u32 *src = from;
 | 
						|
		u32 __iomem *dst = to;
 | 
						|
 | 
						|
		for (size = (size + 3) / 4; size; size--)
 | 
						|
			writel_u(*src++, dst++);
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static inline void fza_moves(const void __iomem *from, void __iomem *to,
 | 
						|
			     unsigned long size)
 | 
						|
{
 | 
						|
	if (sizeof(unsigned long) == 8) {
 | 
						|
		const u64 __iomem *src = from;
 | 
						|
		const u32 __iomem *src_trail;
 | 
						|
		u64 __iomem *dst = to;
 | 
						|
		u32 __iomem *dst_trail;
 | 
						|
 | 
						|
		for (size = (size + 3) / 4; size > 1; size -= 2)
 | 
						|
			writeq_u(readq_u(src++), dst++);
 | 
						|
		if (size) {
 | 
						|
			src_trail = (u32 __iomem *)src;
 | 
						|
			dst_trail = (u32 __iomem *)dst;
 | 
						|
			writel_u(readl_u(src_trail), dst_trail);
 | 
						|
		}
 | 
						|
	} else {
 | 
						|
		const u32 __iomem *src = from;
 | 
						|
		u32 __iomem *dst = to;
 | 
						|
 | 
						|
		for (size = (size + 3) / 4; size; size--)
 | 
						|
			writel_u(readl_u(src++), dst++);
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static inline void fza_zeros(void __iomem *to, unsigned long size)
 | 
						|
{
 | 
						|
	if (sizeof(unsigned long) == 8) {
 | 
						|
		u64 __iomem *dst = to;
 | 
						|
		u32 __iomem *dst_trail;
 | 
						|
 | 
						|
		for (size = (size + 3) / 4; size > 1; size -= 2)
 | 
						|
			writeq_u(0, dst++);
 | 
						|
		if (size) {
 | 
						|
			dst_trail = (u32 __iomem *)dst;
 | 
						|
			writel_u(0, dst_trail);
 | 
						|
		}
 | 
						|
	} else {
 | 
						|
		u32 __iomem *dst = to;
 | 
						|
 | 
						|
		for (size = (size + 3) / 4; size; size--)
 | 
						|
			writel_u(0, dst++);
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static inline void fza_regs_dump(struct fza_private *fp)
 | 
						|
{
 | 
						|
	pr_debug("%s: iomem registers:\n", fp->name);
 | 
						|
	pr_debug(" reset:           0x%04x\n", readw_o(&fp->regs->reset));
 | 
						|
	pr_debug(" interrupt event: 0x%04x\n", readw_u(&fp->regs->int_event));
 | 
						|
	pr_debug(" status:          0x%04x\n", readw_u(&fp->regs->status));
 | 
						|
	pr_debug(" interrupt mask:  0x%04x\n", readw_u(&fp->regs->int_mask));
 | 
						|
	pr_debug(" control A:       0x%04x\n", readw_u(&fp->regs->control_a));
 | 
						|
	pr_debug(" control B:       0x%04x\n", readw_u(&fp->regs->control_b));
 | 
						|
}
 | 
						|
 | 
						|
static inline void fza_do_reset(struct fza_private *fp)
 | 
						|
{
 | 
						|
	/* Reset the board. */
 | 
						|
	writew_o(FZA_RESET_INIT, &fp->regs->reset);
 | 
						|
	readw_o(&fp->regs->reset);	/* Synchronize. */
 | 
						|
	readw_o(&fp->regs->reset);	/* Read it back for a small delay. */
 | 
						|
	writew_o(FZA_RESET_CLR, &fp->regs->reset);
 | 
						|
 | 
						|
	/* Enable all interrupt events we handle. */
 | 
						|
	writew_o(fp->int_mask, &fp->regs->int_mask);
 | 
						|
	readw_o(&fp->regs->int_mask);	/* Synchronize. */
 | 
						|
}
 | 
						|
 | 
						|
static inline void fza_do_shutdown(struct fza_private *fp)
 | 
						|
{
 | 
						|
	/* Disable the driver mode. */
 | 
						|
	writew_o(FZA_CONTROL_B_IDLE, &fp->regs->control_b);
 | 
						|
 | 
						|
	/* And reset the board. */
 | 
						|
	writew_o(FZA_RESET_INIT, &fp->regs->reset);
 | 
						|
	readw_o(&fp->regs->reset);	/* Synchronize. */
 | 
						|
	writew_o(FZA_RESET_CLR, &fp->regs->reset);
 | 
						|
	readw_o(&fp->regs->reset);	/* Synchronize. */
 | 
						|
}
 | 
						|
 | 
						|
static int fza_reset(struct fza_private *fp)
 | 
						|
{
 | 
						|
	unsigned long flags;
 | 
						|
	uint status, state;
 | 
						|
	long t;
 | 
						|
 | 
						|
	pr_info("%s: resetting the board...\n", fp->name);
 | 
						|
 | 
						|
	spin_lock_irqsave(&fp->lock, flags);
 | 
						|
	fp->state_chg_flag = 0;
 | 
						|
	fza_do_reset(fp);
 | 
						|
	spin_unlock_irqrestore(&fp->lock, flags);
 | 
						|
 | 
						|
	/* DEC says RESET needs up to 30 seconds to complete.  My DEFZA-AA
 | 
						|
	 * rev. C03 happily finishes in 9.7 seconds. :-)  But we need to
 | 
						|
	 * be on the safe side...
 | 
						|
	 */
 | 
						|
	t = wait_event_timeout(fp->state_chg_wait, fp->state_chg_flag,
 | 
						|
			       45 * HZ);
 | 
						|
	status = readw_u(&fp->regs->status);
 | 
						|
	state = FZA_STATUS_GET_STATE(status);
 | 
						|
	if (fp->state_chg_flag == 0) {
 | 
						|
		pr_err("%s: RESET timed out!, state %x\n", fp->name, state);
 | 
						|
		return -EIO;
 | 
						|
	}
 | 
						|
	if (state != FZA_STATE_UNINITIALIZED) {
 | 
						|
		pr_err("%s: RESET failed!, state %x, failure ID %x\n",
 | 
						|
		       fp->name, state, FZA_STATUS_GET_TEST(status));
 | 
						|
		return -EIO;
 | 
						|
	}
 | 
						|
	pr_info("%s: OK\n", fp->name);
 | 
						|
	pr_debug("%s: RESET: %lums elapsed\n", fp->name,
 | 
						|
		 (45 * HZ - t) * 1000 / HZ);
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static struct fza_ring_cmd __iomem *fza_cmd_send(struct net_device *dev,
 | 
						|
						 int command)
 | 
						|
{
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
	struct fza_ring_cmd __iomem *ring = fp->ring_cmd + fp->ring_cmd_index;
 | 
						|
	unsigned int old_mask, new_mask;
 | 
						|
	union fza_cmd_buf __iomem *buf;
 | 
						|
	struct netdev_hw_addr *ha;
 | 
						|
	int i;
 | 
						|
 | 
						|
	old_mask = fp->int_mask;
 | 
						|
	new_mask = old_mask & ~FZA_MASK_STATE_CHG;
 | 
						|
	writew_u(new_mask, &fp->regs->int_mask);
 | 
						|
	readw_o(&fp->regs->int_mask);			/* Synchronize. */
 | 
						|
	fp->int_mask = new_mask;
 | 
						|
 | 
						|
	buf = fp->mmio + readl_u(&ring->buffer);
 | 
						|
 | 
						|
	if ((readl_u(&ring->cmd_own) & FZA_RING_OWN_MASK) !=
 | 
						|
	    FZA_RING_OWN_HOST) {
 | 
						|
		pr_warn("%s: command buffer full, command: %u!\n", fp->name,
 | 
						|
			command);
 | 
						|
		return NULL;
 | 
						|
	}
 | 
						|
 | 
						|
	switch (command) {
 | 
						|
	case FZA_RING_CMD_INIT:
 | 
						|
		writel_u(FZA_RING_TX_MODE, &buf->init.tx_mode);
 | 
						|
		writel_u(FZA_RING_RX_SIZE, &buf->init.hst_rx_size);
 | 
						|
		fza_zeros(&buf->init.counters, sizeof(buf->init.counters));
 | 
						|
		break;
 | 
						|
 | 
						|
	case FZA_RING_CMD_MODCAM:
 | 
						|
		i = 0;
 | 
						|
		fza_writes(&hw_addr_purger, &buf->cam.hw_addr[i++],
 | 
						|
			   sizeof(*buf->cam.hw_addr));
 | 
						|
		fza_writes(&hw_addr_beacon, &buf->cam.hw_addr[i++],
 | 
						|
			   sizeof(*buf->cam.hw_addr));
 | 
						|
		netdev_for_each_mc_addr(ha, dev) {
 | 
						|
			if (i >= FZA_CMD_CAM_SIZE)
 | 
						|
				break;
 | 
						|
			fza_writes(ha->addr, &buf->cam.hw_addr[i++],
 | 
						|
				   sizeof(*buf->cam.hw_addr));
 | 
						|
		}
 | 
						|
		while (i < FZA_CMD_CAM_SIZE)
 | 
						|
			fza_zeros(&buf->cam.hw_addr[i++],
 | 
						|
				  sizeof(*buf->cam.hw_addr));
 | 
						|
		break;
 | 
						|
 | 
						|
	case FZA_RING_CMD_PARAM:
 | 
						|
		writel_u(loopback, &buf->param.loop_mode);
 | 
						|
		writel_u(fp->t_max, &buf->param.t_max);
 | 
						|
		writel_u(fp->t_req, &buf->param.t_req);
 | 
						|
		writel_u(fp->tvx, &buf->param.tvx);
 | 
						|
		writel_u(fp->lem_threshold, &buf->param.lem_threshold);
 | 
						|
		fza_writes(&fp->station_id, &buf->param.station_id,
 | 
						|
			   sizeof(buf->param.station_id));
 | 
						|
		/* Convert to milliseconds due to buggy firmware. */
 | 
						|
		writel_u(fp->rtoken_timeout / 12500,
 | 
						|
			 &buf->param.rtoken_timeout);
 | 
						|
		writel_u(fp->ring_purger, &buf->param.ring_purger);
 | 
						|
		break;
 | 
						|
 | 
						|
	case FZA_RING_CMD_MODPROM:
 | 
						|
		if (dev->flags & IFF_PROMISC) {
 | 
						|
			writel_u(1, &buf->modprom.llc_prom);
 | 
						|
			writel_u(1, &buf->modprom.smt_prom);
 | 
						|
		} else {
 | 
						|
			writel_u(0, &buf->modprom.llc_prom);
 | 
						|
			writel_u(0, &buf->modprom.smt_prom);
 | 
						|
		}
 | 
						|
		if (dev->flags & IFF_ALLMULTI ||
 | 
						|
		    netdev_mc_count(dev) > FZA_CMD_CAM_SIZE - 2)
 | 
						|
			writel_u(1, &buf->modprom.llc_multi);
 | 
						|
		else
 | 
						|
			writel_u(0, &buf->modprom.llc_multi);
 | 
						|
		writel_u(1, &buf->modprom.llc_bcast);
 | 
						|
		break;
 | 
						|
	}
 | 
						|
 | 
						|
	/* Trigger the command. */
 | 
						|
	writel_u(FZA_RING_OWN_FZA | command, &ring->cmd_own);
 | 
						|
	writew_o(FZA_CONTROL_A_CMD_POLL, &fp->regs->control_a);
 | 
						|
 | 
						|
	fp->ring_cmd_index = (fp->ring_cmd_index + 1) % FZA_RING_CMD_SIZE;
 | 
						|
 | 
						|
	fp->int_mask = old_mask;
 | 
						|
	writew_u(fp->int_mask, &fp->regs->int_mask);
 | 
						|
 | 
						|
	return ring;
 | 
						|
}
 | 
						|
 | 
						|
static int fza_init_send(struct net_device *dev,
 | 
						|
			 struct fza_cmd_init *__iomem *init)
 | 
						|
{
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
	struct fza_ring_cmd __iomem *ring;
 | 
						|
	unsigned long flags;
 | 
						|
	u32 stat;
 | 
						|
	long t;
 | 
						|
 | 
						|
	spin_lock_irqsave(&fp->lock, flags);
 | 
						|
	fp->cmd_done_flag = 0;
 | 
						|
	ring = fza_cmd_send(dev, FZA_RING_CMD_INIT);
 | 
						|
	spin_unlock_irqrestore(&fp->lock, flags);
 | 
						|
	if (!ring)
 | 
						|
		/* This should never happen in the uninitialized state,
 | 
						|
		 * so do not try to recover and just consider it fatal.
 | 
						|
		 */
 | 
						|
		return -ENOBUFS;
 | 
						|
 | 
						|
	/* INIT may take quite a long time (160ms for my C03). */
 | 
						|
	t = wait_event_timeout(fp->cmd_done_wait, fp->cmd_done_flag, 3 * HZ);
 | 
						|
	if (fp->cmd_done_flag == 0) {
 | 
						|
		pr_err("%s: INIT command timed out!, state %x\n", fp->name,
 | 
						|
		       FZA_STATUS_GET_STATE(readw_u(&fp->regs->status)));
 | 
						|
		return -EIO;
 | 
						|
	}
 | 
						|
	stat = readl_u(&ring->stat);
 | 
						|
	if (stat != FZA_RING_STAT_SUCCESS) {
 | 
						|
		pr_err("%s: INIT command failed!, status %02x, state %x\n",
 | 
						|
		       fp->name, stat,
 | 
						|
		       FZA_STATUS_GET_STATE(readw_u(&fp->regs->status)));
 | 
						|
		return -EIO;
 | 
						|
	}
 | 
						|
	pr_debug("%s: INIT: %lums elapsed\n", fp->name,
 | 
						|
		 (3 * HZ - t) * 1000 / HZ);
 | 
						|
 | 
						|
	if (init)
 | 
						|
		*init = fp->mmio + readl_u(&ring->buffer);
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static void fza_rx_init(struct fza_private *fp)
 | 
						|
{
 | 
						|
	int i;
 | 
						|
 | 
						|
	/* Fill the host receive descriptor ring. */
 | 
						|
	for (i = 0; i < FZA_RING_RX_SIZE; i++) {
 | 
						|
		writel_o(0, &fp->ring_hst_rx[i].rmc);
 | 
						|
		writel_o((fp->rx_dma[i] + 0x1000) >> 9,
 | 
						|
			 &fp->ring_hst_rx[i].buffer1);
 | 
						|
		writel_o(fp->rx_dma[i] >> 9 | FZA_RING_OWN_FZA,
 | 
						|
			 &fp->ring_hst_rx[i].buf0_own);
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static void fza_set_rx_mode(struct net_device *dev)
 | 
						|
{
 | 
						|
	fza_cmd_send(dev, FZA_RING_CMD_MODCAM);
 | 
						|
	fza_cmd_send(dev, FZA_RING_CMD_MODPROM);
 | 
						|
}
 | 
						|
 | 
						|
union fza_buffer_txp {
 | 
						|
	struct fza_buffer_tx *data_ptr;
 | 
						|
	struct fza_buffer_tx __iomem *mmio_ptr;
 | 
						|
};
 | 
						|
 | 
						|
static int fza_do_xmit(union fza_buffer_txp ub, int len,
 | 
						|
		       struct net_device *dev, int smt)
 | 
						|
{
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
	struct fza_buffer_tx __iomem *rmc_tx_ptr;
 | 
						|
	int i, first, frag_len, left_len;
 | 
						|
	u32 own, rmc;
 | 
						|
 | 
						|
	if (((((fp->ring_rmc_txd_index - 1 + fp->ring_rmc_tx_size) -
 | 
						|
	       fp->ring_rmc_tx_index) % fp->ring_rmc_tx_size) *
 | 
						|
	     FZA_TX_BUFFER_SIZE) < len)
 | 
						|
		return 1;
 | 
						|
 | 
						|
	first = fp->ring_rmc_tx_index;
 | 
						|
 | 
						|
	left_len = len;
 | 
						|
	frag_len = FZA_TX_BUFFER_SIZE;
 | 
						|
	/* First descriptor is relinquished last. */
 | 
						|
	own = FZA_RING_TX_OWN_HOST;
 | 
						|
	/* First descriptor carries frame length; we don't use cut-through. */
 | 
						|
	rmc = FZA_RING_TX_SOP | FZA_RING_TX_VBC | len;
 | 
						|
	do {
 | 
						|
		i = fp->ring_rmc_tx_index;
 | 
						|
		rmc_tx_ptr = &fp->buffer_tx[i];
 | 
						|
 | 
						|
		if (left_len < FZA_TX_BUFFER_SIZE)
 | 
						|
			frag_len = left_len;
 | 
						|
		left_len -= frag_len;
 | 
						|
 | 
						|
		/* Length must be a multiple of 4 as only word writes are
 | 
						|
		 * permitted!
 | 
						|
		 */
 | 
						|
		frag_len = (frag_len + 3) & ~3;
 | 
						|
		if (smt)
 | 
						|
			fza_moves(ub.mmio_ptr, rmc_tx_ptr, frag_len);
 | 
						|
		else
 | 
						|
			fza_writes(ub.data_ptr, rmc_tx_ptr, frag_len);
 | 
						|
 | 
						|
		if (left_len == 0)
 | 
						|
			rmc |= FZA_RING_TX_EOP;		/* Mark last frag. */
 | 
						|
 | 
						|
		writel_o(rmc, &fp->ring_rmc_tx[i].rmc);
 | 
						|
		writel_o(own, &fp->ring_rmc_tx[i].own);
 | 
						|
 | 
						|
		ub.data_ptr++;
 | 
						|
		fp->ring_rmc_tx_index = (fp->ring_rmc_tx_index + 1) %
 | 
						|
					fp->ring_rmc_tx_size;
 | 
						|
 | 
						|
		/* Settings for intermediate frags. */
 | 
						|
		own = FZA_RING_TX_OWN_RMC;
 | 
						|
		rmc = 0;
 | 
						|
	} while (left_len > 0);
 | 
						|
 | 
						|
	if (((((fp->ring_rmc_txd_index - 1 + fp->ring_rmc_tx_size) -
 | 
						|
	       fp->ring_rmc_tx_index) % fp->ring_rmc_tx_size) *
 | 
						|
	     FZA_TX_BUFFER_SIZE) < dev->mtu + dev->hard_header_len) {
 | 
						|
		netif_stop_queue(dev);
 | 
						|
		pr_debug("%s: queue stopped\n", fp->name);
 | 
						|
	}
 | 
						|
 | 
						|
	writel_o(FZA_RING_TX_OWN_RMC, &fp->ring_rmc_tx[first].own);
 | 
						|
 | 
						|
	/* Go, go, go! */
 | 
						|
	writew_o(FZA_CONTROL_A_TX_POLL, &fp->regs->control_a);
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int fza_do_recv_smt(struct fza_buffer_tx *data_ptr, int len,
 | 
						|
			   u32 rmc, struct net_device *dev)
 | 
						|
{
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
	struct fza_buffer_tx __iomem *smt_rx_ptr;
 | 
						|
	u32 own;
 | 
						|
	int i;
 | 
						|
 | 
						|
	i = fp->ring_smt_rx_index;
 | 
						|
	own = readl_o(&fp->ring_smt_rx[i].own);
 | 
						|
	if ((own & FZA_RING_OWN_MASK) == FZA_RING_OWN_FZA)
 | 
						|
		return 1;
 | 
						|
 | 
						|
	smt_rx_ptr = fp->mmio + readl_u(&fp->ring_smt_rx[i].buffer);
 | 
						|
 | 
						|
	/* Length must be a multiple of 4 as only word writes are permitted! */
 | 
						|
	fza_writes(data_ptr, smt_rx_ptr, (len + 3) & ~3);
 | 
						|
 | 
						|
	writel_o(rmc, &fp->ring_smt_rx[i].rmc);
 | 
						|
	writel_o(FZA_RING_OWN_FZA, &fp->ring_smt_rx[i].own);
 | 
						|
 | 
						|
	fp->ring_smt_rx_index =
 | 
						|
		(fp->ring_smt_rx_index + 1) % fp->ring_smt_rx_size;
 | 
						|
 | 
						|
	/* Grab it! */
 | 
						|
	writew_o(FZA_CONTROL_A_SMT_RX_POLL, &fp->regs->control_a);
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static void fza_tx(struct net_device *dev)
 | 
						|
{
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
	u32 own, rmc;
 | 
						|
	int i;
 | 
						|
 | 
						|
	while (1) {
 | 
						|
		i = fp->ring_rmc_txd_index;
 | 
						|
		if (i == fp->ring_rmc_tx_index)
 | 
						|
			break;
 | 
						|
		own = readl_o(&fp->ring_rmc_tx[i].own);
 | 
						|
		if ((own & FZA_RING_OWN_MASK) == FZA_RING_TX_OWN_RMC)
 | 
						|
			break;
 | 
						|
 | 
						|
		rmc = readl_u(&fp->ring_rmc_tx[i].rmc);
 | 
						|
		/* Only process the first descriptor. */
 | 
						|
		if ((rmc & FZA_RING_TX_SOP) != 0) {
 | 
						|
			if ((rmc & FZA_RING_TX_DCC_MASK) ==
 | 
						|
			    FZA_RING_TX_DCC_SUCCESS) {
 | 
						|
				int pkt_len = (rmc & FZA_RING_PBC_MASK) - 3;
 | 
						|
								/* Omit PRH. */
 | 
						|
 | 
						|
				fp->stats.tx_packets++;
 | 
						|
				fp->stats.tx_bytes += pkt_len;
 | 
						|
			} else {
 | 
						|
				fp->stats.tx_errors++;
 | 
						|
				switch (rmc & FZA_RING_TX_DCC_MASK) {
 | 
						|
				case FZA_RING_TX_DCC_DTP_SOP:
 | 
						|
				case FZA_RING_TX_DCC_DTP:
 | 
						|
				case FZA_RING_TX_DCC_ABORT:
 | 
						|
					fp->stats.tx_aborted_errors++;
 | 
						|
					break;
 | 
						|
				case FZA_RING_TX_DCC_UNDRRUN:
 | 
						|
					fp->stats.tx_fifo_errors++;
 | 
						|
					break;
 | 
						|
				case FZA_RING_TX_DCC_PARITY:
 | 
						|
				default:
 | 
						|
					break;
 | 
						|
				}
 | 
						|
			}
 | 
						|
		}
 | 
						|
 | 
						|
		fp->ring_rmc_txd_index = (fp->ring_rmc_txd_index + 1) %
 | 
						|
					 fp->ring_rmc_tx_size;
 | 
						|
	}
 | 
						|
 | 
						|
	if (((((fp->ring_rmc_txd_index - 1 + fp->ring_rmc_tx_size) -
 | 
						|
	       fp->ring_rmc_tx_index) % fp->ring_rmc_tx_size) *
 | 
						|
	     FZA_TX_BUFFER_SIZE) >= dev->mtu + dev->hard_header_len) {
 | 
						|
		if (fp->queue_active) {
 | 
						|
			netif_wake_queue(dev);
 | 
						|
			pr_debug("%s: queue woken\n", fp->name);
 | 
						|
		}
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static inline int fza_rx_err(struct fza_private *fp,
 | 
						|
			     const u32 rmc, const u8 fc)
 | 
						|
{
 | 
						|
	int len, min_len, max_len;
 | 
						|
 | 
						|
	len = rmc & FZA_RING_PBC_MASK;
 | 
						|
 | 
						|
	if (unlikely((rmc & FZA_RING_RX_BAD) != 0)) {
 | 
						|
		fp->stats.rx_errors++;
 | 
						|
 | 
						|
		/* Check special status codes. */
 | 
						|
		if ((rmc & (FZA_RING_RX_CRC | FZA_RING_RX_RRR_MASK |
 | 
						|
			    FZA_RING_RX_DA_MASK | FZA_RING_RX_SA_MASK)) ==
 | 
						|
		     (FZA_RING_RX_CRC | FZA_RING_RX_RRR_DADDR |
 | 
						|
		      FZA_RING_RX_DA_CAM | FZA_RING_RX_SA_ALIAS)) {
 | 
						|
			if (len >= 8190)
 | 
						|
				fp->stats.rx_length_errors++;
 | 
						|
			return 1;
 | 
						|
		}
 | 
						|
		if ((rmc & (FZA_RING_RX_CRC | FZA_RING_RX_RRR_MASK |
 | 
						|
			    FZA_RING_RX_DA_MASK | FZA_RING_RX_SA_MASK)) ==
 | 
						|
		     (FZA_RING_RX_CRC | FZA_RING_RX_RRR_DADDR |
 | 
						|
		      FZA_RING_RX_DA_CAM | FZA_RING_RX_SA_CAM)) {
 | 
						|
			/* Halt the interface to trigger a reset. */
 | 
						|
			writew_o(FZA_CONTROL_A_HALT, &fp->regs->control_a);
 | 
						|
			readw_o(&fp->regs->control_a);	/* Synchronize. */
 | 
						|
			return 1;
 | 
						|
		}
 | 
						|
 | 
						|
		/* Check the MAC status. */
 | 
						|
		switch (rmc & FZA_RING_RX_RRR_MASK) {
 | 
						|
		case FZA_RING_RX_RRR_OK:
 | 
						|
			if ((rmc & FZA_RING_RX_CRC) != 0)
 | 
						|
				fp->stats.rx_crc_errors++;
 | 
						|
			else if ((rmc & FZA_RING_RX_FSC_MASK) == 0 ||
 | 
						|
				 (rmc & FZA_RING_RX_FSB_ERR) != 0)
 | 
						|
				fp->stats.rx_frame_errors++;
 | 
						|
			return 1;
 | 
						|
		case FZA_RING_RX_RRR_SADDR:
 | 
						|
		case FZA_RING_RX_RRR_DADDR:
 | 
						|
		case FZA_RING_RX_RRR_ABORT:
 | 
						|
			/* Halt the interface to trigger a reset. */
 | 
						|
			writew_o(FZA_CONTROL_A_HALT, &fp->regs->control_a);
 | 
						|
			readw_o(&fp->regs->control_a);	/* Synchronize. */
 | 
						|
			return 1;
 | 
						|
		case FZA_RING_RX_RRR_LENGTH:
 | 
						|
			fp->stats.rx_frame_errors++;
 | 
						|
			return 1;
 | 
						|
		default:
 | 
						|
			return 1;
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	/* Packet received successfully; validate the length. */
 | 
						|
	switch (fc & FDDI_FC_K_FORMAT_MASK) {
 | 
						|
	case FDDI_FC_K_FORMAT_MANAGEMENT:
 | 
						|
		if ((fc & FDDI_FC_K_CLASS_MASK) == FDDI_FC_K_CLASS_ASYNC)
 | 
						|
			min_len = 37;
 | 
						|
		else
 | 
						|
			min_len = 17;
 | 
						|
		break;
 | 
						|
	case FDDI_FC_K_FORMAT_LLC:
 | 
						|
		min_len = 20;
 | 
						|
		break;
 | 
						|
	default:
 | 
						|
		min_len = 17;
 | 
						|
		break;
 | 
						|
	}
 | 
						|
	max_len = 4495;
 | 
						|
	if (len < min_len || len > max_len) {
 | 
						|
		fp->stats.rx_errors++;
 | 
						|
		fp->stats.rx_length_errors++;
 | 
						|
		return 1;
 | 
						|
	}
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static void fza_rx(struct net_device *dev)
 | 
						|
{
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
	struct sk_buff *skb, *newskb;
 | 
						|
	struct fza_fddihdr *frame;
 | 
						|
	dma_addr_t dma, newdma;
 | 
						|
	u32 own, rmc, buf;
 | 
						|
	int i, len;
 | 
						|
	u8 fc;
 | 
						|
 | 
						|
	while (1) {
 | 
						|
		i = fp->ring_hst_rx_index;
 | 
						|
		own = readl_o(&fp->ring_hst_rx[i].buf0_own);
 | 
						|
		if ((own & FZA_RING_OWN_MASK) == FZA_RING_OWN_FZA)
 | 
						|
			break;
 | 
						|
 | 
						|
		rmc = readl_u(&fp->ring_hst_rx[i].rmc);
 | 
						|
		skb = fp->rx_skbuff[i];
 | 
						|
		dma = fp->rx_dma[i];
 | 
						|
 | 
						|
		/* The RMC doesn't count the preamble and the starting
 | 
						|
		 * delimiter.  We fix it up here for a total of 3 octets.
 | 
						|
		 */
 | 
						|
		dma_rmb();
 | 
						|
		len = (rmc & FZA_RING_PBC_MASK) + 3;
 | 
						|
		frame = (struct fza_fddihdr *)skb->data;
 | 
						|
 | 
						|
		/* We need to get at real FC. */
 | 
						|
		dma_sync_single_for_cpu(fp->bdev,
 | 
						|
					dma +
 | 
						|
					((u8 *)&frame->hdr.fc - (u8 *)frame),
 | 
						|
					sizeof(frame->hdr.fc),
 | 
						|
					DMA_FROM_DEVICE);
 | 
						|
		fc = frame->hdr.fc;
 | 
						|
 | 
						|
		if (fza_rx_err(fp, rmc, fc))
 | 
						|
			goto err_rx;
 | 
						|
 | 
						|
		/* We have to 512-byte-align RX buffers... */
 | 
						|
		newskb = fza_alloc_skb_irq(dev, FZA_RX_BUFFER_SIZE + 511);
 | 
						|
		if (newskb) {
 | 
						|
			fza_skb_align(newskb, 512);
 | 
						|
			newdma = dma_map_single(fp->bdev, newskb->data,
 | 
						|
						FZA_RX_BUFFER_SIZE,
 | 
						|
						DMA_FROM_DEVICE);
 | 
						|
			if (dma_mapping_error(fp->bdev, newdma)) {
 | 
						|
				dev_kfree_skb_irq(newskb);
 | 
						|
				newskb = NULL;
 | 
						|
			}
 | 
						|
		}
 | 
						|
		if (newskb) {
 | 
						|
			int pkt_len = len - 7;	/* Omit P, SD and FCS. */
 | 
						|
			int is_multi;
 | 
						|
			int rx_stat;
 | 
						|
 | 
						|
			dma_unmap_single(fp->bdev, dma, FZA_RX_BUFFER_SIZE,
 | 
						|
					 DMA_FROM_DEVICE);
 | 
						|
 | 
						|
			/* Queue SMT frames to the SMT receive ring. */
 | 
						|
			if ((fc & (FDDI_FC_K_CLASS_MASK |
 | 
						|
				   FDDI_FC_K_FORMAT_MASK)) ==
 | 
						|
			     (FDDI_FC_K_CLASS_ASYNC |
 | 
						|
			      FDDI_FC_K_FORMAT_MANAGEMENT) &&
 | 
						|
			    (rmc & FZA_RING_RX_DA_MASK) !=
 | 
						|
			     FZA_RING_RX_DA_PROM) {
 | 
						|
				if (fza_do_recv_smt((struct fza_buffer_tx *)
 | 
						|
						    skb->data, len, rmc,
 | 
						|
						    dev)) {
 | 
						|
					writel_o(FZA_CONTROL_A_SMT_RX_OVFL,
 | 
						|
						 &fp->regs->control_a);
 | 
						|
				}
 | 
						|
			}
 | 
						|
 | 
						|
			is_multi = ((frame->hdr.daddr[0] & 0x01) != 0);
 | 
						|
 | 
						|
			skb_reserve(skb, 3);	/* Skip over P and SD. */
 | 
						|
			skb_put(skb, pkt_len);	/* And cut off FCS. */
 | 
						|
			skb->protocol = fddi_type_trans(skb, dev);
 | 
						|
 | 
						|
			rx_stat = netif_rx(skb);
 | 
						|
			if (rx_stat != NET_RX_DROP) {
 | 
						|
				fp->stats.rx_packets++;
 | 
						|
				fp->stats.rx_bytes += pkt_len;
 | 
						|
				if (is_multi)
 | 
						|
					fp->stats.multicast++;
 | 
						|
			} else {
 | 
						|
				fp->stats.rx_dropped++;
 | 
						|
			}
 | 
						|
 | 
						|
			skb = newskb;
 | 
						|
			dma = newdma;
 | 
						|
			fp->rx_skbuff[i] = skb;
 | 
						|
			fp->rx_dma[i] = dma;
 | 
						|
		} else {
 | 
						|
			fp->stats.rx_dropped++;
 | 
						|
			pr_notice("%s: memory squeeze, dropping packet\n",
 | 
						|
				  fp->name);
 | 
						|
		}
 | 
						|
 | 
						|
err_rx:
 | 
						|
		writel_o(0, &fp->ring_hst_rx[i].rmc);
 | 
						|
		buf = (dma + 0x1000) >> 9;
 | 
						|
		writel_o(buf, &fp->ring_hst_rx[i].buffer1);
 | 
						|
		buf = dma >> 9 | FZA_RING_OWN_FZA;
 | 
						|
		writel_o(buf, &fp->ring_hst_rx[i].buf0_own);
 | 
						|
		fp->ring_hst_rx_index =
 | 
						|
			(fp->ring_hst_rx_index + 1) % fp->ring_hst_rx_size;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static void fza_tx_smt(struct net_device *dev)
 | 
						|
{
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
	struct fza_buffer_tx __iomem *smt_tx_ptr;
 | 
						|
	int i, len;
 | 
						|
	u32 own;
 | 
						|
 | 
						|
	while (1) {
 | 
						|
		i = fp->ring_smt_tx_index;
 | 
						|
		own = readl_o(&fp->ring_smt_tx[i].own);
 | 
						|
		if ((own & FZA_RING_OWN_MASK) == FZA_RING_OWN_FZA)
 | 
						|
			break;
 | 
						|
 | 
						|
		smt_tx_ptr = fp->mmio + readl_u(&fp->ring_smt_tx[i].buffer);
 | 
						|
		len = readl_u(&fp->ring_smt_tx[i].rmc) & FZA_RING_PBC_MASK;
 | 
						|
 | 
						|
		if (!netif_queue_stopped(dev)) {
 | 
						|
			if (dev_nit_active(dev)) {
 | 
						|
				struct fza_buffer_tx *skb_data_ptr;
 | 
						|
				struct sk_buff *skb;
 | 
						|
 | 
						|
				/* Length must be a multiple of 4 as only word
 | 
						|
				 * reads are permitted!
 | 
						|
				 */
 | 
						|
				skb = fza_alloc_skb_irq(dev, (len + 3) & ~3);
 | 
						|
				if (!skb)
 | 
						|
					goto err_no_skb;	/* Drop. */
 | 
						|
 | 
						|
				skb_data_ptr = (struct fza_buffer_tx *)
 | 
						|
					       skb->data;
 | 
						|
 | 
						|
				fza_reads(smt_tx_ptr, skb_data_ptr,
 | 
						|
					  (len + 3) & ~3);
 | 
						|
				skb->dev = dev;
 | 
						|
				skb_reserve(skb, 3);	/* Skip over PRH. */
 | 
						|
				skb_put(skb, len - 3);
 | 
						|
				skb_reset_network_header(skb);
 | 
						|
 | 
						|
				dev_queue_xmit_nit(skb, dev);
 | 
						|
 | 
						|
				dev_kfree_skb_irq(skb);
 | 
						|
 | 
						|
err_no_skb:
 | 
						|
				;
 | 
						|
			}
 | 
						|
 | 
						|
			/* Queue the frame to the RMC transmit ring. */
 | 
						|
			fza_do_xmit((union fza_buffer_txp)
 | 
						|
				    { .mmio_ptr = smt_tx_ptr },
 | 
						|
				    len, dev, 1);
 | 
						|
		}
 | 
						|
 | 
						|
		writel_o(FZA_RING_OWN_FZA, &fp->ring_smt_tx[i].own);
 | 
						|
		fp->ring_smt_tx_index =
 | 
						|
			(fp->ring_smt_tx_index + 1) % fp->ring_smt_tx_size;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static void fza_uns(struct net_device *dev)
 | 
						|
{
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
	u32 own;
 | 
						|
	int i;
 | 
						|
 | 
						|
	while (1) {
 | 
						|
		i = fp->ring_uns_index;
 | 
						|
		own = readl_o(&fp->ring_uns[i].own);
 | 
						|
		if ((own & FZA_RING_OWN_MASK) == FZA_RING_OWN_FZA)
 | 
						|
			break;
 | 
						|
 | 
						|
		if (readl_u(&fp->ring_uns[i].id) == FZA_RING_UNS_RX_OVER) {
 | 
						|
			fp->stats.rx_errors++;
 | 
						|
			fp->stats.rx_over_errors++;
 | 
						|
		}
 | 
						|
 | 
						|
		writel_o(FZA_RING_OWN_FZA, &fp->ring_uns[i].own);
 | 
						|
		fp->ring_uns_index =
 | 
						|
			(fp->ring_uns_index + 1) % FZA_RING_UNS_SIZE;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static void fza_tx_flush(struct net_device *dev)
 | 
						|
{
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
	u32 own;
 | 
						|
	int i;
 | 
						|
 | 
						|
	/* Clean up the SMT TX ring. */
 | 
						|
	i = fp->ring_smt_tx_index;
 | 
						|
	do {
 | 
						|
		writel_o(FZA_RING_OWN_FZA, &fp->ring_smt_tx[i].own);
 | 
						|
		fp->ring_smt_tx_index =
 | 
						|
			(fp->ring_smt_tx_index + 1) % fp->ring_smt_tx_size;
 | 
						|
 | 
						|
	} while (i != fp->ring_smt_tx_index);
 | 
						|
 | 
						|
	/* Clean up the RMC TX ring. */
 | 
						|
	i = fp->ring_rmc_tx_index;
 | 
						|
	do {
 | 
						|
		own = readl_o(&fp->ring_rmc_tx[i].own);
 | 
						|
		if ((own & FZA_RING_OWN_MASK) == FZA_RING_TX_OWN_RMC) {
 | 
						|
			u32 rmc = readl_u(&fp->ring_rmc_tx[i].rmc);
 | 
						|
 | 
						|
			writel_u(rmc | FZA_RING_TX_DTP,
 | 
						|
				 &fp->ring_rmc_tx[i].rmc);
 | 
						|
		}
 | 
						|
		fp->ring_rmc_tx_index =
 | 
						|
			(fp->ring_rmc_tx_index + 1) % fp->ring_rmc_tx_size;
 | 
						|
 | 
						|
	} while (i != fp->ring_rmc_tx_index);
 | 
						|
 | 
						|
	/* Done. */
 | 
						|
	writew_o(FZA_CONTROL_A_FLUSH_DONE, &fp->regs->control_a);
 | 
						|
}
 | 
						|
 | 
						|
static irqreturn_t fza_interrupt(int irq, void *dev_id)
 | 
						|
{
 | 
						|
	struct net_device *dev = dev_id;
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
	uint int_event;
 | 
						|
 | 
						|
	/* Get interrupt events. */
 | 
						|
	int_event = readw_o(&fp->regs->int_event) & fp->int_mask;
 | 
						|
	if (int_event == 0)
 | 
						|
		return IRQ_NONE;
 | 
						|
 | 
						|
	/* Clear the events. */
 | 
						|
	writew_u(int_event, &fp->regs->int_event);
 | 
						|
 | 
						|
	/* Now handle the events.  The order matters. */
 | 
						|
 | 
						|
	/* Command finished interrupt. */
 | 
						|
	if ((int_event & FZA_EVENT_CMD_DONE) != 0) {
 | 
						|
		fp->irq_count_cmd_done++;
 | 
						|
 | 
						|
		spin_lock(&fp->lock);
 | 
						|
		fp->cmd_done_flag = 1;
 | 
						|
		wake_up(&fp->cmd_done_wait);
 | 
						|
		spin_unlock(&fp->lock);
 | 
						|
	}
 | 
						|
 | 
						|
	/* Transmit finished interrupt. */
 | 
						|
	if ((int_event & FZA_EVENT_TX_DONE) != 0) {
 | 
						|
		fp->irq_count_tx_done++;
 | 
						|
		fza_tx(dev);
 | 
						|
	}
 | 
						|
 | 
						|
	/* Host receive interrupt. */
 | 
						|
	if ((int_event & FZA_EVENT_RX_POLL) != 0) {
 | 
						|
		fp->irq_count_rx_poll++;
 | 
						|
		fza_rx(dev);
 | 
						|
	}
 | 
						|
 | 
						|
	/* SMT transmit interrupt. */
 | 
						|
	if ((int_event & FZA_EVENT_SMT_TX_POLL) != 0) {
 | 
						|
		fp->irq_count_smt_tx_poll++;
 | 
						|
		fza_tx_smt(dev);
 | 
						|
	}
 | 
						|
 | 
						|
	/* Transmit ring flush request. */
 | 
						|
	if ((int_event & FZA_EVENT_FLUSH_TX) != 0) {
 | 
						|
		fp->irq_count_flush_tx++;
 | 
						|
		fza_tx_flush(dev);
 | 
						|
	}
 | 
						|
 | 
						|
	/* Link status change interrupt. */
 | 
						|
	if ((int_event & FZA_EVENT_LINK_ST_CHG) != 0) {
 | 
						|
		uint status;
 | 
						|
 | 
						|
		fp->irq_count_link_st_chg++;
 | 
						|
		status = readw_u(&fp->regs->status);
 | 
						|
		if (FZA_STATUS_GET_LINK(status) == FZA_LINK_ON) {
 | 
						|
			netif_carrier_on(dev);
 | 
						|
			pr_info("%s: link available\n", fp->name);
 | 
						|
		} else {
 | 
						|
			netif_carrier_off(dev);
 | 
						|
			pr_info("%s: link unavailable\n", fp->name);
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	/* Unsolicited event interrupt. */
 | 
						|
	if ((int_event & FZA_EVENT_UNS_POLL) != 0) {
 | 
						|
		fp->irq_count_uns_poll++;
 | 
						|
		fza_uns(dev);
 | 
						|
	}
 | 
						|
 | 
						|
	/* State change interrupt. */
 | 
						|
	if ((int_event & FZA_EVENT_STATE_CHG) != 0) {
 | 
						|
		uint status, state;
 | 
						|
 | 
						|
		fp->irq_count_state_chg++;
 | 
						|
 | 
						|
		status = readw_u(&fp->regs->status);
 | 
						|
		state = FZA_STATUS_GET_STATE(status);
 | 
						|
		pr_debug("%s: state change: %x\n", fp->name, state);
 | 
						|
		switch (state) {
 | 
						|
		case FZA_STATE_RESET:
 | 
						|
			break;
 | 
						|
 | 
						|
		case FZA_STATE_UNINITIALIZED:
 | 
						|
			netif_carrier_off(dev);
 | 
						|
			del_timer_sync(&fp->reset_timer);
 | 
						|
			fp->ring_cmd_index = 0;
 | 
						|
			fp->ring_uns_index = 0;
 | 
						|
			fp->ring_rmc_tx_index = 0;
 | 
						|
			fp->ring_rmc_txd_index = 0;
 | 
						|
			fp->ring_hst_rx_index = 0;
 | 
						|
			fp->ring_smt_tx_index = 0;
 | 
						|
			fp->ring_smt_rx_index = 0;
 | 
						|
			if (fp->state > state) {
 | 
						|
				pr_info("%s: OK\n", fp->name);
 | 
						|
				fza_cmd_send(dev, FZA_RING_CMD_INIT);
 | 
						|
			}
 | 
						|
			break;
 | 
						|
 | 
						|
		case FZA_STATE_INITIALIZED:
 | 
						|
			if (fp->state > state) {
 | 
						|
				fza_set_rx_mode(dev);
 | 
						|
				fza_cmd_send(dev, FZA_RING_CMD_PARAM);
 | 
						|
			}
 | 
						|
			break;
 | 
						|
 | 
						|
		case FZA_STATE_RUNNING:
 | 
						|
		case FZA_STATE_MAINTENANCE:
 | 
						|
			fp->state = state;
 | 
						|
			fza_rx_init(fp);
 | 
						|
			fp->queue_active = 1;
 | 
						|
			netif_wake_queue(dev);
 | 
						|
			pr_debug("%s: queue woken\n", fp->name);
 | 
						|
			break;
 | 
						|
 | 
						|
		case FZA_STATE_HALTED:
 | 
						|
			fp->queue_active = 0;
 | 
						|
			netif_stop_queue(dev);
 | 
						|
			pr_debug("%s: queue stopped\n", fp->name);
 | 
						|
			del_timer_sync(&fp->reset_timer);
 | 
						|
			pr_warn("%s: halted, reason: %x\n", fp->name,
 | 
						|
				FZA_STATUS_GET_HALT(status));
 | 
						|
			fza_regs_dump(fp);
 | 
						|
			pr_info("%s: resetting the board...\n", fp->name);
 | 
						|
			fza_do_reset(fp);
 | 
						|
			fp->timer_state = 0;
 | 
						|
			fp->reset_timer.expires = jiffies + 45 * HZ;
 | 
						|
			add_timer(&fp->reset_timer);
 | 
						|
			break;
 | 
						|
 | 
						|
		default:
 | 
						|
			pr_warn("%s: undefined state: %x\n", fp->name, state);
 | 
						|
			break;
 | 
						|
		}
 | 
						|
 | 
						|
		spin_lock(&fp->lock);
 | 
						|
		fp->state_chg_flag = 1;
 | 
						|
		wake_up(&fp->state_chg_wait);
 | 
						|
		spin_unlock(&fp->lock);
 | 
						|
	}
 | 
						|
 | 
						|
	return IRQ_HANDLED;
 | 
						|
}
 | 
						|
 | 
						|
static void fza_reset_timer(struct timer_list *t)
 | 
						|
{
 | 
						|
	struct fza_private *fp = from_timer(fp, t, reset_timer);
 | 
						|
 | 
						|
	if (!fp->timer_state) {
 | 
						|
		pr_err("%s: RESET timed out!\n", fp->name);
 | 
						|
		pr_info("%s: trying harder...\n", fp->name);
 | 
						|
 | 
						|
		/* Assert the board reset. */
 | 
						|
		writew_o(FZA_RESET_INIT, &fp->regs->reset);
 | 
						|
		readw_o(&fp->regs->reset);		/* Synchronize. */
 | 
						|
 | 
						|
		fp->timer_state = 1;
 | 
						|
		fp->reset_timer.expires = jiffies + HZ;
 | 
						|
	} else {
 | 
						|
		/* Clear the board reset. */
 | 
						|
		writew_u(FZA_RESET_CLR, &fp->regs->reset);
 | 
						|
 | 
						|
		/* Enable all interrupt events we handle. */
 | 
						|
		writew_o(fp->int_mask, &fp->regs->int_mask);
 | 
						|
		readw_o(&fp->regs->int_mask);		/* Synchronize. */
 | 
						|
 | 
						|
		fp->timer_state = 0;
 | 
						|
		fp->reset_timer.expires = jiffies + 45 * HZ;
 | 
						|
	}
 | 
						|
	add_timer(&fp->reset_timer);
 | 
						|
}
 | 
						|
 | 
						|
static int fza_set_mac_address(struct net_device *dev, void *addr)
 | 
						|
{
 | 
						|
	return -EOPNOTSUPP;
 | 
						|
}
 | 
						|
 | 
						|
static netdev_tx_t fza_start_xmit(struct sk_buff *skb, struct net_device *dev)
 | 
						|
{
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
	unsigned int old_mask, new_mask;
 | 
						|
	int ret;
 | 
						|
	u8 fc;
 | 
						|
 | 
						|
	skb_push(skb, 3);			/* Make room for PRH. */
 | 
						|
 | 
						|
	/* Decode FC to set PRH. */
 | 
						|
	fc = skb->data[3];
 | 
						|
	skb->data[0] = 0;
 | 
						|
	skb->data[1] = 0;
 | 
						|
	skb->data[2] = FZA_PRH2_NORMAL;
 | 
						|
	if ((fc & FDDI_FC_K_CLASS_MASK) == FDDI_FC_K_CLASS_SYNC)
 | 
						|
		skb->data[0] |= FZA_PRH0_FRAME_SYNC;
 | 
						|
	switch (fc & FDDI_FC_K_FORMAT_MASK) {
 | 
						|
	case FDDI_FC_K_FORMAT_MANAGEMENT:
 | 
						|
		if ((fc & FDDI_FC_K_CONTROL_MASK) == 0) {
 | 
						|
			/* Token. */
 | 
						|
			skb->data[0] |= FZA_PRH0_TKN_TYPE_IMM;
 | 
						|
			skb->data[1] |= FZA_PRH1_TKN_SEND_NONE;
 | 
						|
		} else {
 | 
						|
			/* SMT or MAC. */
 | 
						|
			skb->data[0] |= FZA_PRH0_TKN_TYPE_UNR;
 | 
						|
			skb->data[1] |= FZA_PRH1_TKN_SEND_UNR;
 | 
						|
		}
 | 
						|
		skb->data[1] |= FZA_PRH1_CRC_NORMAL;
 | 
						|
		break;
 | 
						|
	case FDDI_FC_K_FORMAT_LLC:
 | 
						|
	case FDDI_FC_K_FORMAT_FUTURE:
 | 
						|
		skb->data[0] |= FZA_PRH0_TKN_TYPE_UNR;
 | 
						|
		skb->data[1] |= FZA_PRH1_CRC_NORMAL | FZA_PRH1_TKN_SEND_UNR;
 | 
						|
		break;
 | 
						|
	case FDDI_FC_K_FORMAT_IMPLEMENTOR:
 | 
						|
		skb->data[0] |= FZA_PRH0_TKN_TYPE_UNR;
 | 
						|
		skb->data[1] |= FZA_PRH1_TKN_SEND_ORIG;
 | 
						|
		break;
 | 
						|
	}
 | 
						|
 | 
						|
	/* SMT transmit interrupts may sneak frames into the RMC
 | 
						|
	 * transmit ring.  We disable them while queueing a frame
 | 
						|
	 * to maintain consistency.
 | 
						|
	 */
 | 
						|
	old_mask = fp->int_mask;
 | 
						|
	new_mask = old_mask & ~FZA_MASK_SMT_TX_POLL;
 | 
						|
	writew_u(new_mask, &fp->regs->int_mask);
 | 
						|
	readw_o(&fp->regs->int_mask);			/* Synchronize. */
 | 
						|
	fp->int_mask = new_mask;
 | 
						|
	ret = fza_do_xmit((union fza_buffer_txp)
 | 
						|
			  { .data_ptr = (struct fza_buffer_tx *)skb->data },
 | 
						|
			  skb->len, dev, 0);
 | 
						|
	fp->int_mask = old_mask;
 | 
						|
	writew_u(fp->int_mask, &fp->regs->int_mask);
 | 
						|
 | 
						|
	if (ret) {
 | 
						|
		/* Probably an SMT packet filled the remaining space,
 | 
						|
		 * so just stop the queue, but don't report it as an error.
 | 
						|
		 */
 | 
						|
		netif_stop_queue(dev);
 | 
						|
		pr_debug("%s: queue stopped\n", fp->name);
 | 
						|
		fp->stats.tx_dropped++;
 | 
						|
	}
 | 
						|
 | 
						|
	dev_kfree_skb(skb);
 | 
						|
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
static int fza_open(struct net_device *dev)
 | 
						|
{
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
	struct fza_ring_cmd __iomem *ring;
 | 
						|
	struct sk_buff *skb;
 | 
						|
	unsigned long flags;
 | 
						|
	dma_addr_t dma;
 | 
						|
	int ret, i;
 | 
						|
	u32 stat;
 | 
						|
	long t;
 | 
						|
 | 
						|
	for (i = 0; i < FZA_RING_RX_SIZE; i++) {
 | 
						|
		/* We have to 512-byte-align RX buffers... */
 | 
						|
		skb = fza_alloc_skb(dev, FZA_RX_BUFFER_SIZE + 511);
 | 
						|
		if (skb) {
 | 
						|
			fza_skb_align(skb, 512);
 | 
						|
			dma = dma_map_single(fp->bdev, skb->data,
 | 
						|
					     FZA_RX_BUFFER_SIZE,
 | 
						|
					     DMA_FROM_DEVICE);
 | 
						|
			if (dma_mapping_error(fp->bdev, dma)) {
 | 
						|
				dev_kfree_skb(skb);
 | 
						|
				skb = NULL;
 | 
						|
			}
 | 
						|
		}
 | 
						|
		if (!skb) {
 | 
						|
			for (--i; i >= 0; i--) {
 | 
						|
				dma_unmap_single(fp->bdev, fp->rx_dma[i],
 | 
						|
						 FZA_RX_BUFFER_SIZE,
 | 
						|
						 DMA_FROM_DEVICE);
 | 
						|
				dev_kfree_skb(fp->rx_skbuff[i]);
 | 
						|
				fp->rx_dma[i] = 0;
 | 
						|
				fp->rx_skbuff[i] = NULL;
 | 
						|
			}
 | 
						|
			return -ENOMEM;
 | 
						|
		}
 | 
						|
		fp->rx_skbuff[i] = skb;
 | 
						|
		fp->rx_dma[i] = dma;
 | 
						|
	}
 | 
						|
 | 
						|
	ret = fza_init_send(dev, NULL);
 | 
						|
	if (ret != 0)
 | 
						|
		return ret;
 | 
						|
 | 
						|
	/* Purger and Beacon multicasts need to be supplied before PARAM. */
 | 
						|
	fza_set_rx_mode(dev);
 | 
						|
 | 
						|
	spin_lock_irqsave(&fp->lock, flags);
 | 
						|
	fp->cmd_done_flag = 0;
 | 
						|
	ring = fza_cmd_send(dev, FZA_RING_CMD_PARAM);
 | 
						|
	spin_unlock_irqrestore(&fp->lock, flags);
 | 
						|
	if (!ring)
 | 
						|
		return -ENOBUFS;
 | 
						|
 | 
						|
	t = wait_event_timeout(fp->cmd_done_wait, fp->cmd_done_flag, 3 * HZ);
 | 
						|
	if (fp->cmd_done_flag == 0) {
 | 
						|
		pr_err("%s: PARAM command timed out!, state %x\n", fp->name,
 | 
						|
		       FZA_STATUS_GET_STATE(readw_u(&fp->regs->status)));
 | 
						|
		return -EIO;
 | 
						|
	}
 | 
						|
	stat = readl_u(&ring->stat);
 | 
						|
	if (stat != FZA_RING_STAT_SUCCESS) {
 | 
						|
		pr_err("%s: PARAM command failed!, status %02x, state %x\n",
 | 
						|
		       fp->name, stat,
 | 
						|
		       FZA_STATUS_GET_STATE(readw_u(&fp->regs->status)));
 | 
						|
		return -EIO;
 | 
						|
	}
 | 
						|
	pr_debug("%s: PARAM: %lums elapsed\n", fp->name,
 | 
						|
		 (3 * HZ - t) * 1000 / HZ);
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int fza_close(struct net_device *dev)
 | 
						|
{
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
	unsigned long flags;
 | 
						|
	uint state;
 | 
						|
	long t;
 | 
						|
	int i;
 | 
						|
 | 
						|
	netif_stop_queue(dev);
 | 
						|
	pr_debug("%s: queue stopped\n", fp->name);
 | 
						|
 | 
						|
	del_timer_sync(&fp->reset_timer);
 | 
						|
	spin_lock_irqsave(&fp->lock, flags);
 | 
						|
	fp->state = FZA_STATE_UNINITIALIZED;
 | 
						|
	fp->state_chg_flag = 0;
 | 
						|
	/* Shut the interface down. */
 | 
						|
	writew_o(FZA_CONTROL_A_SHUT, &fp->regs->control_a);
 | 
						|
	readw_o(&fp->regs->control_a);			/* Synchronize. */
 | 
						|
	spin_unlock_irqrestore(&fp->lock, flags);
 | 
						|
 | 
						|
	/* DEC says SHUT needs up to 10 seconds to complete. */
 | 
						|
	t = wait_event_timeout(fp->state_chg_wait, fp->state_chg_flag,
 | 
						|
			       15 * HZ);
 | 
						|
	state = FZA_STATUS_GET_STATE(readw_o(&fp->regs->status));
 | 
						|
	if (fp->state_chg_flag == 0) {
 | 
						|
		pr_err("%s: SHUT timed out!, state %x\n", fp->name, state);
 | 
						|
		return -EIO;
 | 
						|
	}
 | 
						|
	if (state != FZA_STATE_UNINITIALIZED) {
 | 
						|
		pr_err("%s: SHUT failed!, state %x\n", fp->name, state);
 | 
						|
		return -EIO;
 | 
						|
	}
 | 
						|
	pr_debug("%s: SHUT: %lums elapsed\n", fp->name,
 | 
						|
		 (15 * HZ - t) * 1000 / HZ);
 | 
						|
 | 
						|
	for (i = 0; i < FZA_RING_RX_SIZE; i++)
 | 
						|
		if (fp->rx_skbuff[i]) {
 | 
						|
			dma_unmap_single(fp->bdev, fp->rx_dma[i],
 | 
						|
					 FZA_RX_BUFFER_SIZE, DMA_FROM_DEVICE);
 | 
						|
			dev_kfree_skb(fp->rx_skbuff[i]);
 | 
						|
			fp->rx_dma[i] = 0;
 | 
						|
			fp->rx_skbuff[i] = NULL;
 | 
						|
		}
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static struct net_device_stats *fza_get_stats(struct net_device *dev)
 | 
						|
{
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
 | 
						|
	return &fp->stats;
 | 
						|
}
 | 
						|
 | 
						|
static int fza_probe(struct device *bdev)
 | 
						|
{
 | 
						|
	static const struct net_device_ops netdev_ops = {
 | 
						|
		.ndo_open = fza_open,
 | 
						|
		.ndo_stop = fza_close,
 | 
						|
		.ndo_start_xmit = fza_start_xmit,
 | 
						|
		.ndo_set_rx_mode = fza_set_rx_mode,
 | 
						|
		.ndo_set_mac_address = fza_set_mac_address,
 | 
						|
		.ndo_get_stats = fza_get_stats,
 | 
						|
	};
 | 
						|
	static int version_printed;
 | 
						|
	char rom_rev[4], fw_rev[4], rmc_rev[4];
 | 
						|
	struct tc_dev *tdev = to_tc_dev(bdev);
 | 
						|
	struct fza_cmd_init __iomem *init;
 | 
						|
	resource_size_t start, len;
 | 
						|
	struct net_device *dev;
 | 
						|
	struct fza_private *fp;
 | 
						|
	uint smt_ver, pmd_type;
 | 
						|
	void __iomem *mmio;
 | 
						|
	uint hw_addr[2];
 | 
						|
	int ret, i;
 | 
						|
 | 
						|
	if (!version_printed) {
 | 
						|
		pr_info("%s", version);
 | 
						|
		version_printed = 1;
 | 
						|
	}
 | 
						|
 | 
						|
	dev = alloc_fddidev(sizeof(*fp));
 | 
						|
	if (!dev)
 | 
						|
		return -ENOMEM;
 | 
						|
	SET_NETDEV_DEV(dev, bdev);
 | 
						|
 | 
						|
	fp = netdev_priv(dev);
 | 
						|
	dev_set_drvdata(bdev, dev);
 | 
						|
 | 
						|
	fp->bdev = bdev;
 | 
						|
	fp->name = dev_name(bdev);
 | 
						|
 | 
						|
	/* Request the I/O MEM resource. */
 | 
						|
	start = tdev->resource.start;
 | 
						|
	len = tdev->resource.end - start + 1;
 | 
						|
	if (!request_mem_region(start, len, dev_name(bdev))) {
 | 
						|
		pr_err("%s: cannot reserve MMIO region\n", fp->name);
 | 
						|
		ret = -EBUSY;
 | 
						|
		goto err_out_kfree;
 | 
						|
	}
 | 
						|
 | 
						|
	/* MMIO mapping setup. */
 | 
						|
	mmio = ioremap(start, len);
 | 
						|
	if (!mmio) {
 | 
						|
		pr_err("%s: cannot map MMIO\n", fp->name);
 | 
						|
		ret = -ENOMEM;
 | 
						|
		goto err_out_resource;
 | 
						|
	}
 | 
						|
 | 
						|
	/* Initialize the new device structure. */
 | 
						|
	switch (loopback) {
 | 
						|
	case FZA_LOOP_NORMAL:
 | 
						|
	case FZA_LOOP_INTERN:
 | 
						|
	case FZA_LOOP_EXTERN:
 | 
						|
		break;
 | 
						|
	default:
 | 
						|
		loopback = FZA_LOOP_NORMAL;
 | 
						|
	}
 | 
						|
 | 
						|
	fp->mmio = mmio;
 | 
						|
	dev->irq = tdev->interrupt;
 | 
						|
 | 
						|
	pr_info("%s: DEC FDDIcontroller 700 or 700-C at 0x%08llx, irq %d\n",
 | 
						|
		fp->name, (long long)tdev->resource.start, dev->irq);
 | 
						|
	pr_debug("%s: mapped at: 0x%p\n", fp->name, mmio);
 | 
						|
 | 
						|
	fp->regs = mmio + FZA_REG_BASE;
 | 
						|
	fp->ring_cmd = mmio + FZA_RING_CMD;
 | 
						|
	fp->ring_uns = mmio + FZA_RING_UNS;
 | 
						|
 | 
						|
	init_waitqueue_head(&fp->state_chg_wait);
 | 
						|
	init_waitqueue_head(&fp->cmd_done_wait);
 | 
						|
	spin_lock_init(&fp->lock);
 | 
						|
	fp->int_mask = FZA_MASK_NORMAL;
 | 
						|
 | 
						|
	timer_setup(&fp->reset_timer, fza_reset_timer, 0);
 | 
						|
 | 
						|
	/* Sanitize the board. */
 | 
						|
	fza_regs_dump(fp);
 | 
						|
	fza_do_shutdown(fp);
 | 
						|
 | 
						|
	ret = request_irq(dev->irq, fza_interrupt, IRQF_SHARED, fp->name, dev);
 | 
						|
	if (ret != 0) {
 | 
						|
		pr_err("%s: unable to get IRQ %d!\n", fp->name, dev->irq);
 | 
						|
		goto err_out_map;
 | 
						|
	}
 | 
						|
 | 
						|
	/* Enable the driver mode. */
 | 
						|
	writew_o(FZA_CONTROL_B_DRIVER, &fp->regs->control_b);
 | 
						|
 | 
						|
	/* For some reason transmit done interrupts can trigger during
 | 
						|
	 * reset.  This avoids a division error in the handler.
 | 
						|
	 */
 | 
						|
	fp->ring_rmc_tx_size = FZA_RING_TX_SIZE;
 | 
						|
 | 
						|
	ret = fza_reset(fp);
 | 
						|
	if (ret != 0)
 | 
						|
		goto err_out_irq;
 | 
						|
 | 
						|
	ret = fza_init_send(dev, &init);
 | 
						|
	if (ret != 0)
 | 
						|
		goto err_out_irq;
 | 
						|
 | 
						|
	fza_reads(&init->hw_addr, &hw_addr, sizeof(hw_addr));
 | 
						|
	dev_addr_set(dev, (u8 *)&hw_addr);
 | 
						|
 | 
						|
	fza_reads(&init->rom_rev, &rom_rev, sizeof(rom_rev));
 | 
						|
	fza_reads(&init->fw_rev, &fw_rev, sizeof(fw_rev));
 | 
						|
	fza_reads(&init->rmc_rev, &rmc_rev, sizeof(rmc_rev));
 | 
						|
	for (i = 3; i >= 0 && rom_rev[i] == ' '; i--)
 | 
						|
		rom_rev[i] = 0;
 | 
						|
	for (i = 3; i >= 0 && fw_rev[i] == ' '; i--)
 | 
						|
		fw_rev[i] = 0;
 | 
						|
	for (i = 3; i >= 0 && rmc_rev[i] == ' '; i--)
 | 
						|
		rmc_rev[i] = 0;
 | 
						|
 | 
						|
	fp->ring_rmc_tx = mmio + readl_u(&init->rmc_tx);
 | 
						|
	fp->ring_rmc_tx_size = readl_u(&init->rmc_tx_size);
 | 
						|
	fp->ring_hst_rx = mmio + readl_u(&init->hst_rx);
 | 
						|
	fp->ring_hst_rx_size = readl_u(&init->hst_rx_size);
 | 
						|
	fp->ring_smt_tx = mmio + readl_u(&init->smt_tx);
 | 
						|
	fp->ring_smt_tx_size = readl_u(&init->smt_tx_size);
 | 
						|
	fp->ring_smt_rx = mmio + readl_u(&init->smt_rx);
 | 
						|
	fp->ring_smt_rx_size = readl_u(&init->smt_rx_size);
 | 
						|
 | 
						|
	fp->buffer_tx = mmio + FZA_TX_BUFFER_ADDR(readl_u(&init->rmc_tx));
 | 
						|
 | 
						|
	fp->t_max = readl_u(&init->def_t_max);
 | 
						|
	fp->t_req = readl_u(&init->def_t_req);
 | 
						|
	fp->tvx = readl_u(&init->def_tvx);
 | 
						|
	fp->lem_threshold = readl_u(&init->lem_threshold);
 | 
						|
	fza_reads(&init->def_station_id, &fp->station_id,
 | 
						|
		  sizeof(fp->station_id));
 | 
						|
	fp->rtoken_timeout = readl_u(&init->rtoken_timeout);
 | 
						|
	fp->ring_purger = readl_u(&init->ring_purger);
 | 
						|
 | 
						|
	smt_ver = readl_u(&init->smt_ver);
 | 
						|
	pmd_type = readl_u(&init->pmd_type);
 | 
						|
 | 
						|
	pr_debug("%s: INIT parameters:\n", fp->name);
 | 
						|
	pr_debug("        tx_mode: %u\n", readl_u(&init->tx_mode));
 | 
						|
	pr_debug("    hst_rx_size: %u\n", readl_u(&init->hst_rx_size));
 | 
						|
	pr_debug("        rmc_rev: %.4s\n", rmc_rev);
 | 
						|
	pr_debug("        rom_rev: %.4s\n", rom_rev);
 | 
						|
	pr_debug("         fw_rev: %.4s\n", fw_rev);
 | 
						|
	pr_debug("       mop_type: %u\n", readl_u(&init->mop_type));
 | 
						|
	pr_debug("         hst_rx: 0x%08x\n", readl_u(&init->hst_rx));
 | 
						|
	pr_debug("         rmc_tx: 0x%08x\n", readl_u(&init->rmc_tx));
 | 
						|
	pr_debug("    rmc_tx_size: %u\n", readl_u(&init->rmc_tx_size));
 | 
						|
	pr_debug("         smt_tx: 0x%08x\n", readl_u(&init->smt_tx));
 | 
						|
	pr_debug("    smt_tx_size: %u\n", readl_u(&init->smt_tx_size));
 | 
						|
	pr_debug("         smt_rx: 0x%08x\n", readl_u(&init->smt_rx));
 | 
						|
	pr_debug("    smt_rx_size: %u\n", readl_u(&init->smt_rx_size));
 | 
						|
	/* TC systems are always LE, so don't bother swapping. */
 | 
						|
	pr_debug("        hw_addr: 0x%02x%02x%02x%02x%02x%02x%02x%02x\n",
 | 
						|
		 (readl_u(&init->hw_addr[0]) >> 0) & 0xff,
 | 
						|
		 (readl_u(&init->hw_addr[0]) >> 8) & 0xff,
 | 
						|
		 (readl_u(&init->hw_addr[0]) >> 16) & 0xff,
 | 
						|
		 (readl_u(&init->hw_addr[0]) >> 24) & 0xff,
 | 
						|
		 (readl_u(&init->hw_addr[1]) >> 0) & 0xff,
 | 
						|
		 (readl_u(&init->hw_addr[1]) >> 8) & 0xff,
 | 
						|
		 (readl_u(&init->hw_addr[1]) >> 16) & 0xff,
 | 
						|
		 (readl_u(&init->hw_addr[1]) >> 24) & 0xff);
 | 
						|
	pr_debug("      def_t_req: %u\n", readl_u(&init->def_t_req));
 | 
						|
	pr_debug("        def_tvx: %u\n", readl_u(&init->def_tvx));
 | 
						|
	pr_debug("      def_t_max: %u\n", readl_u(&init->def_t_max));
 | 
						|
	pr_debug("  lem_threshold: %u\n", readl_u(&init->lem_threshold));
 | 
						|
	/* Don't bother swapping, see above. */
 | 
						|
	pr_debug(" def_station_id: 0x%02x%02x%02x%02x%02x%02x%02x%02x\n",
 | 
						|
		 (readl_u(&init->def_station_id[0]) >> 0) & 0xff,
 | 
						|
		 (readl_u(&init->def_station_id[0]) >> 8) & 0xff,
 | 
						|
		 (readl_u(&init->def_station_id[0]) >> 16) & 0xff,
 | 
						|
		 (readl_u(&init->def_station_id[0]) >> 24) & 0xff,
 | 
						|
		 (readl_u(&init->def_station_id[1]) >> 0) & 0xff,
 | 
						|
		 (readl_u(&init->def_station_id[1]) >> 8) & 0xff,
 | 
						|
		 (readl_u(&init->def_station_id[1]) >> 16) & 0xff,
 | 
						|
		 (readl_u(&init->def_station_id[1]) >> 24) & 0xff);
 | 
						|
	pr_debug("   pmd_type_alt: %u\n", readl_u(&init->pmd_type_alt));
 | 
						|
	pr_debug("        smt_ver: %u\n", readl_u(&init->smt_ver));
 | 
						|
	pr_debug(" rtoken_timeout: %u\n", readl_u(&init->rtoken_timeout));
 | 
						|
	pr_debug("    ring_purger: %u\n", readl_u(&init->ring_purger));
 | 
						|
	pr_debug("    smt_ver_max: %u\n", readl_u(&init->smt_ver_max));
 | 
						|
	pr_debug("    smt_ver_min: %u\n", readl_u(&init->smt_ver_min));
 | 
						|
	pr_debug("       pmd_type: %u\n", readl_u(&init->pmd_type));
 | 
						|
 | 
						|
	pr_info("%s: model %s, address %pMF\n",
 | 
						|
		fp->name,
 | 
						|
		pmd_type == FZA_PMD_TYPE_TW ?
 | 
						|
			"700-C (DEFZA-CA), ThinWire PMD selected" :
 | 
						|
			pmd_type == FZA_PMD_TYPE_STP ?
 | 
						|
				"700-C (DEFZA-CA), STP PMD selected" :
 | 
						|
				"700 (DEFZA-AA), MMF PMD",
 | 
						|
		dev->dev_addr);
 | 
						|
	pr_info("%s: ROM rev. %.4s, firmware rev. %.4s, RMC rev. %.4s, "
 | 
						|
		"SMT ver. %u\n", fp->name, rom_rev, fw_rev, rmc_rev, smt_ver);
 | 
						|
 | 
						|
	/* Now that we fetched initial parameters just shut the interface
 | 
						|
	 * until opened.
 | 
						|
	 */
 | 
						|
	ret = fza_close(dev);
 | 
						|
	if (ret != 0)
 | 
						|
		goto err_out_irq;
 | 
						|
 | 
						|
	/* The FZA-specific entries in the device structure. */
 | 
						|
	dev->netdev_ops = &netdev_ops;
 | 
						|
 | 
						|
	ret = register_netdev(dev);
 | 
						|
	if (ret != 0)
 | 
						|
		goto err_out_irq;
 | 
						|
 | 
						|
	pr_info("%s: registered as %s\n", fp->name, dev->name);
 | 
						|
	fp->name = (const char *)dev->name;
 | 
						|
 | 
						|
	get_device(bdev);
 | 
						|
	return 0;
 | 
						|
 | 
						|
err_out_irq:
 | 
						|
	del_timer_sync(&fp->reset_timer);
 | 
						|
	fza_do_shutdown(fp);
 | 
						|
	free_irq(dev->irq, dev);
 | 
						|
 | 
						|
err_out_map:
 | 
						|
	iounmap(mmio);
 | 
						|
 | 
						|
err_out_resource:
 | 
						|
	release_mem_region(start, len);
 | 
						|
 | 
						|
err_out_kfree:
 | 
						|
	pr_err("%s: initialization failure, aborting!\n", fp->name);
 | 
						|
	free_netdev(dev);
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
static int fza_remove(struct device *bdev)
 | 
						|
{
 | 
						|
	struct net_device *dev = dev_get_drvdata(bdev);
 | 
						|
	struct fza_private *fp = netdev_priv(dev);
 | 
						|
	struct tc_dev *tdev = to_tc_dev(bdev);
 | 
						|
	resource_size_t start, len;
 | 
						|
 | 
						|
	put_device(bdev);
 | 
						|
 | 
						|
	unregister_netdev(dev);
 | 
						|
 | 
						|
	del_timer_sync(&fp->reset_timer);
 | 
						|
	fza_do_shutdown(fp);
 | 
						|
	free_irq(dev->irq, dev);
 | 
						|
 | 
						|
	iounmap(fp->mmio);
 | 
						|
 | 
						|
	start = tdev->resource.start;
 | 
						|
	len = tdev->resource.end - start + 1;
 | 
						|
	release_mem_region(start, len);
 | 
						|
 | 
						|
	free_netdev(dev);
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static struct tc_device_id const fza_tc_table[] = {
 | 
						|
	{ "DEC     ", "PMAF-AA " },
 | 
						|
	{ }
 | 
						|
};
 | 
						|
MODULE_DEVICE_TABLE(tc, fza_tc_table);
 | 
						|
 | 
						|
static struct tc_driver fza_driver = {
 | 
						|
	.id_table	= fza_tc_table,
 | 
						|
	.driver		= {
 | 
						|
		.name	= "defza",
 | 
						|
		.bus	= &tc_bus_type,
 | 
						|
		.probe	= fza_probe,
 | 
						|
		.remove	= fza_remove,
 | 
						|
	},
 | 
						|
};
 | 
						|
 | 
						|
static int fza_init(void)
 | 
						|
{
 | 
						|
	return tc_register_driver(&fza_driver);
 | 
						|
}
 | 
						|
 | 
						|
static void fza_exit(void)
 | 
						|
{
 | 
						|
	tc_unregister_driver(&fza_driver);
 | 
						|
}
 | 
						|
 | 
						|
module_init(fza_init);
 | 
						|
module_exit(fza_exit);
 |