config root man

Current Path : /sys/amd64/compile/hs32/modules/usr/src/sys/modules/i2c/iicbus/@/dev/qlxgb/

FreeBSD hs32.drive.ne.jp 9.1-RELEASE FreeBSD 9.1-RELEASE #1: Wed Jan 14 12:18:08 JST 2015 root@hs32.drive.ne.jp:/sys/amd64/compile/hs32 amd64
Upload File :
Current File : //sys/amd64/compile/hs32/modules/usr/src/sys/modules/i2c/iicbus/@/dev/qlxgb/qla_isr.c

/*
 * Copyright (c) 2010-2011 Qlogic Corporation
 * All rights reserved.
 *
 *  Redistribution and use in source and binary forms, with or without
 *  modification, are permitted provided that the following conditions
 *  are met:
 *
 *  1. Redistributions of source code must retain the above copyright
 *     notice, this list of conditions and the following disclaimer.
 *  2. Redistributions in binary form must reproduce the above copyright
 *     notice, this list of conditions and the following disclaimer in the
 *     documentation and/or other materials provided with the distribution.
 *
 *  THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
 *  AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 *  IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 *  ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
 *  LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
 *  CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
 *  SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
 *  INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
 *  CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
 *  ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
 *  POSSIBILITY OF SUCH DAMAGE.
 */

/*
 * File: qla_isr.c
 * Author : David C Somayajulu, Qlogic Corporation, Aliso Viejo, CA 92656.
 */

#include <sys/cdefs.h>
__FBSDID("$FreeBSD: release/9.1.0/sys/dev/qlxgb/qla_isr.c 227064 2011-11-03 21:20:22Z bz $");

#include "qla_os.h"
#include "qla_reg.h"
#include "qla_hw.h"
#include "qla_def.h"
#include "qla_inline.h"
#include "qla_ver.h"
#include "qla_glbl.h"
#include "qla_dbg.h"

static void qla_replenish_normal_rx(qla_host_t *ha, qla_sds_t *sdsp);
static void qla_replenish_jumbo_rx(qla_host_t *ha, qla_sds_t *sdsp);

/*
 * Name: qla_rx_intr
 * Function: Handles normal ethernet frames received
 */
static void
qla_rx_intr(qla_host_t *ha, uint64_t data, uint32_t sds_idx,
	struct lro_ctrl *lro)
{
	uint32_t idx, length, status, ring;
	qla_rx_buf_t *rxb;
	struct mbuf *mp;
	struct ifnet *ifp = ha->ifp;
	qla_sds_t *sdsp;
	struct ether_vlan_header *eh;
	
	sdsp = &ha->hw.sds[sds_idx];
	
	ring = (uint32_t)Q8_STAT_DESC_TYPE(data);
	idx = (uint32_t)Q8_STAT_DESC_HANDLE(data);
	length = (uint32_t)Q8_STAT_DESC_TOTAL_LENGTH(data);
	status = (uint32_t)Q8_STAT_DESC_STATUS(data);

	if (ring == 0) {
		if ((idx >= NUM_RX_DESCRIPTORS) || (length > MCLBYTES)) {
			device_printf(ha->pci_dev, "%s: ring[%d] index[0x%08x]"
				" len[0x%08x] invalid\n",
				__func__, ring, idx, length);
			return;
		}
	} else {
		if ((idx >= NUM_RX_JUMBO_DESCRIPTORS)||(length > MJUM9BYTES)) {
			device_printf(ha->pci_dev, "%s: ring[%d] index[0x%08x]"
				" len[0x%08x] invalid\n",
				__func__, ring, idx, length);
			return;
		}
	}

	if (ring == 0)
		rxb = &ha->rx_buf[idx];
	else 
		rxb = &ha->rx_jbuf[idx];

	QL_ASSERT((rxb != NULL),\
		("%s: [r, i, sds_idx]=[%d, 0x%x, %d] rxb != NULL\n",\
		 __func__, ring, idx, sds_idx));

	mp = rxb->m_head;

	QL_ASSERT((mp != NULL),\
		("%s: [r,i,rxb, sds_idx]=[%d, 0x%x, %p, %d] mp != NULL\n",\
		 __func__, ring, idx, rxb, sds_idx));

	bus_dmamap_sync(ha->rx_tag, rxb->map, BUS_DMASYNC_POSTREAD);

	if (ring == 0) {
		rxb->m_head = NULL;
		rxb->next = sdsp->rxb_free;
		sdsp->rxb_free = rxb;
		sdsp->rx_free++;
	} else {
		rxb->m_head = NULL;
		rxb->next = sdsp->rxjb_free;
		sdsp->rxjb_free = rxb;
		sdsp->rxj_free++;
	}
	
	mp->m_len = length;
	mp->m_pkthdr.len = length;
	mp->m_pkthdr.rcvif = ifp;
	
	eh = mtod(mp, struct ether_vlan_header *);

	if (eh->evl_encap_proto == htons(ETHERTYPE_VLAN)) {
		uint32_t *data = (uint32_t *)eh;

		mp->m_pkthdr.ether_vtag = ntohs(eh->evl_tag);
		mp->m_flags |= M_VLANTAG;

		*(data + 3) = *(data + 2);
		*(data + 2) = *(data + 1);
		*(data + 1) = *data;

		m_adj(mp, ETHER_VLAN_ENCAP_LEN);
	}

	if (status == Q8_STAT_DESC_STATUS_CHKSUM_OK) {
		mp->m_pkthdr.csum_flags = (CSUM_IP_CHECKED | CSUM_IP_VALID);
	} else {
		mp->m_pkthdr.csum_flags = 0;
	}

	if (lro->lro_cnt && (tcp_lro_rx(lro, mp, 0) == 0)) {
		/* LRO packet has been successfuly queued */
	} else {
		(*ifp->if_input)(ifp, mp);
	}

	if (sdsp->rx_free > std_replenish)
		qla_replenish_normal_rx(ha, sdsp);

	if (sdsp->rxj_free > jumbo_replenish)
		qla_replenish_jumbo_rx(ha, sdsp);

	return;
}

static void
qla_replenish_jumbo_rx(qla_host_t *ha, qla_sds_t *sdsp)
{
	qla_rx_buf_t *rxb;
	int count = jumbo_replenish;
	uint32_t rxj_next;

	if (!mtx_trylock(&ha->rxj_lock))
		return;

	rxj_next = ha->hw.rxj_next;

	while (count--) {
		rxb = sdsp->rxjb_free;

		if (rxb == NULL)
			break;

		sdsp->rxjb_free = rxb->next;
		sdsp->rxj_free--;


		if (qla_get_mbuf(ha, rxb, NULL, RDS_RING_INDEX_JUMBO) == 0) {
			qla_set_hw_rcv_desc(ha, RDS_RING_INDEX_JUMBO,
				ha->hw.rxj_in, rxb->handle, rxb->paddr,
				(rxb->m_head)->m_pkthdr.len);
			ha->hw.rxj_in++;
			if (ha->hw.rxj_in == NUM_RX_JUMBO_DESCRIPTORS)
				ha->hw.rxj_in = 0;
			ha->hw.rxj_next++;
			if (ha->hw.rxj_next == NUM_RX_JUMBO_DESCRIPTORS)
				ha->hw.rxj_next = 0;
		} else {
			device_printf(ha->pci_dev,
				"%s: qla_get_mbuf [1,(%d),(%d)] failed\n",
				__func__, ha->hw.rxj_in, rxb->handle);

			rxb->m_head = NULL;
			rxb->next = sdsp->rxjb_free;
			sdsp->rxjb_free = rxb;
			sdsp->rxj_free++;

			break;
		}
	}

	if (rxj_next != ha->hw.rxj_next) {
		QL_UPDATE_RDS_PRODUCER_INDEX(ha, 1, ha->hw.rxj_next);
	}
	mtx_unlock(&ha->rxj_lock);
}

static void
qla_replenish_normal_rx(qla_host_t *ha, qla_sds_t *sdsp)
{
	qla_rx_buf_t *rxb;
	int count = std_replenish;
	uint32_t rx_next;

	if (!mtx_trylock(&ha->rx_lock))
		return;

	rx_next = ha->hw.rx_next;

	while (count--) {
		rxb = sdsp->rxb_free;

		if (rxb == NULL)
			break;

		sdsp->rxb_free = rxb->next;
		sdsp->rx_free--;

		if (qla_get_mbuf(ha, rxb, NULL, RDS_RING_INDEX_NORMAL) == 0) {
			qla_set_hw_rcv_desc(ha, RDS_RING_INDEX_NORMAL,
				ha->hw.rx_in, rxb->handle, rxb->paddr,
				(rxb->m_head)->m_pkthdr.len);
			ha->hw.rx_in++;
			if (ha->hw.rx_in == NUM_RX_DESCRIPTORS)
				ha->hw.rx_in = 0;
			ha->hw.rx_next++;
			if (ha->hw.rx_next == NUM_RX_DESCRIPTORS)
				ha->hw.rx_next = 0;
		} else {
			device_printf(ha->pci_dev,
				"%s: qla_get_mbuf [0,(%d),(%d)] failed\n",
				__func__, ha->hw.rx_in, rxb->handle);

			rxb->m_head = NULL;
			rxb->next = sdsp->rxb_free;
			sdsp->rxb_free = rxb;
			sdsp->rx_free++;

			break;
		}
	}

	if (rx_next != ha->hw.rx_next) {
		QL_UPDATE_RDS_PRODUCER_INDEX(ha, 0, ha->hw.rx_next);
	}
	mtx_unlock(&ha->rx_lock);
}

/*
 * Name: qla_isr
 * Function: Main Interrupt Service Routine
 */
static uint32_t
qla_rcv_isr(qla_host_t *ha, uint32_t sds_idx, uint32_t count)
{
	device_t dev;
	qla_hw_t *hw;
	uint32_t comp_idx, desc_count;
	q80_stat_desc_t *sdesc;
	struct lro_ctrl *lro;
	struct lro_entry *queued;
	uint32_t ret = 0;

	dev = ha->pci_dev;
	hw = &ha->hw;

	hw->sds[sds_idx].rcv_active = 1;
	if (ha->flags.stop_rcv) {
		hw->sds[sds_idx].rcv_active = 0;
		return 0;
	}

	QL_DPRINT2((dev, "%s: [%d]enter\n", __func__, sds_idx));

	/*
	 * receive interrupts
	 */
	comp_idx = hw->sds[sds_idx].sdsr_next;
	lro = &hw->sds[sds_idx].lro;

	while (count--) {

		sdesc = (q80_stat_desc_t *)
				&hw->sds[sds_idx].sds_ring_base[comp_idx];

		if (Q8_STAT_DESC_OWNER((sdesc->data[0])) !=
			Q8_STAT_DESC_OWNER_HOST) {
			QL_DPRINT2((dev, "%s:  data %p sdsr_next 0x%08x\n",
				__func__, (void *)sdesc->data[0], comp_idx));
			break;
		}

		desc_count = Q8_STAT_DESC_COUNT((sdesc->data[0]));

		switch (Q8_STAT_DESC_OPCODE((sdesc->data[0]))) {

		case Q8_STAT_DESC_OPCODE_RCV_PKT:
		case Q8_STAT_DESC_OPCODE_SYN_OFFLOAD:
			qla_rx_intr(ha, (sdesc->data[0]), sds_idx, lro);
			
			break;

		default:
			device_printf(dev, "%s: default 0x%llx!\n", __func__,
					(long long unsigned int)sdesc->data[0]);
			break;
		}

		while (desc_count--) {
			sdesc->data[0] =
				Q8_STAT_DESC_SET_OWNER(Q8_STAT_DESC_OWNER_FW);
			comp_idx = (comp_idx + 1) & (NUM_STATUS_DESCRIPTORS-1);
			sdesc = (q80_stat_desc_t *)
				&hw->sds[sds_idx].sds_ring_base[comp_idx];
		}
	}

	while((!SLIST_EMPTY(&lro->lro_active))) {
		queued = SLIST_FIRST(&lro->lro_active);
		SLIST_REMOVE_HEAD(&lro->lro_active, next);
		tcp_lro_flush(lro, queued);
	}

	if (hw->sds[sds_idx].sdsr_next != comp_idx) {
		QL_UPDATE_SDS_CONSUMER_INDEX(ha, sds_idx, comp_idx);
	}
	hw->sds[sds_idx].sdsr_next = comp_idx;

	sdesc = (q80_stat_desc_t *)&hw->sds[sds_idx].sds_ring_base[comp_idx];
	if ((sds_idx == 0) && (Q8_STAT_DESC_OWNER((sdesc->data[0])) ==
					Q8_STAT_DESC_OWNER_HOST)) {
		ret = -1;
	}

	hw->sds[sds_idx].rcv_active = 0;
	return (ret);
}

void
qla_isr(void *arg)
{
	qla_ivec_t *ivec = arg;
	qla_host_t *ha;
	uint32_t sds_idx;
	uint32_t ret;

	ha = ivec->ha;
	sds_idx = ivec->irq_rid - 1;

	if (sds_idx >= ha->hw.num_sds_rings) {
		device_printf(ha->pci_dev, "%s: bogus sds_idx 0x%x\n", __func__,
			sds_idx);
		
		return;
	}

	if (sds_idx == 0)
		taskqueue_enqueue(ha->tx_tq, &ha->tx_task);

	ret = qla_rcv_isr(ha, sds_idx, rcv_pkt_thres);

	if (sds_idx == 0)
		taskqueue_enqueue(ha->tx_tq, &ha->tx_task);

	if (ret) {
		taskqueue_enqueue(ha->irq_vec[sds_idx].rcv_tq,
			&ha->irq_vec[sds_idx].rcv_task);
	} else {
		QL_ENABLE_INTERRUPTS(ha, sds_idx);
	}
}

void
qla_rcv(void *context, int pending)
{
	qla_ivec_t *ivec = context;
	qla_host_t *ha;
	device_t dev;
	qla_hw_t *hw;
	uint32_t sds_idx;
	uint32_t ret;
	struct ifnet *ifp;

	ha = ivec->ha;
	dev = ha->pci_dev;
	hw = &ha->hw;
	sds_idx = ivec->irq_rid - 1;
	ifp = ha->ifp;

	do {
		if (sds_idx == 0) {
			if (qla_le32_to_host(*(hw->tx_cons)) != hw->txr_comp) {
				taskqueue_enqueue(ha->tx_tq, &ha->tx_task);
			} else if ((ifp->if_snd.ifq_head != NULL) &&
					QL_RUNNING(ifp)) {
				taskqueue_enqueue(ha->tx_tq, &ha->tx_task);
			}
		}
		ret = qla_rcv_isr(ha, sds_idx, rcv_pkt_thres_d);
	} while (ret);

	if (sds_idx == 0)
		taskqueue_enqueue(ha->tx_tq, &ha->tx_task);

	QL_ENABLE_INTERRUPTS(ha, sds_idx);
}


Man Man