diff options
author | André Fabian Silva Delgado <emulatorman@parabola.nu> | 2015-12-15 14:52:16 -0300 |
---|---|---|
committer | André Fabian Silva Delgado <emulatorman@parabola.nu> | 2015-12-15 14:52:16 -0300 |
commit | 8d91c1e411f55d7ea91b1183a2e9f8088fb4d5be (patch) | |
tree | e9891aa6c295060d065adffd610c4f49ecf884f3 /drivers/staging/unisys/visornic/visornic_main.c | |
parent | a71852147516bc1cb5b0b3cbd13639bfd4022dc8 (diff) |
Linux-libre 4.3.2-gnu
Diffstat (limited to 'drivers/staging/unisys/visornic/visornic_main.c')
-rw-r--r-- | drivers/staging/unisys/visornic/visornic_main.c | 1138 |
1 files changed, 586 insertions, 552 deletions
diff --git a/drivers/staging/unisys/visornic/visornic_main.c b/drivers/staging/unisys/visornic/visornic_main.c index 710074437..9d3c1e282 100644 --- a/drivers/staging/unisys/visornic/visornic_main.c +++ b/drivers/staging/unisys/visornic/visornic_main.c @@ -1,10 +1,9 @@ /* Copyright (c) 2012 - 2015 UNISYS CORPORATION * All rights reserved. * - * This program is free software; you can redistribute it and/or modify - * it under the terms of the GNU General Public License as published by - * the Free Software Foundation; either version 2 of the License, or (at - * your option) any later version. + * This program is free software; you can redistribute it and/or modify it + * under the terms and conditions of the GNU General Public License, + * version 2, as published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of @@ -20,15 +19,16 @@ */ #include <linux/debugfs.h> -#include <linux/netdevice.h> #include <linux/etherdevice.h> -#include <linux/skbuff.h> +#include <linux/netdevice.h> #include <linux/kthread.h> +#include <linux/skbuff.h> +#include <linux/rtnetlink.h> #include "visorbus.h" #include "iochannel.h" -#define VISORNIC_INFINITE_RESPONSE_WAIT 0 +#define VISORNIC_INFINITE_RSP_WAIT 0 #define VISORNICSOPENMAX 32 #define MAXDEVICES 16384 @@ -61,7 +61,6 @@ static const struct file_operations debugfs_enable_ints_fops = { .write = enable_ints_write, }; -static struct workqueue_struct *visornic_serverdown_workqueue; static struct workqueue_struct *visornic_timeout_reset_workqueue; /* GUIDS for director channel type supported by this driver. */ @@ -72,6 +71,15 @@ static struct visor_channeltype_descriptor visornic_channel_types[] = { { SPAR_VNIC_CHANNEL_PROTOCOL_UUID, "ultravnic" }, { NULL_UUID_LE, NULL } }; +MODULE_DEVICE_TABLE(visorbus, visornic_channel_types); +/* + * FIXME XXX: This next line of code must be fixed and removed before + * acceptance into the 'normal' part of the kernel. It is only here as a place + * holder to get module autoloading functionality working for visorbus. Code + * must be added to scripts/mode/file2alias.c, etc., to get this working + * properly. + */ +MODULE_ALIAS("visorbus:" SPAR_VNIC_CHANNEL_PROTOCOL_UUID_STR); /* This is used to tell the visor bus driver which types of visor devices * we support, and what functions to call when a visor device that we support @@ -90,12 +98,6 @@ static struct visor_driver visornic_driver = { .channel_interrupt = NULL, }; -struct visor_thread_info { - struct task_struct *task; - struct completion has_stopped; - int id; -}; - struct chanstat { unsigned long got_rcv; unsigned long got_enbdisack; @@ -104,6 +106,7 @@ struct chanstat { unsigned long sent_enbdis; unsigned long sent_promisc; unsigned long sent_post; + unsigned long sent_post_failed; unsigned long sent_xmit; unsigned long reject_count; unsigned long extra_rcvbufs_sent; @@ -111,7 +114,6 @@ struct chanstat { struct visornic_devdata { int devnum; - int thread_wait_ms; unsigned short enabled; /* 0 disabled 1 enabled to receive */ unsigned short enab_dis_acked; /* NET_RCV_ENABLE/DISABLE acked by * IOPART @@ -119,7 +121,6 @@ struct visornic_devdata { struct visor_device *dev; char name[99]; struct list_head list_all; /* < link within list_all_devices list */ - struct kref kref; struct net_device *netdev; struct net_device_stats net_stats; atomic_t interrupt_rcvd; @@ -137,20 +138,21 @@ struct visornic_devdata { atomic_t num_rcvbuf_in_iovm; unsigned long alloc_failed_in_if_needed_cnt; unsigned long alloc_failed_in_repost_rtn_cnt; - int max_outstanding_net_xmits; /* absolute max number of outstanding - * xmits - should never hit this - */ - int upper_threshold_net_xmits; /* high water mark for calling - * netif_stop_queue() - */ - int lower_threshold_net_xmits; /* high water mark for calling - * netif_wake_queue() - */ + unsigned long max_outstanding_net_xmits; /* absolute max number of + * outstanding xmits - should + * never hit this + */ + unsigned long upper_threshold_net_xmits; /* high water mark for + * calling netif_stop_queue() + */ + unsigned long lower_threshold_net_xmits; /* high water mark for calling + * netif_wake_queue() + */ struct sk_buff_head xmitbufhead; /* xmitbufhead is the head of the * xmit buffer list that have been * sent to the IOPART end */ - struct work_struct serverdown_completion; + visorbus_state_complete_func server_down_complete_func; struct work_struct timeout_reset; struct uiscmdrsp *cmdrsp_rcv; /* cmdrsp_rcv is used for * posting/unposting rcv buffers @@ -161,8 +163,8 @@ struct visornic_devdata { */ bool server_down; /* IOPART is down */ bool server_change_state; /* Processing SERVER_CHANGESTATE msg */ + bool going_away; /* device is being torn down */ struct dentry *eth_debugfs_dir; - struct visor_thread_info threadinfo; u64 interrupts_rcvd; u64 interrupts_notme; u64 interrupts_disabled; @@ -194,16 +196,19 @@ struct visornic_devdata { int queuefullmsg_logged; struct chanstat chstat; + struct timer_list irq_poll_timer; + struct napi_struct napi; + struct uiscmdrsp cmdrsp[SIZEOF_CMDRSP]; }; -/* array of open devices maintained by open() and close() */ -static struct net_device *num_visornic_open[VISORNICSOPENMAX]; /* List of all visornic_devdata structs, * linked via the list_all member */ static LIST_HEAD(list_all_devices); static DEFINE_SPINLOCK(lock_all_devices); +static int visornic_poll(struct napi_struct *napi, int budget); +static void poll_for_irq(unsigned long v); /** * visor_copy_fragsinfo_from_skb( @@ -223,9 +228,25 @@ visor_copy_fragsinfo_from_skb(struct sk_buff *skb, unsigned int firstfraglen, struct phys_info frags[]) { unsigned int count = 0, ii, size, offset = 0, numfrags; + unsigned int total_count; numfrags = skb_shinfo(skb)->nr_frags; + /* + * Compute the number of fragments this skb has, and if its more than + * frag array can hold, linearize the skb + */ + total_count = numfrags + (firstfraglen / PI_PAGE_SIZE); + if (firstfraglen % PI_PAGE_SIZE) + total_count++; + + if (total_count > frags_max) { + if (skb_linearize(skb)) + return -EINVAL; + numfrags = skb_shinfo(skb)->nr_frags; + firstfraglen = 0; + } + while (firstfraglen) { if (count == frags_max) return -EINVAL; @@ -256,8 +277,16 @@ visor_copy_fragsinfo_from_skb(struct sk_buff *skb, unsigned int firstfraglen, page_offset, skb_shinfo(skb)->frags[ii]. size, count, frags_max, frags); - if (!count) - return -EIO; + /* + * add_physinfo_entries only returns + * zero if the frags array is out of room + * That should never happen because we + * fail above, if count+numfrags > frags_max. + * Given that theres no recovery mechanism from putting + * half a packet in the I/O channel, panic here as this + * should never happen + */ + BUG_ON(!count); } } if (skb_shinfo(skb)->frag_list) { @@ -279,222 +308,15 @@ visor_copy_fragsinfo_from_skb(struct sk_buff *skb, unsigned int firstfraglen, return count; } -/** - * visort_thread_start - starts thread for the device - * @thrinfo: The thread to start - * @threadfn: Function the thread starts - * @thrcontext: Context to pass to the thread, i.e. devdata - * @name: string describing name of thread - * - * Starts a thread for the device, currently only thread is - * process_incoming_rsps - * Returns 0 on success; - */ -static int visor_thread_start(struct visor_thread_info *thrinfo, - int (*threadfn)(void *), - void *thrcontext, char *name) -{ - /* used to stop the thread */ - init_completion(&thrinfo->has_stopped); - thrinfo->task = kthread_run(threadfn, thrcontext, name); - if (IS_ERR(thrinfo->task)) { - thrinfo->id = 0; - return -EINVAL; - } - thrinfo->id = thrinfo->task->pid; - return 0; -} - -/** - * visor_thread_stop - stop a thread for the device - * @thrinfo: The thread to stop - * - * Stop the thread and wait for completion for a minute - * Returns void. - */ -static void visor_thread_stop(struct visor_thread_info *thrinfo) -{ - if (!thrinfo->id) - return; /* thread not running */ - - kthread_stop(thrinfo->task); - /* give up if the thread has NOT died in 1 minute */ - if (wait_for_completion_timeout(&thrinfo->has_stopped, 60 * HZ)) - thrinfo->id = 0; -} - -/* DebugFS code */ -static ssize_t info_debugfs_read(struct file *file, char __user *buf, - size_t len, loff_t *offset) -{ - int i; - ssize_t bytes_read = 0; - int str_pos = 0; - struct visornic_devdata *devdata; - char *vbuf; - - if (len > MAX_BUF) - len = MAX_BUF; - vbuf = kzalloc(len, GFP_KERNEL); - if (!vbuf) - return -ENOMEM; - - /* for each vnic channel - * dump out channel specific data - */ - for (i = 0; i < VISORNICSOPENMAX; i++) { - if (!num_visornic_open[i]) - continue; - - devdata = netdev_priv(num_visornic_open[i]); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - "Vnic i = %d\n", i); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - "netdev = %s (0x%p), MAC Addr %pM\n", - num_visornic_open[i]->name, - num_visornic_open[i], - num_visornic_open[i]->dev_addr); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - "VisorNic Dev Info = 0x%p\n", devdata); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " num_rcv_bufs = %d\n", - devdata->num_rcv_bufs); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " max_oustanding_next_xmits = %d\n", - devdata->max_outstanding_net_xmits); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " upper_threshold_net_xmits = %d\n", - devdata->upper_threshold_net_xmits); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " lower_threshold_net_xmits = %d\n", - devdata->lower_threshold_net_xmits); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " queuefullmsg_logged = %d\n", - devdata->queuefullmsg_logged); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " chstat.got_rcv = %lu\n", - devdata->chstat.got_rcv); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " chstat.got_enbdisack = %lu\n", - devdata->chstat.got_enbdisack); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " chstat.got_xmit_done = %lu\n", - devdata->chstat.got_xmit_done); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " chstat.xmit_fail = %lu\n", - devdata->chstat.xmit_fail); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " chstat.sent_enbdis = %lu\n", - devdata->chstat.sent_enbdis); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " chstat.sent_promisc = %lu\n", - devdata->chstat.sent_promisc); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " chstat.sent_post = %lu\n", - devdata->chstat.sent_post); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " chstat.sent_xmit = %lu\n", - devdata->chstat.sent_xmit); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " chstat.reject_count = %lu\n", - devdata->chstat.reject_count); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " chstat.extra_rcvbufs_sent = %lu\n", - devdata->chstat.extra_rcvbufs_sent); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " n_rcv0 = %lu\n", devdata->n_rcv0); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " n_rcv1 = %lu\n", devdata->n_rcv1); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " n_rcv2 = %lu\n", devdata->n_rcv2); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " n_rcvx = %lu\n", devdata->n_rcvx); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " num_rcvbuf_in_iovm = %d\n", - atomic_read(&devdata->num_rcvbuf_in_iovm)); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " alloc_failed_in_if_needed_cnt = %lu\n", - devdata->alloc_failed_in_if_needed_cnt); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " alloc_failed_in_repost_rtn_cnt = %lu\n", - devdata->alloc_failed_in_repost_rtn_cnt); - /* str_pos += scnprintf(vbuf + str_pos, len - str_pos, - * " inner_loop_limit_reached_cnt = %lu\n", - * devdata->inner_loop_limit_reached_cnt); - */ - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " found_repost_rcvbuf_cnt = %lu\n", - devdata->found_repost_rcvbuf_cnt); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " repost_found_skb_cnt = %lu\n", - devdata->repost_found_skb_cnt); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " n_repost_deficit = %lu\n", - devdata->n_repost_deficit); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " bad_rcv_buf = %lu\n", - devdata->bad_rcv_buf); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " n_rcv_packets_not_accepted = %lu\n", - devdata->n_rcv_packets_not_accepted); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " interrupts_rcvd = %llu\n", - devdata->interrupts_rcvd); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " interrupts_notme = %llu\n", - devdata->interrupts_notme); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " interrupts_disabled = %llu\n", - devdata->interrupts_disabled); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " busy_cnt = %llu\n", - devdata->busy_cnt); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " flow_control_upper_hits = %llu\n", - devdata->flow_control_upper_hits); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " flow_control_lower_hits = %llu\n", - devdata->flow_control_lower_hits); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " thread_wait_ms = %d\n", - devdata->thread_wait_ms); - str_pos += scnprintf(vbuf + str_pos, len - str_pos, - " netif_queue = %s\n", - netif_queue_stopped(devdata->netdev) ? - "stopped" : "running"); - } - bytes_read = simple_read_from_buffer(buf, len, offset, vbuf, str_pos); - kfree(vbuf); - return bytes_read; -} - static ssize_t enable_ints_write(struct file *file, const char __user *buffer, size_t count, loff_t *ppos) { - char buf[4]; - int i, new_value; - struct visornic_devdata *devdata; - - if (count >= ARRAY_SIZE(buf)) - return -EINVAL; - - buf[count] = '\0'; - if (copy_from_user(buf, buffer, count)) - return -EFAULT; - - i = kstrtoint(buf, 10, &new_value); - if (i != 0) - return -EFAULT; - - /* set all counts to new_value usually 0 */ - for (i = 0; i < VISORNICSOPENMAX; i++) { - if (num_visornic_open[i]) { - devdata = netdev_priv(num_visornic_open[i]); - /* TODO update features bit in channel */ - } - } - + /* + * Don't want to break ABI here by having a debugfs + * file that no longer exists or is writable, so + * lets just make this a vestigual function + */ return count; } @@ -509,44 +331,29 @@ static ssize_t enable_ints_write(struct file *file, * Returns void. */ static void -visornic_serverdown_complete(struct work_struct *work) +visornic_serverdown_complete(struct visornic_devdata *devdata) { - struct visornic_devdata *devdata; struct net_device *netdev; - unsigned long flags; - int i = 0, count = 0; - devdata = container_of(work, struct visornic_devdata, - serverdown_completion); netdev = devdata->netdev; - /* Stop using datachan */ - visor_thread_stop(&devdata->threadinfo); + /* Stop polling for interrupts */ + del_timer_sync(&devdata->irq_poll_timer); - /* Inform Linux that the link is down */ - netif_carrier_off(netdev); - netif_stop_queue(netdev); + rtnl_lock(); + dev_close(netdev); + rtnl_unlock(); - /* Free the skb for XMITs that haven't been serviced by the server - * We shouldn't have to inform Linux about these IOs because they - * are "lost in the ethernet" - */ - skb_queue_purge(&devdata->xmitbufhead); - - spin_lock_irqsave(&devdata->priv_lock, flags); - /* free rcv buffers */ - for (i = 0; i < devdata->num_rcv_bufs; i++) { - if (devdata->rcvbuf[i]) { - kfree_skb(devdata->rcvbuf[i]); - devdata->rcvbuf[i] = NULL; - count++; - } - } atomic_set(&devdata->num_rcvbuf_in_iovm, 0); - spin_unlock_irqrestore(&devdata->priv_lock, flags); + devdata->chstat.sent_xmit = 0; + devdata->chstat.got_xmit_done = 0; + + if (devdata->server_down_complete_func) + (*devdata->server_down_complete_func)(devdata->dev, 0); devdata->server_down = true; devdata->server_change_state = false; + devdata->server_down_complete_func = NULL; } /** @@ -558,15 +365,31 @@ visornic_serverdown_complete(struct work_struct *work) * Returns 0 if we scheduled the work, -EINVAL on error. */ static int -visornic_serverdown(struct visornic_devdata *devdata) +visornic_serverdown(struct visornic_devdata *devdata, + visorbus_state_complete_func complete_func) { + unsigned long flags; + + spin_lock_irqsave(&devdata->priv_lock, flags); if (!devdata->server_down && !devdata->server_change_state) { + if (devdata->going_away) { + spin_unlock_irqrestore(&devdata->priv_lock, flags); + dev_dbg(&devdata->dev->device, + "%s aborting because device removal pending\n", + __func__); + return -ENODEV; + } devdata->server_change_state = true; - queue_work(visornic_serverdown_workqueue, - &devdata->serverdown_completion); + devdata->server_down_complete_func = complete_func; + spin_unlock_irqrestore(&devdata->priv_lock, flags); + visornic_serverdown_complete(devdata); } else if (devdata->server_change_state) { + dev_dbg(&devdata->dev->device, "%s changing state\n", + __func__); + spin_unlock_irqrestore(&devdata->priv_lock, flags); return -EINVAL; - } + } else + spin_unlock_irqrestore(&devdata->priv_lock, flags); return 0; } @@ -625,11 +448,14 @@ post_skb(struct uiscmdrsp *cmdrsp, if ((cmdrsp->net.rcvpost.frag.pi_off + skb->len) <= PI_PAGE_SIZE) { cmdrsp->net.type = NET_RCV_POST; cmdrsp->cmdtype = CMD_NET_TYPE; - visorchannel_signalinsert(devdata->dev->visorchannel, + if (visorchannel_signalinsert(devdata->dev->visorchannel, IOCHAN_TO_IOPART, - cmdrsp); - atomic_inc(&devdata->num_rcvbuf_in_iovm); - devdata->chstat.sent_post++; + cmdrsp)) { + atomic_inc(&devdata->num_rcvbuf_in_iovm); + devdata->chstat.sent_post++; + } else { + devdata->chstat.sent_post_failed++; + } } } @@ -651,10 +477,10 @@ send_enbdis(struct net_device *netdev, int state, devdata->cmdrsp_rcv->net.enbdis.context = netdev; devdata->cmdrsp_rcv->net.type = NET_RCV_ENBDIS; devdata->cmdrsp_rcv->cmdtype = CMD_NET_TYPE; - visorchannel_signalinsert(devdata->dev->visorchannel, + if (visorchannel_signalinsert(devdata->dev->visorchannel, IOCHAN_TO_IOPART, - devdata->cmdrsp_rcv); - devdata->chstat.sent_enbdis++; + devdata->cmdrsp_rcv)) + devdata->chstat.sent_enbdis++; } /** @@ -676,9 +502,6 @@ visornic_disable_with_timeout(struct net_device *netdev, const int timeout) unsigned long flags; int wait = 0; - /* stop the transmit queue so nothing more can be transmitted */ - netif_stop_queue(netdev); - /* send a msg telling the other end we are stopping incoming pkts */ spin_lock_irqsave(&devdata->priv_lock, flags); devdata->enabled = 0; @@ -695,12 +518,14 @@ visornic_disable_with_timeout(struct net_device *netdev, const int timeout) * when it gets a disable. */ spin_lock_irqsave(&devdata->priv_lock, flags); - while ((timeout == VISORNIC_INFINITE_RESPONSE_WAIT) || + while ((timeout == VISORNIC_INFINITE_RSP_WAIT) || (wait < timeout)) { if (devdata->enab_dis_acked) break; if (devdata->server_down || devdata->server_change_state) { spin_unlock_irqrestore(&devdata->priv_lock, flags); + dev_dbg(&netdev->dev, "%s server went away\n", + __func__); return -EIO; } set_current_state(TASK_INTERRUPTIBLE); @@ -722,10 +547,16 @@ visornic_disable_with_timeout(struct net_device *netdev, const int timeout) break; } } - /* we've set enabled to 0, so we can give up the lock. */ spin_unlock_irqrestore(&devdata->priv_lock, flags); + /* stop the transmit queue so nothing more can be transmitted */ + netif_stop_queue(netdev); + + napi_disable(&devdata->napi); + + skb_queue_purge(&devdata->xmitbufhead); + /* Free rcv buffers - other end has automatically unposed them on * disable */ @@ -736,13 +567,6 @@ visornic_disable_with_timeout(struct net_device *netdev, const int timeout) } } - /* remove references from array */ - for (i = 0; i < VISORNICSOPENMAX; i++) - if (num_visornic_open[i] == netdev) { - num_visornic_open[i] = NULL; - break; - } - return 0; } @@ -814,11 +638,15 @@ visornic_enable_with_timeout(struct net_device *netdev, const int timeout) * gets a disable. */ i = init_rcv_bufs(netdev, devdata); - if (i < 0) + if (i < 0) { + dev_err(&netdev->dev, + "%s failed to init rcv bufs (%d)\n", __func__, i); return i; + } spin_lock_irqsave(&devdata->priv_lock, flags); devdata->enabled = 1; + devdata->enab_dis_acked = 0; /* now we're ready, let's send an ENB to uisnic but until we get * an ACK back from uisnic, we'll drop the packets @@ -829,15 +657,18 @@ visornic_enable_with_timeout(struct net_device *netdev, const int timeout) /* send enable and wait for ack -- don't hold lock when sending enable * because if the queue is full, insert might sleep. */ + napi_enable(&devdata->napi); send_enbdis(netdev, 1, devdata); spin_lock_irqsave(&devdata->priv_lock, flags); - while ((timeout == VISORNIC_INFINITE_RESPONSE_WAIT) || + while ((timeout == VISORNIC_INFINITE_RSP_WAIT) || (wait < timeout)) { if (devdata->enab_dis_acked) break; if (devdata->server_down || devdata->server_change_state) { spin_unlock_irqrestore(&devdata->priv_lock, flags); + dev_dbg(&netdev->dev, "%s server went away\n", + __func__); return -EIO; } set_current_state(TASK_INTERRUPTIBLE); @@ -848,19 +679,13 @@ visornic_enable_with_timeout(struct net_device *netdev, const int timeout) spin_unlock_irqrestore(&devdata->priv_lock, flags); - if (!devdata->enab_dis_acked) + if (!devdata->enab_dis_acked) { + dev_err(&netdev->dev, "%s missing ACK\n", __func__); return -EIO; - - /* find an open slot in the array to save off VisorNic references - * for debug - */ - for (i = 0; i < VISORNICSOPENMAX; i++) { - if (!num_visornic_open[i]) { - num_visornic_open[i] = netdev; - break; - } } + netif_start_queue(netdev); + return 0; } @@ -882,20 +707,29 @@ visornic_timeout_reset(struct work_struct *work) devdata = container_of(work, struct visornic_devdata, timeout_reset); netdev = devdata->netdev; - netif_stop_queue(netdev); - response = visornic_disable_with_timeout(netdev, 100); + rtnl_lock(); + if (!netif_running(netdev)) { + rtnl_unlock(); + return; + } + + response = visornic_disable_with_timeout(netdev, + VISORNIC_INFINITE_RSP_WAIT); if (response) goto call_serverdown; - response = visornic_enable_with_timeout(netdev, 100); + response = visornic_enable_with_timeout(netdev, + VISORNIC_INFINITE_RSP_WAIT); if (response) goto call_serverdown; - netif_wake_queue(netdev); + + rtnl_unlock(); return; call_serverdown: - visornic_serverdown(devdata); + visornic_serverdown(devdata, NULL); + rtnl_unlock(); } /** @@ -908,12 +742,7 @@ call_serverdown: static int visornic_open(struct net_device *netdev) { - visornic_enable_with_timeout(netdev, VISORNIC_INFINITE_RESPONSE_WAIT); - - /* start the interface's transmit queue, allowing it to accept - * packets for transmission - */ - netif_start_queue(netdev); + visornic_enable_with_timeout(netdev, VISORNIC_INFINITE_RSP_WAIT); return 0; } @@ -928,13 +757,59 @@ visornic_open(struct net_device *netdev) static int visornic_close(struct net_device *netdev) { - netif_stop_queue(netdev); - visornic_disable_with_timeout(netdev, VISORNIC_INFINITE_RESPONSE_WAIT); + visornic_disable_with_timeout(netdev, VISORNIC_INFINITE_RSP_WAIT); return 0; } /** + * devdata_xmits_outstanding - compute outstanding xmits + * @devdata: visornic_devdata for device + * + * Return value is the number of outstanding xmits. + */ +static unsigned long devdata_xmits_outstanding(struct visornic_devdata *devdata) +{ + if (devdata->chstat.sent_xmit >= devdata->chstat.got_xmit_done) + return devdata->chstat.sent_xmit - + devdata->chstat.got_xmit_done; + else + return (ULONG_MAX - devdata->chstat.got_xmit_done + + devdata->chstat.sent_xmit + 1); +} + +/** + * vnic_hit_high_watermark + * @devdata: indicates visornic device we are checking + * @high_watermark: max num of unacked xmits we will tolerate, + * before we will start throttling + * + * Returns true iff the number of unacked xmits sent to + * the IO partition is >= high_watermark. + */ +static inline bool vnic_hit_high_watermark(struct visornic_devdata *devdata, + ulong high_watermark) +{ + return (devdata_xmits_outstanding(devdata) >= high_watermark); +} + +/** + * vnic_hit_low_watermark + * @devdata: indicates visornic device we are checking + * @low_watermark: we will wait until the num of unacked xmits + * drops to this value or lower before we start + * transmitting again + * + * Returns true iff the number of unacked xmits sent to + * the IO partition is <= low_watermark. + */ +static inline bool vnic_hit_low_watermark(struct visornic_devdata *devdata, + ulong low_watermark) +{ + return (devdata_xmits_outstanding(devdata) <= low_watermark); +} + +/** * visornic_xmit - send a packet to the IO Partition * @skb: Packet to be sent * @netdev: net device the packet is being sent from @@ -944,7 +819,7 @@ visornic_close(struct net_device *netdev) * function is protected from concurrent calls by a spinlock xmit_lock * in the net_device struct, but as soon as the function returns it * can be called again. - * Returns NETDEV_TX_OK for success, NETDEV_TX_BUSY for error. + * Returns NETDEV_TX_OK. */ static int visornic_xmit(struct sk_buff *skb, struct net_device *netdev) @@ -961,7 +836,10 @@ visornic_xmit(struct sk_buff *skb, struct net_device *netdev) devdata->server_change_state) { spin_unlock_irqrestore(&devdata->priv_lock, flags); devdata->busy_cnt++; - return NETDEV_TX_BUSY; + dev_dbg(&netdev->dev, + "%s busy - queue stopped\n", __func__); + kfree_skb(skb); + return NETDEV_TX_OK; } /* sk_buff struct is used to host network data throughout all the @@ -979,7 +857,11 @@ visornic_xmit(struct sk_buff *skb, struct net_device *netdev) if (firstfraglen < ETH_HEADER_SIZE) { spin_unlock_irqrestore(&devdata->priv_lock, flags); devdata->busy_cnt++; - return NETDEV_TX_BUSY; + dev_err(&netdev->dev, + "%s busy - first frag too small (%d)\n", + __func__, firstfraglen); + kfree_skb(skb); + return NETDEV_TX_OK; } if ((len < ETH_MIN_PACKET_SIZE) && @@ -1002,13 +884,8 @@ visornic_xmit(struct sk_buff *skb, struct net_device *netdev) /* save the pointer to skb -- we'll need it for completion */ cmdrsp->net.buf = skb; - if (((devdata->chstat.sent_xmit >= devdata->chstat.got_xmit_done) && - (devdata->chstat.sent_xmit - devdata->chstat.got_xmit_done >= - devdata->max_outstanding_net_xmits)) || - ((devdata->chstat.sent_xmit < devdata->chstat.got_xmit_done) && - (ULONG_MAX - devdata->chstat.got_xmit_done + - devdata->chstat.sent_xmit >= - devdata->max_outstanding_net_xmits))) { + if (vnic_hit_high_watermark(devdata, + devdata->max_outstanding_net_xmits)) { /* too many NET_XMITs queued over to IOVM - need to wait */ devdata->chstat.reject_count++; @@ -1018,7 +895,11 @@ visornic_xmit(struct sk_buff *skb, struct net_device *netdev) netif_stop_queue(netdev); spin_unlock_irqrestore(&devdata->priv_lock, flags); devdata->busy_cnt++; - return NETDEV_TX_BUSY; + dev_dbg(&netdev->dev, + "%s busy - waiting for iovm to catch up\n", + __func__); + kfree_skb(skb); + return NETDEV_TX_OK; } if (devdata->queuefullmsg_logged) devdata->queuefullmsg_logged = 0; @@ -1055,10 +936,13 @@ visornic_xmit(struct sk_buff *skb, struct net_device *netdev) visor_copy_fragsinfo_from_skb(skb, firstfraglen, MAX_PHYS_INFO, cmdrsp->net.xmt.frags); - if (cmdrsp->net.xmt.num_frags == -1) { + if (cmdrsp->net.xmt.num_frags < 0) { spin_unlock_irqrestore(&devdata->priv_lock, flags); devdata->busy_cnt++; - return NETDEV_TX_BUSY; + dev_err(&netdev->dev, + "%s busy - copy frags failed\n", __func__); + kfree_skb(skb); + return NETDEV_TX_OK; } if (!visorchannel_signalinsert(devdata->dev->visorchannel, @@ -1066,18 +950,15 @@ visornic_xmit(struct sk_buff *skb, struct net_device *netdev) netif_stop_queue(netdev); spin_unlock_irqrestore(&devdata->priv_lock, flags); devdata->busy_cnt++; - return NETDEV_TX_BUSY; + dev_dbg(&netdev->dev, + "%s busy - signalinsert failed\n", __func__); + kfree_skb(skb); + return NETDEV_TX_OK; } /* Track the skbs that have been sent to the IOVM for XMIT */ skb_queue_head(&devdata->xmitbufhead, skb); - /* set the last transmission start time - * linux doc says: Do not forget to update netdev->trans_start to - * jiffies after each new tx packet is given to the hardware. - */ - netdev->trans_start = jiffies; - /* update xmt stats */ devdata->net_stats.tx_packets++; devdata->net_stats.tx_bytes += skb->len; @@ -1086,18 +967,16 @@ visornic_xmit(struct sk_buff *skb, struct net_device *netdev) /* check to see if we have hit the high watermark for * netif_stop_queue() */ - if (((devdata->chstat.sent_xmit >= devdata->chstat.got_xmit_done) && - (devdata->chstat.sent_xmit - devdata->chstat.got_xmit_done >= - devdata->upper_threshold_net_xmits)) || - ((devdata->chstat.sent_xmit < devdata->chstat.got_xmit_done) && - (ULONG_MAX - devdata->chstat.got_xmit_done + - devdata->chstat.sent_xmit >= - devdata->upper_threshold_net_xmits))) { + if (vnic_hit_high_watermark(devdata, + devdata->upper_threshold_net_xmits)) { /* too many NET_XMITs queued over to IOVM - need to wait */ netif_stop_queue(netdev); /* calling stop queue - call * netif_wake_queue() after lower * threshold */ + dev_dbg(&netdev->dev, + "%s busy - invoking iovm flow control\n", + __func__); devdata->flow_control_upper_hits++; } spin_unlock_irqrestore(&devdata->priv_lock, flags); @@ -1121,21 +1000,6 @@ visornic_get_stats(struct net_device *netdev) } /** - * visornic_ioctl - ioctl function for netdevice. - * @netdev: netdevice - * @ifr: ignored - * @cmd: ignored - * - * Currently not supported. - * Returns EOPNOTSUPP - */ -static int -visornic_ioctl(struct net_device *netdev, struct ifreq *ifr, int cmd) -{ - return -EOPNOTSUPP; -} - -/** * visornic_change_mtu - changes mtu of device. * @netdev: netdevice * @new_mtu: value of new mtu @@ -1201,15 +1065,24 @@ visornic_xmit_timeout(struct net_device *netdev) unsigned long flags; spin_lock_irqsave(&devdata->priv_lock, flags); + if (devdata->going_away) { + spin_unlock_irqrestore(&devdata->priv_lock, flags); + dev_dbg(&devdata->dev->device, + "%s aborting because device removal pending\n", + __func__); + return; + } + /* Ensure that a ServerDown message hasn't been received */ if (!devdata->enabled || (devdata->server_down && !devdata->server_change_state)) { + dev_dbg(&netdev->dev, "%s no processing\n", + __func__); spin_unlock_irqrestore(&devdata->priv_lock, flags); return; } - spin_unlock_irqrestore(&devdata->priv_lock, flags); - queue_work(visornic_timeout_reset_workqueue, &devdata->timeout_reset); + spin_unlock_irqrestore(&devdata->priv_lock, flags); } /** @@ -1281,7 +1154,6 @@ repost_return(struct uiscmdrsp *cmdrsp, struct visornic_devdata *devdata, devdata->bad_rcv_buf++; } } - atomic_dec(&devdata->usage); return status; } @@ -1293,18 +1165,16 @@ repost_return(struct uiscmdrsp *cmdrsp, struct visornic_devdata *devdata, * it up the stack. * Returns void */ -static void +static int visornic_rx(struct uiscmdrsp *cmdrsp) { struct visornic_devdata *devdata; struct sk_buff *skb, *prev, *curr; struct net_device *netdev; - int cc, currsize, off, status; + int cc, currsize, off; struct ethhdr *eth; unsigned long flags; -#ifdef DEBUG - struct phys_info testfrags[MAX_PHYS_INFO]; -#endif + int rx_count = 0; /* post new rcv buf to the other end using the cmdrsp we have at hand * post it without holding lock - but we'll use the signal lock to @@ -1314,37 +1184,21 @@ visornic_rx(struct uiscmdrsp *cmdrsp) skb = cmdrsp->net.buf; netdev = skb->dev; - if (!netdev) { - /* We must have previously downed this network device and - * this skb and device is no longer valid. This also means - * the skb reference was removed from devdata->rcvbuf so no - * need to search for it. - * All we can do is free the skb and return. - * Note: We crash if we try to log this here. - */ - kfree_skb(skb); - return; - } - devdata = netdev_priv(netdev); spin_lock_irqsave(&devdata->priv_lock, flags); atomic_dec(&devdata->num_rcvbuf_in_iovm); - /* update rcv stats - call it with priv_lock held */ - devdata->net_stats.rx_packets++; - devdata->net_stats.rx_bytes = skb->len; - - atomic_inc(&devdata->usage); /* don't want a close to happen before - * we're done here - */ - /* set length to how much was ACTUALLY received - * NOTE: rcv_done_len includes actual length of data rcvd * including ethhdr */ skb->len = cmdrsp->net.rcv.rcv_done_len; + /* update rcv stats - call it with priv_lock held */ + devdata->net_stats.rx_packets++; + devdata->net_stats.rx_bytes += skb->len; + /* test enabled while holding lock */ if (!(devdata->enabled && devdata->enab_dis_acked)) { /* don't process it unless we're in enable mode and until @@ -1352,7 +1206,7 @@ visornic_rx(struct uiscmdrsp *cmdrsp) */ spin_unlock_irqrestore(&devdata->priv_lock, flags); repost_return(cmdrsp, devdata, skb, netdev); - return; + return rx_count; } spin_unlock_irqrestore(&devdata->priv_lock, flags); @@ -1371,7 +1225,7 @@ visornic_rx(struct uiscmdrsp *cmdrsp) if (repost_return(cmdrsp, devdata, skb, netdev) < 0) dev_err(&devdata->netdev->dev, "repost_return failed"); - return; + return rx_count; } /* length rcvd is greater than firstfrag in this skb rcv buf */ skb->tail += RCVPOST_BUF_SIZE; /* amount in skb->data */ @@ -1386,7 +1240,7 @@ visornic_rx(struct uiscmdrsp *cmdrsp) if (repost_return(cmdrsp, devdata, skb, netdev) < 0) dev_err(&devdata->netdev->dev, "repost_return failed"); - return; + return rx_count; } skb->tail += skb->len; skb->data_len = 0; /* nothing rcvd in frag_list */ @@ -1405,7 +1259,7 @@ visornic_rx(struct uiscmdrsp *cmdrsp) if (cmdrsp->net.rcv.rcvbuf[0] != skb) { if (repost_return(cmdrsp, devdata, skb, netdev) < 0) dev_err(&devdata->netdev->dev, "repost_return failed"); - return; + return rx_count; } if (cmdrsp->net.rcv.numrcvbufs > 1) { @@ -1431,29 +1285,12 @@ visornic_rx(struct uiscmdrsp *cmdrsp) curr->data_len = 0; off += currsize; } -#ifdef DEBUG /* assert skb->len == off */ if (skb->len != off) { - dev_err(&devdata->netdev->dev, - "%s something wrong; skb->len:%d != off:%d\n", - netdev->name, skb->len, off); - } - /* test code */ - cc = util_copy_fragsinfo_from_skb("rcvchaintest", skb, - RCVPOST_BUF_SIZE, - MAX_PHYS_INFO, testfrags); - if (cc != cmdrsp->net.rcv.numrcvbufs) { - dev_err(&devdata->netdev->dev, - "**** %s Something wrong; rcvd chain length %d different from one we calculated %d\n", - netdev->name, cmdrsp->net.rcv.numrcvbufs, cc); + netdev_err(devdata->netdev, + "something wrong; skb->len:%d != off:%d\n", + skb->len, off); } - for (i = 0; i < cc; i++) { - dev_inf(&devdata->netdev->dev, - "test:RCVPOST_BUF_SIZE:%d[%d] pfn:%llu off:0x%x len:%d\n", - RCVPOST_BUF_SIZE, i, testfrags[i].pi_pfn, - testfrags[i].pi_off, testfrags[i].pi_len); - } -#endif } /* set up packet's protocl type using ethernet header - this @@ -1505,10 +1342,11 @@ visornic_rx(struct uiscmdrsp *cmdrsp) /* drop packet - don't forward it up to OS */ devdata->n_rcv_packets_not_accepted++; repost_return(cmdrsp, devdata, skb, netdev); - return; + return rx_count; } while (0); - status = netif_rx(skb); + rx_count++; + netif_receive_skb(skb); /* netif_rx returns various values, but "in practice most drivers * ignore the return value */ @@ -1520,6 +1358,7 @@ visornic_rx(struct uiscmdrsp *cmdrsp) * new rcv buffer. */ repost_return(cmdrsp, devdata, skb, netdev); + return rx_count; } /** @@ -1545,14 +1384,11 @@ devdata_initialize(struct visornic_devdata *devdata, struct visor_device *dev) spin_unlock(&dev_num_pool_lock); if (devnum == MAXDEVICES) devnum = -1; - if (devnum < 0) { - kfree(devdata); + if (devnum < 0) return NULL; - } devdata->devnum = devnum; devdata->dev = dev; strncpy(devdata->name, dev_name(&dev->device), sizeof(devdata->name)); - kref_init(&devdata->kref); spin_lock(&lock_all_devices); list_add_tail(&devdata->list_all, &list_all_devices); spin_unlock(&lock_all_devices); @@ -1560,24 +1396,23 @@ devdata_initialize(struct visornic_devdata *devdata, struct visor_device *dev) } /** - * devdata_release - Frees up a devdata - * @mykref: kref to the devdata + * devdata_release - Frees up references in devdata + * @devdata: struct to clean up * - * Frees up a devdata. + * Frees up references in devdata. * Returns void */ -static void devdata_release(struct kref *mykref) +static void devdata_release(struct visornic_devdata *devdata) { - struct visornic_devdata *devdata = - container_of(mykref, struct visornic_devdata, kref); - spin_lock(&dev_num_pool_lock); clear_bit(devdata->devnum, dev_num_pool); spin_unlock(&dev_num_pool_lock); spin_lock(&lock_all_devices); list_del(&devdata->list_all); spin_unlock(&lock_all_devices); - kfree(devdata); + kfree(devdata->rcvbuf); + kfree(devdata->cmdrsp_rcv); + kfree(devdata->xmit_cmdrsp); } static const struct net_device_ops visornic_dev_ops = { @@ -1585,12 +1420,163 @@ static const struct net_device_ops visornic_dev_ops = { .ndo_stop = visornic_close, .ndo_start_xmit = visornic_xmit, .ndo_get_stats = visornic_get_stats, - .ndo_do_ioctl = visornic_ioctl, .ndo_change_mtu = visornic_change_mtu, .ndo_tx_timeout = visornic_xmit_timeout, .ndo_set_rx_mode = visornic_set_multi, }; +/* DebugFS code */ +static ssize_t info_debugfs_read(struct file *file, char __user *buf, + size_t len, loff_t *offset) +{ + ssize_t bytes_read = 0; + int str_pos = 0; + struct visornic_devdata *devdata; + struct net_device *dev; + char *vbuf; + + if (len > MAX_BUF) + len = MAX_BUF; + vbuf = kzalloc(len, GFP_KERNEL); + if (!vbuf) + return -ENOMEM; + + /* for each vnic channel + * dump out channel specific data + */ + rcu_read_lock(); + for_each_netdev_rcu(current->nsproxy->net_ns, dev) { + /* + * Only consider netdevs that are visornic, and are open + */ + if ((dev->netdev_ops != &visornic_dev_ops) || + (!netif_queue_stopped(dev))) + continue; + + devdata = netdev_priv(dev); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + "netdev = %s (0x%p), MAC Addr %pM\n", + dev->name, + dev, + dev->dev_addr); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + "VisorNic Dev Info = 0x%p\n", devdata); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " num_rcv_bufs = %d\n", + devdata->num_rcv_bufs); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " max_oustanding_next_xmits = %lu\n", + devdata->max_outstanding_net_xmits); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " upper_threshold_net_xmits = %lu\n", + devdata->upper_threshold_net_xmits); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " lower_threshold_net_xmits = %lu\n", + devdata->lower_threshold_net_xmits); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " queuefullmsg_logged = %d\n", + devdata->queuefullmsg_logged); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " chstat.got_rcv = %lu\n", + devdata->chstat.got_rcv); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " chstat.got_enbdisack = %lu\n", + devdata->chstat.got_enbdisack); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " chstat.got_xmit_done = %lu\n", + devdata->chstat.got_xmit_done); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " chstat.xmit_fail = %lu\n", + devdata->chstat.xmit_fail); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " chstat.sent_enbdis = %lu\n", + devdata->chstat.sent_enbdis); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " chstat.sent_promisc = %lu\n", + devdata->chstat.sent_promisc); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " chstat.sent_post = %lu\n", + devdata->chstat.sent_post); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " chstat.sent_post_failed = %lu\n", + devdata->chstat.sent_post_failed); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " chstat.sent_xmit = %lu\n", + devdata->chstat.sent_xmit); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " chstat.reject_count = %lu\n", + devdata->chstat.reject_count); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " chstat.extra_rcvbufs_sent = %lu\n", + devdata->chstat.extra_rcvbufs_sent); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " n_rcv0 = %lu\n", devdata->n_rcv0); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " n_rcv1 = %lu\n", devdata->n_rcv1); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " n_rcv2 = %lu\n", devdata->n_rcv2); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " n_rcvx = %lu\n", devdata->n_rcvx); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " num_rcvbuf_in_iovm = %d\n", + atomic_read(&devdata->num_rcvbuf_in_iovm)); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " alloc_failed_in_if_needed_cnt = %lu\n", + devdata->alloc_failed_in_if_needed_cnt); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " alloc_failed_in_repost_rtn_cnt = %lu\n", + devdata->alloc_failed_in_repost_rtn_cnt); + /* str_pos += scnprintf(vbuf + str_pos, len - str_pos, + * " inner_loop_limit_reached_cnt = %lu\n", + * devdata->inner_loop_limit_reached_cnt); + */ + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " found_repost_rcvbuf_cnt = %lu\n", + devdata->found_repost_rcvbuf_cnt); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " repost_found_skb_cnt = %lu\n", + devdata->repost_found_skb_cnt); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " n_repost_deficit = %lu\n", + devdata->n_repost_deficit); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " bad_rcv_buf = %lu\n", + devdata->bad_rcv_buf); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " n_rcv_packets_not_accepted = %lu\n", + devdata->n_rcv_packets_not_accepted); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " interrupts_rcvd = %llu\n", + devdata->interrupts_rcvd); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " interrupts_notme = %llu\n", + devdata->interrupts_notme); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " interrupts_disabled = %llu\n", + devdata->interrupts_disabled); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " busy_cnt = %llu\n", + devdata->busy_cnt); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " flow_control_upper_hits = %llu\n", + devdata->flow_control_upper_hits); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " flow_control_lower_hits = %llu\n", + devdata->flow_control_lower_hits); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " netif_queue = %s\n", + netif_queue_stopped(devdata->netdev) ? + "stopped" : "running"); + str_pos += scnprintf(vbuf + str_pos, len - str_pos, + " xmits_outstanding = %lu\n", + devdata_xmits_outstanding(devdata)); + } + rcu_read_unlock(); + bytes_read = simple_read_from_buffer(buf, len, offset, vbuf, str_pos); + kfree(vbuf); + return bytes_read; +} + /** * send_rcv_posts_if_needed * @devdata: visornic device @@ -1644,15 +1630,15 @@ send_rcv_posts_if_needed(struct visornic_devdata *devdata) * Returns when response queue is empty or when the threadd stops. */ static void -drain_queue(struct uiscmdrsp *cmdrsp, struct visornic_devdata *devdata) +service_resp_queue(struct uiscmdrsp *cmdrsp, struct visornic_devdata *devdata, + int *rx_work_done) { unsigned long flags; struct net_device *netdev; - /* drain queue */ - while (1) { - /* TODO: CLIENT ACQUIRE -- Don't really need this at the - * moment */ + /* TODO: CLIENT ACQUIRE -- Don't really need this at the + * moment */ + for (;;) { if (!visorchannel_signalremove(devdata->dev->visorchannel, IOCHAN_FROM_IOPART, cmdrsp)) @@ -1662,7 +1648,7 @@ drain_queue(struct uiscmdrsp *cmdrsp, struct visornic_devdata *devdata) case NET_RCV: devdata->chstat.got_rcv++; /* process incoming packet */ - visornic_rx(cmdrsp); + *rx_work_done += visornic_rx(cmdrsp); break; case NET_XMIT_DONE: spin_lock_irqsave(&devdata->priv_lock, flags); @@ -1678,16 +1664,8 @@ drain_queue(struct uiscmdrsp *cmdrsp, struct visornic_devdata *devdata) * the lower watermark for * netif_wake_queue() */ - if (((devdata->chstat.sent_xmit >= - devdata->chstat.got_xmit_done) && - (devdata->chstat.sent_xmit - - devdata->chstat.got_xmit_done <= - devdata->lower_threshold_net_xmits)) || - ((devdata->chstat.sent_xmit < - devdata->chstat.got_xmit_done) && - (ULONG_MAX - devdata->chstat.got_xmit_done - + devdata->chstat.sent_xmit <= - devdata->lower_threshold_net_xmits))) { + if (vnic_hit_low_watermark(devdata, + devdata->lower_threshold_net_xmits)) { /* enough NET_XMITs completed * so can restart netif queue */ @@ -1738,50 +1716,51 @@ drain_queue(struct uiscmdrsp *cmdrsp, struct visornic_devdata *devdata) break; } /* cmdrsp is now available for reuse */ - - if (kthread_should_stop()) - break; } } +static int visornic_poll(struct napi_struct *napi, int budget) +{ + struct visornic_devdata *devdata = container_of(napi, + struct visornic_devdata, + napi); + int rx_count = 0; + + send_rcv_posts_if_needed(devdata); + service_resp_queue(devdata->cmdrsp, devdata, &rx_count); + + /* + * If there aren't any more packets to receive + * stop the poll + */ + if (rx_count < budget) + napi_complete(napi); + + return rx_count; +} + /** - * process_incoming_rsps - Checks the status of the response queue. + * poll_for_irq - Checks the status of the response queue. * @v: void pointer to the visronic devdata * * Main function of the vnic_incoming thread. Peridocially check the * response queue and drain it if needed. * Returns when thread has stopped. */ -static int -process_incoming_rsps(void *v) +static void +poll_for_irq(unsigned long v) { - struct visornic_devdata *devdata = v; - struct uiscmdrsp *cmdrsp = NULL; - const int SZ = SIZEOF_CMDRSP; + struct visornic_devdata *devdata = (struct visornic_devdata *)v; - cmdrsp = kmalloc(SZ, GFP_ATOMIC); - if (!cmdrsp) - complete_and_exit(&devdata->threadinfo.has_stopped, 0); + if (!visorchannel_signalempty( + devdata->dev->visorchannel, + IOCHAN_FROM_IOPART)) + napi_schedule(&devdata->napi); - while (1) { - wait_event_interruptible_timeout( - devdata->rsp_queue, (atomic_read( - &devdata->interrupt_rcvd) == 1), - msecs_to_jiffies(devdata->thread_wait_ms)); + atomic_set(&devdata->interrupt_rcvd, 0); - /* periodically check to see if there are any rcf bufs which - * need to get sent to the IOSP. This can only happen if - * we run out of memory when trying to allocate skbs. - */ - atomic_set(&devdata->interrupt_rcvd, 0); - send_rcv_posts_if_needed(devdata); - drain_queue(cmdrsp, devdata); - if (kthread_should_stop()) - break; - } + mod_timer(&devdata->irq_poll_timer, msecs_to_jiffies(2)); - kfree(cmdrsp); - complete_and_exit(&devdata->threadinfo.has_stopped, 0); } /** @@ -1801,12 +1780,15 @@ static int visornic_probe(struct visor_device *dev) u64 features; netdev = alloc_etherdev(sizeof(struct visornic_devdata)); - if (!netdev) + if (!netdev) { + dev_err(&dev->device, + "%s alloc_etherdev failed\n", __func__); return -ENOMEM; + } netdev->netdev_ops = &visornic_dev_ops; netdev->watchdog_timeo = (5 * HZ); - netdev->dev.parent = &dev->device; + SET_NETDEV_DEV(netdev, &dev->device); /* Get MAC adddress from channel and read it into the device. */ netdev->addr_len = ETH_ALEN; @@ -1814,16 +1796,23 @@ static int visornic_probe(struct visor_device *dev) vnic.macaddr); err = visorbus_read_channel(dev, channel_offset, netdev->dev_addr, ETH_ALEN); - if (err < 0) + if (err < 0) { + dev_err(&dev->device, + "%s failed to get mac addr from chan (%d)\n", + __func__, err); goto cleanup_netdev; + } devdata = devdata_initialize(netdev_priv(netdev), dev); if (!devdata) { + dev_err(&dev->device, + "%s devdata_initialize failed\n", __func__); err = -ENOMEM; goto cleanup_netdev; } devdata->netdev = netdev; + dev_set_drvdata(&dev->device, devdata); init_waitqueue_head(&devdata->rsp_queue); spin_lock_init(&devdata->priv_lock); devdata->enabled = 0; /* not yet */ @@ -1834,10 +1823,14 @@ static int visornic_probe(struct visor_device *dev) vnic.num_rcv_bufs); err = visorbus_read_channel(dev, channel_offset, &devdata->num_rcv_bufs, 4); - if (err) + if (err) { + dev_err(&dev->device, + "%s failed to get #rcv bufs from chan (%d)\n", + __func__, err); goto cleanup_netdev; + } - devdata->rcvbuf = kmalloc(sizeof(struct sk_buff *) * + devdata->rcvbuf = kzalloc(sizeof(struct sk_buff *) * devdata->num_rcv_bufs, GFP_KERNEL); if (!devdata->rcvbuf) { err = -ENOMEM; @@ -1846,12 +1839,15 @@ static int visornic_probe(struct visor_device *dev) /* set the net_xmit outstanding threshold */ /* always leave two slots open but you should have 3 at a minimum */ + /* note that max_outstanding_net_xmits must be > 0 */ devdata->max_outstanding_net_xmits = - max(3, ((devdata->num_rcv_bufs / 3) - 2)); + max_t(unsigned long, 3, ((devdata->num_rcv_bufs / 3) - 2)); devdata->upper_threshold_net_xmits = - max(2, devdata->max_outstanding_net_xmits - 1); + max_t(unsigned long, + 2, (devdata->max_outstanding_net_xmits - 1)); devdata->lower_threshold_net_xmits = - max(1, devdata->max_outstanding_net_xmits / 2); + max_t(unsigned long, + 1, (devdata->max_outstanding_net_xmits / 2)); skb_queue_head_init(&devdata->xmitbufhead); @@ -1866,8 +1862,6 @@ static int visornic_probe(struct visor_device *dev) err = -ENOMEM; goto cleanup_xmit_cmdrsp; } - INIT_WORK(&devdata->serverdown_completion, - visornic_serverdown_complete); INIT_WORK(&devdata->timeout_reset, visornic_timeout_reset); devdata->server_down = false; devdata->server_change_state = false; @@ -1876,42 +1870,73 @@ static int visornic_probe(struct visor_device *dev) channel_offset = offsetof(struct spar_io_channel_protocol, vnic.mtu); err = visorbus_read_channel(dev, channel_offset, &netdev->mtu, 4); - if (err) + if (err) { + dev_err(&dev->device, + "%s failed to get mtu from chan (%d)\n", + __func__, err); goto cleanup_xmit_cmdrsp; + } /* TODO: Setup Interrupt information */ /* Let's start our threads to get responses */ + netif_napi_add(netdev, &devdata->napi, visornic_poll, 64); + + setup_timer(&devdata->irq_poll_timer, poll_for_irq, + (unsigned long)devdata); + /* + * Note: This time has to start running before the while + * loop below because the napi routine is responsible for + * setting enab_dis_acked + */ + mod_timer(&devdata->irq_poll_timer, msecs_to_jiffies(2)); + channel_offset = offsetof(struct spar_io_channel_protocol, channel_header.features); err = visorbus_read_channel(dev, channel_offset, &features, 8); - if (err) - goto cleanup_xmit_cmdrsp; + if (err) { + dev_err(&dev->device, + "%s failed to get features from chan (%d)\n", + __func__, err); + goto cleanup_napi_add; + } features |= ULTRA_IO_CHANNEL_IS_POLLING; err = visorbus_write_channel(dev, channel_offset, &features, 8); - if (err) - goto cleanup_xmit_cmdrsp; - - devdata->thread_wait_ms = 2; - visor_thread_start(&devdata->threadinfo, process_incoming_rsps, - devdata, "vnic_incoming"); + if (err) { + dev_err(&dev->device, + "%s failed to set features in chan (%d)\n", + __func__, err); + goto cleanup_napi_add; + } err = register_netdev(netdev); - if (err) - goto cleanup_thread_stop; + if (err) { + dev_err(&dev->device, + "%s register_netdev failed (%d)\n", __func__, err); + goto cleanup_napi_add; + } /* create debgug/sysfs directories */ devdata->eth_debugfs_dir = debugfs_create_dir(netdev->name, visornic_debugfs_dir); if (!devdata->eth_debugfs_dir) { + dev_err(&dev->device, + "%s debugfs_create_dir %s failed\n", + __func__, netdev->name); err = -ENOMEM; - goto cleanup_thread_stop; + goto cleanup_register_netdev; } + dev_info(&dev->device, "%s success netdev=%s\n", + __func__, netdev->name); return 0; -cleanup_thread_stop: - visor_thread_stop(&devdata->threadinfo); +cleanup_register_netdev: + unregister_netdev(netdev); + +cleanup_napi_add: + del_timer_sync(&devdata->irq_poll_timer); + netif_napi_del(&devdata->napi); cleanup_xmit_cmdrsp: kfree(devdata->xmit_cmdrsp); @@ -1954,12 +1979,41 @@ static void host_side_disappeared(struct visornic_devdata *devdata) static void visornic_remove(struct visor_device *dev) { struct visornic_devdata *devdata = dev_get_drvdata(&dev->device); + struct net_device *netdev; + unsigned long flags; - if (!devdata) + if (!devdata) { + dev_err(&dev->device, "%s no devdata\n", __func__); + return; + } + spin_lock_irqsave(&devdata->priv_lock, flags); + if (devdata->going_away) { + spin_unlock_irqrestore(&devdata->priv_lock, flags); + dev_err(&dev->device, "%s already being removed\n", __func__); return; + } + devdata->going_away = true; + spin_unlock_irqrestore(&devdata->priv_lock, flags); + netdev = devdata->netdev; + if (!netdev) { + dev_err(&dev->device, "%s not net device\n", __func__); + return; + } + + /* going_away prevents new items being added to the workqueues */ + flush_workqueue(visornic_timeout_reset_workqueue); + + debugfs_remove_recursive(devdata->eth_debugfs_dir); + + unregister_netdev(netdev); /* this will call visornic_close() */ + + del_timer_sync(&devdata->irq_poll_timer); + netif_napi_del(&devdata->napi); + dev_set_drvdata(&dev->device, NULL); host_side_disappeared(devdata); - kref_put(&devdata->kref, devdata_release); + devdata_release(devdata); + free_netdev(netdev); } /** @@ -1980,8 +2034,7 @@ static int visornic_pause(struct visor_device *dev, { struct visornic_devdata *devdata = dev_get_drvdata(&dev->device); - visornic_serverdown(devdata); - complete_func(dev, 0); + visornic_serverdown(devdata, complete_func); return 0; } @@ -2003,37 +2056,40 @@ static int visornic_resume(struct visor_device *dev, unsigned long flags; devdata = dev_get_drvdata(&dev->device); - if (!devdata) + if (!devdata) { + dev_err(&dev->device, "%s no devdata\n", __func__); return -EINVAL; + } netdev = devdata->netdev; - if (devdata->server_down && !devdata->server_change_state) { - devdata->server_change_state = true; - /* Must transition channel to ATTACHED state BEFORE - * we can start using the device again. - * TODO: State transitions - */ - visor_thread_start(&devdata->threadinfo, process_incoming_rsps, - devdata, "vnic_incoming"); - init_rcv_bufs(netdev, devdata); - spin_lock_irqsave(&devdata->priv_lock, flags); - devdata->enabled = 1; - - /* Now we're ready, let's send an ENB to uisnic but until - * we get an ACK back from uisnic, we'll drop the packets - */ - devdata->enab_dis_acked = 0; + spin_lock_irqsave(&devdata->priv_lock, flags); + if (devdata->server_change_state) { spin_unlock_irqrestore(&devdata->priv_lock, flags); - - /* send enable and wait for ack - don't hold lock when - * sending enable because if the queue if sull, insert - * might sleep. - */ - send_enbdis(netdev, 1, devdata); - } else if (devdata->server_change_state) { - return -EIO; + dev_err(&dev->device, "%s server already changing state\n", + __func__); + return -EINVAL; + } + if (!devdata->server_down) { + spin_unlock_irqrestore(&devdata->priv_lock, flags); + dev_err(&dev->device, "%s server not down\n", __func__); + complete_func(dev, 0); + return 0; } + devdata->server_change_state = true; + spin_unlock_irqrestore(&devdata->priv_lock, flags); + + /* Must transition channel to ATTACHED state BEFORE + * we can start using the device again. + * TODO: State transitions + */ + mod_timer(&devdata->irq_poll_timer, msecs_to_jiffies(2)); + + init_rcv_bufs(netdev, devdata); + + rtnl_lock(); + dev_open(netdev); + rtnl_unlock(); complete_func(dev, 0); return 0; @@ -2051,18 +2107,6 @@ static int visornic_init(void) struct dentry *ret; int err = -ENOMEM; - /* create workqueue for serverdown completion */ - visornic_serverdown_workqueue = - create_singlethread_workqueue("visornic_serverdown"); - if (!visornic_serverdown_workqueue) - return -ENOMEM; - - /* create workqueue for tx timeout reset */ - visornic_timeout_reset_workqueue = - create_singlethread_workqueue("visornic_timeout_reset"); - if (!visornic_timeout_reset_workqueue) - return -ENOMEM; - visornic_debugfs_dir = debugfs_create_dir("visornic", NULL); if (!visornic_debugfs_dir) return err; @@ -2076,12 +2120,6 @@ static int visornic_init(void) if (!ret) goto cleanup_debugfs; - /* create workqueue for serverdown completion */ - visornic_serverdown_workqueue = - create_singlethread_workqueue("visornic_serverdown"); - if (!visornic_serverdown_workqueue) - goto cleanup_debugfs; - /* create workqueue for tx timeout reset */ visornic_timeout_reset_workqueue = create_singlethread_workqueue("visornic_timeout_reset"); @@ -2093,12 +2131,11 @@ static int visornic_init(void) if (!dev_num_pool) goto cleanup_workqueue; - visorbus_register_visor_driver(&visornic_driver); - return 0; + err = visorbus_register_visor_driver(&visornic_driver); + if (!err) + return 0; cleanup_workqueue: - flush_workqueue(visornic_serverdown_workqueue); - destroy_workqueue(visornic_serverdown_workqueue); if (visornic_timeout_reset_workqueue) { flush_workqueue(visornic_timeout_reset_workqueue); destroy_workqueue(visornic_timeout_reset_workqueue); @@ -2116,17 +2153,14 @@ cleanup_debugfs: */ static void visornic_cleanup(void) { - if (visornic_serverdown_workqueue) { - flush_workqueue(visornic_serverdown_workqueue); - destroy_workqueue(visornic_serverdown_workqueue); - } + visorbus_unregister_visor_driver(&visornic_driver); + if (visornic_timeout_reset_workqueue) { flush_workqueue(visornic_timeout_reset_workqueue); destroy_workqueue(visornic_timeout_reset_workqueue); } debugfs_remove_recursive(visornic_debugfs_dir); - visorbus_unregister_visor_driver(&visornic_driver); kfree(dev_num_pool); dev_num_pool = NULL; } |