2005-04-17 02:20:36 +04:00
|
|
|
/*
|
|
|
|
* net/sched/sch_generic.c Generic packet scheduler routines.
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU General Public License
|
|
|
|
* as published by the Free Software Foundation; either version
|
|
|
|
* 2 of the License, or (at your option) any later version.
|
|
|
|
*
|
|
|
|
* Authors: Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru>
|
|
|
|
* Jamal Hadi Salim, <hadi@cyberus.ca> 990601
|
|
|
|
* - Ingress support
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/bitops.h>
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/types.h>
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/sched.h>
|
|
|
|
#include <linux/string.h>
|
|
|
|
#include <linux/errno.h>
|
|
|
|
#include <linux/netdevice.h>
|
|
|
|
#include <linux/skbuff.h>
|
|
|
|
#include <linux/rtnetlink.h>
|
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/rcupdate.h>
|
|
|
|
#include <linux/list.h>
|
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 11:04:11 +03:00
|
|
|
#include <linux/slab.h>
|
2013-08-04 00:07:47 +04:00
|
|
|
#include <linux/if_vlan.h>
|
2017-12-06 18:50:28 +03:00
|
|
|
#include <linux/if_macvlan.h>
|
2013-02-12 04:12:03 +04:00
|
|
|
#include <net/sch_generic.h>
|
2005-04-17 02:20:36 +04:00
|
|
|
#include <net/pkt_sched.h>
|
2010-05-12 03:19:48 +04:00
|
|
|
#include <net/dst.h>
|
2017-08-15 22:11:03 +03:00
|
|
|
#include <trace/events/qdisc.h>
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2013-08-31 21:15:33 +04:00
|
|
|
/* Qdisc to use by default */
|
|
|
|
const struct Qdisc_ops *default_qdisc_ops = &pfifo_fast_ops;
|
|
|
|
EXPORT_SYMBOL(default_qdisc_ops);
|
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
/* Main transmission queue. */
|
|
|
|
|
2007-04-17 04:02:10 +04:00
|
|
|
/* Modifications to data participating in scheduling must be protected with
|
2008-08-03 07:02:43 +04:00
|
|
|
* qdisc_lock(qdisc) spinlock.
|
2007-04-17 04:02:10 +04:00
|
|
|
*
|
|
|
|
* The idea is the following:
|
2008-07-16 14:22:39 +04:00
|
|
|
* - enqueue, dequeue are serialized via qdisc root lock
|
|
|
|
* - ingress filtering is also serialized via qdisc root lock
|
2007-04-17 04:02:10 +04:00
|
|
|
* - updates to tree and tree walking are only done under the rtnl mutex.
|
2005-04-17 02:20:36 +04:00
|
|
|
*/
|
|
|
|
|
2008-07-16 13:15:04 +04:00
|
|
|
static inline int dev_requeue_skb(struct sk_buff *skb, struct Qdisc *q)
|
2007-06-11 04:31:24 +04:00
|
|
|
{
|
2008-10-06 21:41:50 +04:00
|
|
|
q->gso_skb = skb;
|
2008-10-08 22:36:22 +04:00
|
|
|
q->qstats.requeues++;
|
2016-06-04 01:05:57 +03:00
|
|
|
qdisc_qstats_backlog_inc(q, skb);
|
2009-08-06 05:44:21 +04:00
|
|
|
q->q.qlen++; /* it's still part of the queue */
|
2008-07-16 13:15:04 +04:00
|
|
|
__netif_schedule(q);
|
2008-10-06 21:41:50 +04:00
|
|
|
|
2007-06-11 04:31:24 +04:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2014-10-04 02:31:07 +04:00
|
|
|
static void try_bulk_dequeue_skb(struct Qdisc *q,
|
|
|
|
struct sk_buff *skb,
|
net_sched: restore qdisc quota fairness limits after bulk dequeue
Restore the quota fairness between qdisc's, that we broke with commit
5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE").
Before that commit, the quota in __qdisc_run() were in packets as
dequeue_skb() would only dequeue a single packet, that assumption
broke with bulk dequeue.
We choose not to account for the number of packets inside the TSO/GSO
packets (accessable via "skb_gso_segs"). As the previous fairness
also had this "defect". Thus, GSO/TSO packets counts as a single
packet.
Further more, we choose to slack on accuracy, by allowing a bulk
dequeue try_bulk_dequeue_skb() to exceed the "packets" limit, only
limited by the BQL bytelimit. This is done because BQL prefers to get
its full budget for appropriate feedback from TX completion.
In future, we might consider reworking this further and, if it allows,
switch to a time-based model, as suggested by Eric. Right now, we only
restore old semantics.
Joint work with Eric, Hannes, Daniel and Jesper. Hannes wrote the
first patch in cooperation with Daniel and Jesper. Eric rewrote the
patch.
Fixes: 5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE")
Signed-off-by: Eric Dumazet <edumazet@google.com>
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-09 14:18:10 +04:00
|
|
|
const struct netdev_queue *txq,
|
|
|
|
int *packets)
|
qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE
Based on DaveM's recent API work on dev_hard_start_xmit(), that allows
sending/processing an entire skb list.
This patch implements qdisc bulk dequeue, by allowing multiple packets
to be dequeued in dequeue_skb().
The optimization principle for this is two fold, (1) to amortize
locking cost and (2) avoid expensive tailptr update for notifying HW.
(1) Several packets are dequeued while holding the qdisc root_lock,
amortizing locking cost over several packet. The dequeued SKB list is
processed under the TXQ lock in dev_hard_start_xmit(), thus also
amortizing the cost of the TXQ lock.
(2) Further more, dev_hard_start_xmit() will utilize the skb->xmit_more
API to delay HW tailptr update, which also reduces the cost per
packet.
One restriction of the new API is that every SKB must belong to the
same TXQ. This patch takes the easy way out, by restricting bulk
dequeue to qdisc's with the TCQ_F_ONETXQUEUE flag, that specifies the
qdisc only have attached a single TXQ.
Some detail about the flow; dev_hard_start_xmit() will process the skb
list, and transmit packets individually towards the driver (see
xmit_one()). In case the driver stops midway in the list, the
remaining skb list is returned by dev_hard_start_xmit(). In
sch_direct_xmit() this returned list is requeued by dev_requeue_skb().
To avoid overshooting the HW limits, which results in requeuing, the
patch limits the amount of bytes dequeued, based on the drivers BQL
limits. In-effect bulking will only happen for BQL enabled drivers.
Small amounts for extra HoL blocking (2x MTU/0.24ms) were
measured at 100Mbit/s, with bulking 8 packets, but the
oscillating nature of the measurement indicate something, like
sched latency might be causing this effect. More comparisons
show, that this oscillation goes away occationally. Thus, we
disregard this artifact completely and remove any "magic" bulking
limit.
For now, as a conservative approach, stop bulking when seeing TSO and
segmented GSO packets. They already benefit from bulking on their own.
A followup patch add this, to allow easier bisect-ability for finding
regressions.
Jointed work with Hannes, Daniel and Florian.
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: Florian Westphal <fw@strlen.de>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-02 00:35:59 +04:00
|
|
|
{
|
2014-10-04 02:31:07 +04:00
|
|
|
int bytelimit = qdisc_avail_bulklimit(txq) - skb->len;
|
qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE
Based on DaveM's recent API work on dev_hard_start_xmit(), that allows
sending/processing an entire skb list.
This patch implements qdisc bulk dequeue, by allowing multiple packets
to be dequeued in dequeue_skb().
The optimization principle for this is two fold, (1) to amortize
locking cost and (2) avoid expensive tailptr update for notifying HW.
(1) Several packets are dequeued while holding the qdisc root_lock,
amortizing locking cost over several packet. The dequeued SKB list is
processed under the TXQ lock in dev_hard_start_xmit(), thus also
amortizing the cost of the TXQ lock.
(2) Further more, dev_hard_start_xmit() will utilize the skb->xmit_more
API to delay HW tailptr update, which also reduces the cost per
packet.
One restriction of the new API is that every SKB must belong to the
same TXQ. This patch takes the easy way out, by restricting bulk
dequeue to qdisc's with the TCQ_F_ONETXQUEUE flag, that specifies the
qdisc only have attached a single TXQ.
Some detail about the flow; dev_hard_start_xmit() will process the skb
list, and transmit packets individually towards the driver (see
xmit_one()). In case the driver stops midway in the list, the
remaining skb list is returned by dev_hard_start_xmit(). In
sch_direct_xmit() this returned list is requeued by dev_requeue_skb().
To avoid overshooting the HW limits, which results in requeuing, the
patch limits the amount of bytes dequeued, based on the drivers BQL
limits. In-effect bulking will only happen for BQL enabled drivers.
Small amounts for extra HoL blocking (2x MTU/0.24ms) were
measured at 100Mbit/s, with bulking 8 packets, but the
oscillating nature of the measurement indicate something, like
sched latency might be causing this effect. More comparisons
show, that this oscillation goes away occationally. Thus, we
disregard this artifact completely and remove any "magic" bulking
limit.
For now, as a conservative approach, stop bulking when seeing TSO and
segmented GSO packets. They already benefit from bulking on their own.
A followup patch add this, to allow easier bisect-ability for finding
regressions.
Jointed work with Hannes, Daniel and Florian.
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: Florian Westphal <fw@strlen.de>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-02 00:35:59 +04:00
|
|
|
|
|
|
|
while (bytelimit > 0) {
|
2014-10-04 02:31:07 +04:00
|
|
|
struct sk_buff *nskb = q->dequeue(q);
|
qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE
Based on DaveM's recent API work on dev_hard_start_xmit(), that allows
sending/processing an entire skb list.
This patch implements qdisc bulk dequeue, by allowing multiple packets
to be dequeued in dequeue_skb().
The optimization principle for this is two fold, (1) to amortize
locking cost and (2) avoid expensive tailptr update for notifying HW.
(1) Several packets are dequeued while holding the qdisc root_lock,
amortizing locking cost over several packet. The dequeued SKB list is
processed under the TXQ lock in dev_hard_start_xmit(), thus also
amortizing the cost of the TXQ lock.
(2) Further more, dev_hard_start_xmit() will utilize the skb->xmit_more
API to delay HW tailptr update, which also reduces the cost per
packet.
One restriction of the new API is that every SKB must belong to the
same TXQ. This patch takes the easy way out, by restricting bulk
dequeue to qdisc's with the TCQ_F_ONETXQUEUE flag, that specifies the
qdisc only have attached a single TXQ.
Some detail about the flow; dev_hard_start_xmit() will process the skb
list, and transmit packets individually towards the driver (see
xmit_one()). In case the driver stops midway in the list, the
remaining skb list is returned by dev_hard_start_xmit(). In
sch_direct_xmit() this returned list is requeued by dev_requeue_skb().
To avoid overshooting the HW limits, which results in requeuing, the
patch limits the amount of bytes dequeued, based on the drivers BQL
limits. In-effect bulking will only happen for BQL enabled drivers.
Small amounts for extra HoL blocking (2x MTU/0.24ms) were
measured at 100Mbit/s, with bulking 8 packets, but the
oscillating nature of the measurement indicate something, like
sched latency might be causing this effect. More comparisons
show, that this oscillation goes away occationally. Thus, we
disregard this artifact completely and remove any "magic" bulking
limit.
For now, as a conservative approach, stop bulking when seeing TSO and
segmented GSO packets. They already benefit from bulking on their own.
A followup patch add this, to allow easier bisect-ability for finding
regressions.
Jointed work with Hannes, Daniel and Florian.
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: Florian Westphal <fw@strlen.de>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-02 00:35:59 +04:00
|
|
|
|
2014-10-04 02:31:07 +04:00
|
|
|
if (!nskb)
|
qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE
Based on DaveM's recent API work on dev_hard_start_xmit(), that allows
sending/processing an entire skb list.
This patch implements qdisc bulk dequeue, by allowing multiple packets
to be dequeued in dequeue_skb().
The optimization principle for this is two fold, (1) to amortize
locking cost and (2) avoid expensive tailptr update for notifying HW.
(1) Several packets are dequeued while holding the qdisc root_lock,
amortizing locking cost over several packet. The dequeued SKB list is
processed under the TXQ lock in dev_hard_start_xmit(), thus also
amortizing the cost of the TXQ lock.
(2) Further more, dev_hard_start_xmit() will utilize the skb->xmit_more
API to delay HW tailptr update, which also reduces the cost per
packet.
One restriction of the new API is that every SKB must belong to the
same TXQ. This patch takes the easy way out, by restricting bulk
dequeue to qdisc's with the TCQ_F_ONETXQUEUE flag, that specifies the
qdisc only have attached a single TXQ.
Some detail about the flow; dev_hard_start_xmit() will process the skb
list, and transmit packets individually towards the driver (see
xmit_one()). In case the driver stops midway in the list, the
remaining skb list is returned by dev_hard_start_xmit(). In
sch_direct_xmit() this returned list is requeued by dev_requeue_skb().
To avoid overshooting the HW limits, which results in requeuing, the
patch limits the amount of bytes dequeued, based on the drivers BQL
limits. In-effect bulking will only happen for BQL enabled drivers.
Small amounts for extra HoL blocking (2x MTU/0.24ms) were
measured at 100Mbit/s, with bulking 8 packets, but the
oscillating nature of the measurement indicate something, like
sched latency might be causing this effect. More comparisons
show, that this oscillation goes away occationally. Thus, we
disregard this artifact completely and remove any "magic" bulking
limit.
For now, as a conservative approach, stop bulking when seeing TSO and
segmented GSO packets. They already benefit from bulking on their own.
A followup patch add this, to allow easier bisect-ability for finding
regressions.
Jointed work with Hannes, Daniel and Florian.
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: Florian Westphal <fw@strlen.de>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-02 00:35:59 +04:00
|
|
|
break;
|
|
|
|
|
2014-10-04 02:31:07 +04:00
|
|
|
bytelimit -= nskb->len; /* covers GSO len */
|
|
|
|
skb->next = nskb;
|
|
|
|
skb = nskb;
|
net_sched: restore qdisc quota fairness limits after bulk dequeue
Restore the quota fairness between qdisc's, that we broke with commit
5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE").
Before that commit, the quota in __qdisc_run() were in packets as
dequeue_skb() would only dequeue a single packet, that assumption
broke with bulk dequeue.
We choose not to account for the number of packets inside the TSO/GSO
packets (accessable via "skb_gso_segs"). As the previous fairness
also had this "defect". Thus, GSO/TSO packets counts as a single
packet.
Further more, we choose to slack on accuracy, by allowing a bulk
dequeue try_bulk_dequeue_skb() to exceed the "packets" limit, only
limited by the BQL bytelimit. This is done because BQL prefers to get
its full budget for appropriate feedback from TX completion.
In future, we might consider reworking this further and, if it allows,
switch to a time-based model, as suggested by Eric. Right now, we only
restore old semantics.
Joint work with Eric, Hannes, Daniel and Jesper. Hannes wrote the
first patch in cooperation with Daniel and Jesper. Eric rewrote the
patch.
Fixes: 5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE")
Signed-off-by: Eric Dumazet <edumazet@google.com>
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-09 14:18:10 +04:00
|
|
|
(*packets)++; /* GSO counts as one pkt */
|
qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE
Based on DaveM's recent API work on dev_hard_start_xmit(), that allows
sending/processing an entire skb list.
This patch implements qdisc bulk dequeue, by allowing multiple packets
to be dequeued in dequeue_skb().
The optimization principle for this is two fold, (1) to amortize
locking cost and (2) avoid expensive tailptr update for notifying HW.
(1) Several packets are dequeued while holding the qdisc root_lock,
amortizing locking cost over several packet. The dequeued SKB list is
processed under the TXQ lock in dev_hard_start_xmit(), thus also
amortizing the cost of the TXQ lock.
(2) Further more, dev_hard_start_xmit() will utilize the skb->xmit_more
API to delay HW tailptr update, which also reduces the cost per
packet.
One restriction of the new API is that every SKB must belong to the
same TXQ. This patch takes the easy way out, by restricting bulk
dequeue to qdisc's with the TCQ_F_ONETXQUEUE flag, that specifies the
qdisc only have attached a single TXQ.
Some detail about the flow; dev_hard_start_xmit() will process the skb
list, and transmit packets individually towards the driver (see
xmit_one()). In case the driver stops midway in the list, the
remaining skb list is returned by dev_hard_start_xmit(). In
sch_direct_xmit() this returned list is requeued by dev_requeue_skb().
To avoid overshooting the HW limits, which results in requeuing, the
patch limits the amount of bytes dequeued, based on the drivers BQL
limits. In-effect bulking will only happen for BQL enabled drivers.
Small amounts for extra HoL blocking (2x MTU/0.24ms) were
measured at 100Mbit/s, with bulking 8 packets, but the
oscillating nature of the measurement indicate something, like
sched latency might be causing this effect. More comparisons
show, that this oscillation goes away occationally. Thus, we
disregard this artifact completely and remove any "magic" bulking
limit.
For now, as a conservative approach, stop bulking when seeing TSO and
segmented GSO packets. They already benefit from bulking on their own.
A followup patch add this, to allow easier bisect-ability for finding
regressions.
Jointed work with Hannes, Daniel and Florian.
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: Florian Westphal <fw@strlen.de>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-02 00:35:59 +04:00
|
|
|
}
|
2014-10-04 02:31:07 +04:00
|
|
|
skb->next = NULL;
|
qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE
Based on DaveM's recent API work on dev_hard_start_xmit(), that allows
sending/processing an entire skb list.
This patch implements qdisc bulk dequeue, by allowing multiple packets
to be dequeued in dequeue_skb().
The optimization principle for this is two fold, (1) to amortize
locking cost and (2) avoid expensive tailptr update for notifying HW.
(1) Several packets are dequeued while holding the qdisc root_lock,
amortizing locking cost over several packet. The dequeued SKB list is
processed under the TXQ lock in dev_hard_start_xmit(), thus also
amortizing the cost of the TXQ lock.
(2) Further more, dev_hard_start_xmit() will utilize the skb->xmit_more
API to delay HW tailptr update, which also reduces the cost per
packet.
One restriction of the new API is that every SKB must belong to the
same TXQ. This patch takes the easy way out, by restricting bulk
dequeue to qdisc's with the TCQ_F_ONETXQUEUE flag, that specifies the
qdisc only have attached a single TXQ.
Some detail about the flow; dev_hard_start_xmit() will process the skb
list, and transmit packets individually towards the driver (see
xmit_one()). In case the driver stops midway in the list, the
remaining skb list is returned by dev_hard_start_xmit(). In
sch_direct_xmit() this returned list is requeued by dev_requeue_skb().
To avoid overshooting the HW limits, which results in requeuing, the
patch limits the amount of bytes dequeued, based on the drivers BQL
limits. In-effect bulking will only happen for BQL enabled drivers.
Small amounts for extra HoL blocking (2x MTU/0.24ms) were
measured at 100Mbit/s, with bulking 8 packets, but the
oscillating nature of the measurement indicate something, like
sched latency might be causing this effect. More comparisons
show, that this oscillation goes away occationally. Thus, we
disregard this artifact completely and remove any "magic" bulking
limit.
For now, as a conservative approach, stop bulking when seeing TSO and
segmented GSO packets. They already benefit from bulking on their own.
A followup patch add this, to allow easier bisect-ability for finding
regressions.
Jointed work with Hannes, Daniel and Florian.
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: Florian Westphal <fw@strlen.de>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-02 00:35:59 +04:00
|
|
|
}
|
|
|
|
|
net_sched: generalize bulk dequeue
When qdisc bulk dequeue was added in linux-3.18 (commit
5772e9a3463b "qdisc: bulk dequeue support for qdiscs
with TCQ_F_ONETXQUEUE"), it was constrained to some
specific qdiscs.
With some extra care, we can extend this to all qdiscs,
so that typical traffic shaping solutions can benefit from
small batches (8 packets in this patch).
For example, HTB is often used on some multi queue device.
And bonding/team are multi queue devices...
Idea is to bulk-dequeue packets mapping to the same transmit queue.
This brings between 35 and 80 % performance increase in HTB setup
under pressure on a bonding setup :
1) NUMA node contention : 610,000 pps -> 1,110,000 pps
2) No node contention : 1,380,000 pps -> 1,930,000 pps
Now we should work to add batches on the enqueue() side ;)
Signed-off-by: Eric Dumazet <edumazet@google.com>
Cc: John Fastabend <john.r.fastabend@intel.com>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Hannes Frederic Sowa <hannes@stressinduktion.org>
Cc: Florian Westphal <fw@strlen.de>
Cc: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-06-22 09:16:52 +03:00
|
|
|
/* This variant of try_bulk_dequeue_skb() makes sure
|
|
|
|
* all skbs in the chain are for the same txq
|
|
|
|
*/
|
|
|
|
static void try_bulk_dequeue_skb_slow(struct Qdisc *q,
|
|
|
|
struct sk_buff *skb,
|
|
|
|
int *packets)
|
|
|
|
{
|
|
|
|
int mapping = skb_get_queue_mapping(skb);
|
|
|
|
struct sk_buff *nskb;
|
|
|
|
int cnt = 0;
|
|
|
|
|
|
|
|
do {
|
|
|
|
nskb = q->dequeue(q);
|
|
|
|
if (!nskb)
|
|
|
|
break;
|
|
|
|
if (unlikely(skb_get_queue_mapping(nskb) != mapping)) {
|
|
|
|
q->skb_bad_txq = nskb;
|
|
|
|
qdisc_qstats_backlog_inc(q, nskb);
|
|
|
|
q->q.qlen++;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
skb->next = nskb;
|
|
|
|
skb = nskb;
|
|
|
|
} while (++cnt < 8);
|
|
|
|
(*packets) += cnt;
|
|
|
|
skb->next = NULL;
|
|
|
|
}
|
|
|
|
|
qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE
Based on DaveM's recent API work on dev_hard_start_xmit(), that allows
sending/processing an entire skb list.
This patch implements qdisc bulk dequeue, by allowing multiple packets
to be dequeued in dequeue_skb().
The optimization principle for this is two fold, (1) to amortize
locking cost and (2) avoid expensive tailptr update for notifying HW.
(1) Several packets are dequeued while holding the qdisc root_lock,
amortizing locking cost over several packet. The dequeued SKB list is
processed under the TXQ lock in dev_hard_start_xmit(), thus also
amortizing the cost of the TXQ lock.
(2) Further more, dev_hard_start_xmit() will utilize the skb->xmit_more
API to delay HW tailptr update, which also reduces the cost per
packet.
One restriction of the new API is that every SKB must belong to the
same TXQ. This patch takes the easy way out, by restricting bulk
dequeue to qdisc's with the TCQ_F_ONETXQUEUE flag, that specifies the
qdisc only have attached a single TXQ.
Some detail about the flow; dev_hard_start_xmit() will process the skb
list, and transmit packets individually towards the driver (see
xmit_one()). In case the driver stops midway in the list, the
remaining skb list is returned by dev_hard_start_xmit(). In
sch_direct_xmit() this returned list is requeued by dev_requeue_skb().
To avoid overshooting the HW limits, which results in requeuing, the
patch limits the amount of bytes dequeued, based on the drivers BQL
limits. In-effect bulking will only happen for BQL enabled drivers.
Small amounts for extra HoL blocking (2x MTU/0.24ms) were
measured at 100Mbit/s, with bulking 8 packets, but the
oscillating nature of the measurement indicate something, like
sched latency might be causing this effect. More comparisons
show, that this oscillation goes away occationally. Thus, we
disregard this artifact completely and remove any "magic" bulking
limit.
For now, as a conservative approach, stop bulking when seeing TSO and
segmented GSO packets. They already benefit from bulking on their own.
A followup patch add this, to allow easier bisect-ability for finding
regressions.
Jointed work with Hannes, Daniel and Florian.
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: Florian Westphal <fw@strlen.de>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-02 00:35:59 +04:00
|
|
|
/* Note that dequeue_skb can possibly return a SKB list (via skb->next).
|
|
|
|
* A requeued skb (via q->gso_skb) can also be a SKB list.
|
|
|
|
*/
|
net_sched: restore qdisc quota fairness limits after bulk dequeue
Restore the quota fairness between qdisc's, that we broke with commit
5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE").
Before that commit, the quota in __qdisc_run() were in packets as
dequeue_skb() would only dequeue a single packet, that assumption
broke with bulk dequeue.
We choose not to account for the number of packets inside the TSO/GSO
packets (accessable via "skb_gso_segs"). As the previous fairness
also had this "defect". Thus, GSO/TSO packets counts as a single
packet.
Further more, we choose to slack on accuracy, by allowing a bulk
dequeue try_bulk_dequeue_skb() to exceed the "packets" limit, only
limited by the BQL bytelimit. This is done because BQL prefers to get
its full budget for appropriate feedback from TX completion.
In future, we might consider reworking this further and, if it allows,
switch to a time-based model, as suggested by Eric. Right now, we only
restore old semantics.
Joint work with Eric, Hannes, Daniel and Jesper. Hannes wrote the
first patch in cooperation with Daniel and Jesper. Eric rewrote the
patch.
Fixes: 5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE")
Signed-off-by: Eric Dumazet <edumazet@google.com>
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-09 14:18:10 +04:00
|
|
|
static struct sk_buff *dequeue_skb(struct Qdisc *q, bool *validate,
|
|
|
|
int *packets)
|
2007-06-11 04:31:24 +04:00
|
|
|
{
|
2008-10-06 20:54:39 +04:00
|
|
|
struct sk_buff *skb = q->gso_skb;
|
2012-12-11 19:54:33 +04:00
|
|
|
const struct netdev_queue *txq = q->dev_queue;
|
2008-10-06 20:54:39 +04:00
|
|
|
|
net_sched: restore qdisc quota fairness limits after bulk dequeue
Restore the quota fairness between qdisc's, that we broke with commit
5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE").
Before that commit, the quota in __qdisc_run() were in packets as
dequeue_skb() would only dequeue a single packet, that assumption
broke with bulk dequeue.
We choose not to account for the number of packets inside the TSO/GSO
packets (accessable via "skb_gso_segs"). As the previous fairness
also had this "defect". Thus, GSO/TSO packets counts as a single
packet.
Further more, we choose to slack on accuracy, by allowing a bulk
dequeue try_bulk_dequeue_skb() to exceed the "packets" limit, only
limited by the BQL bytelimit. This is done because BQL prefers to get
its full budget for appropriate feedback from TX completion.
In future, we might consider reworking this further and, if it allows,
switch to a time-based model, as suggested by Eric. Right now, we only
restore old semantics.
Joint work with Eric, Hannes, Daniel and Jesper. Hannes wrote the
first patch in cooperation with Daniel and Jesper. Eric rewrote the
patch.
Fixes: 5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE")
Signed-off-by: Eric Dumazet <edumazet@google.com>
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-09 14:18:10 +04:00
|
|
|
*packets = 1;
|
2008-09-23 09:16:23 +04:00
|
|
|
if (unlikely(skb)) {
|
net_sched: generalize bulk dequeue
When qdisc bulk dequeue was added in linux-3.18 (commit
5772e9a3463b "qdisc: bulk dequeue support for qdiscs
with TCQ_F_ONETXQUEUE"), it was constrained to some
specific qdiscs.
With some extra care, we can extend this to all qdiscs,
so that typical traffic shaping solutions can benefit from
small batches (8 packets in this patch).
For example, HTB is often used on some multi queue device.
And bonding/team are multi queue devices...
Idea is to bulk-dequeue packets mapping to the same transmit queue.
This brings between 35 and 80 % performance increase in HTB setup
under pressure on a bonding setup :
1) NUMA node contention : 610,000 pps -> 1,110,000 pps
2) No node contention : 1,380,000 pps -> 1,930,000 pps
Now we should work to add batches on the enqueue() side ;)
Signed-off-by: Eric Dumazet <edumazet@google.com>
Cc: John Fastabend <john.r.fastabend@intel.com>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Hannes Frederic Sowa <hannes@stressinduktion.org>
Cc: Florian Westphal <fw@strlen.de>
Cc: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-06-22 09:16:52 +03:00
|
|
|
/* skb in gso_skb were already validated */
|
|
|
|
*validate = false;
|
2008-09-23 09:16:23 +04:00
|
|
|
/* check the reason of requeuing without tx lock first */
|
2014-08-27 13:11:27 +04:00
|
|
|
txq = skb_get_tx_queue(txq->dev, skb);
|
2011-11-28 20:32:44 +04:00
|
|
|
if (!netif_xmit_frozen_or_stopped(txq)) {
|
2008-10-06 21:41:50 +04:00
|
|
|
q->gso_skb = NULL;
|
2016-06-04 01:05:57 +03:00
|
|
|
qdisc_qstats_backlog_dec(q, skb);
|
2009-08-06 05:44:21 +04:00
|
|
|
q->q.qlen--;
|
|
|
|
} else
|
2008-09-23 09:16:23 +04:00
|
|
|
skb = NULL;
|
2017-08-15 22:11:03 +03:00
|
|
|
goto trace;
|
net_sched: generalize bulk dequeue
When qdisc bulk dequeue was added in linux-3.18 (commit
5772e9a3463b "qdisc: bulk dequeue support for qdiscs
with TCQ_F_ONETXQUEUE"), it was constrained to some
specific qdiscs.
With some extra care, we can extend this to all qdiscs,
so that typical traffic shaping solutions can benefit from
small batches (8 packets in this patch).
For example, HTB is often used on some multi queue device.
And bonding/team are multi queue devices...
Idea is to bulk-dequeue packets mapping to the same transmit queue.
This brings between 35 and 80 % performance increase in HTB setup
under pressure on a bonding setup :
1) NUMA node contention : 610,000 pps -> 1,110,000 pps
2) No node contention : 1,380,000 pps -> 1,930,000 pps
Now we should work to add batches on the enqueue() side ;)
Signed-off-by: Eric Dumazet <edumazet@google.com>
Cc: John Fastabend <john.r.fastabend@intel.com>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Hannes Frederic Sowa <hannes@stressinduktion.org>
Cc: Florian Westphal <fw@strlen.de>
Cc: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-06-22 09:16:52 +03:00
|
|
|
}
|
|
|
|
*validate = true;
|
|
|
|
skb = q->skb_bad_txq;
|
|
|
|
if (unlikely(skb)) {
|
|
|
|
/* check the reason of requeuing without tx lock first */
|
|
|
|
txq = skb_get_tx_queue(txq->dev, skb);
|
|
|
|
if (!netif_xmit_frozen_or_stopped(txq)) {
|
|
|
|
q->skb_bad_txq = NULL;
|
|
|
|
qdisc_qstats_backlog_dec(q, skb);
|
|
|
|
q->q.qlen--;
|
|
|
|
goto bulk;
|
2014-08-31 06:13:51 +04:00
|
|
|
}
|
2017-08-15 22:11:03 +03:00
|
|
|
skb = NULL;
|
|
|
|
goto trace;
|
net_sched: generalize bulk dequeue
When qdisc bulk dequeue was added in linux-3.18 (commit
5772e9a3463b "qdisc: bulk dequeue support for qdiscs
with TCQ_F_ONETXQUEUE"), it was constrained to some
specific qdiscs.
With some extra care, we can extend this to all qdiscs,
so that typical traffic shaping solutions can benefit from
small batches (8 packets in this patch).
For example, HTB is often used on some multi queue device.
And bonding/team are multi queue devices...
Idea is to bulk-dequeue packets mapping to the same transmit queue.
This brings between 35 and 80 % performance increase in HTB setup
under pressure on a bonding setup :
1) NUMA node contention : 610,000 pps -> 1,110,000 pps
2) No node contention : 1,380,000 pps -> 1,930,000 pps
Now we should work to add batches on the enqueue() side ;)
Signed-off-by: Eric Dumazet <edumazet@google.com>
Cc: John Fastabend <john.r.fastabend@intel.com>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Hannes Frederic Sowa <hannes@stressinduktion.org>
Cc: Florian Westphal <fw@strlen.de>
Cc: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-06-22 09:16:52 +03:00
|
|
|
}
|
|
|
|
if (!(q->flags & TCQ_F_ONETXQUEUE) ||
|
|
|
|
!netif_xmit_frozen_or_stopped(txq))
|
|
|
|
skb = q->dequeue(q);
|
|
|
|
if (skb) {
|
|
|
|
bulk:
|
|
|
|
if (qdisc_may_bulk(q))
|
|
|
|
try_bulk_dequeue_skb(q, skb, txq, packets);
|
|
|
|
else
|
|
|
|
try_bulk_dequeue_skb_slow(q, skb, packets);
|
2008-09-23 09:16:23 +04:00
|
|
|
}
|
2017-08-15 22:11:03 +03:00
|
|
|
trace:
|
|
|
|
trace_qdisc_dequeue(q, txq, *packets, skb);
|
2007-06-11 04:31:24 +04:00
|
|
|
return skb;
|
|
|
|
}
|
|
|
|
|
2007-02-09 17:25:16 +03:00
|
|
|
/*
|
2014-09-02 18:35:33 +04:00
|
|
|
* Transmit possibly several skbs, and handle the return status as
|
2016-06-06 19:37:15 +03:00
|
|
|
* required. Owning running seqcount bit guarantees that
|
2014-09-02 18:35:33 +04:00
|
|
|
* only one CPU can execute this function.
|
[NET]: qdisc_restart - readability changes plus one bug fix.
New changes :
- Incorporated Peter Waskiewicz's comments.
- Re-added back one warning message (on driver returning wrong value).
Previous changes :
- Converted to use switch/case code which looks neater.
- "if (ret == NETDEV_TX_LOCKED && lockless)" is buggy, and the lockless
check should be removed, since driver will return NETDEV_TX_LOCKED only
if lockless is true and driver has to do the locking. In the original
code as well as the latest code, this code can result in a bug where
if LLTX is not set for a driver (lockless == 0) but the driver is written
wrongly to do a trylock (despite LLTX being set), the driver returns
LOCKED. But since lockless is zero, the packet is requeue'd instead of
calling collision code which will issue warning and free up the skb.
Instead this skb will be retried with this driver next time, and the same
result will ensue. Removing this check will catch these driver bugs instead
of hiding the problem. I am keeping this change to readability section
since :
a. it is confusing to check two things as it is; and
b. it is difficult to keep this check in the changed 'switch' code.
- Changed some names, like try_get_tx_pkt to dev_dequeue_skb (as that is
the work being done and easier to understand) and do_dev_requeue to
dev_requeue_skb, merged handle_dev_cpu_collision and tx_islocked to
dev_handle_collision (handle_dev_cpu_collision is a small routine with only
one caller, so there is no need to have two separate routines which also
results in getting rid of two macros, etc.
- Removed an XXX comment as it should never fail (I suspect this was related
to batch skb WIP, Jamal ?). Converted some functions to original coding
style of having the return values and the function name on same line, eg
prio2list.
Signed-off-by: Krishna Kumar <krkumar2@in.ibm.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2007-06-25 06:56:09 +04:00
|
|
|
*
|
|
|
|
* Returns to the caller:
|
|
|
|
* 0 - queue is empty or throttled.
|
|
|
|
* >0 - queue is not empty.
|
|
|
|
*/
|
2009-08-06 05:44:21 +04:00
|
|
|
int sch_direct_xmit(struct sk_buff *skb, struct Qdisc *q,
|
|
|
|
struct net_device *dev, struct netdev_queue *txq,
|
2014-10-04 02:31:07 +04:00
|
|
|
spinlock_t *root_lock, bool validate)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
2007-11-14 07:40:55 +03:00
|
|
|
int ret = NETDEV_TX_BUSY;
|
2008-07-16 12:42:40 +04:00
|
|
|
|
|
|
|
/* And release qdisc */
|
|
|
|
spin_unlock(root_lock);
|
2007-06-11 04:31:24 +04:00
|
|
|
|
2014-10-04 02:31:07 +04:00
|
|
|
/* Note that we validate skb (GSO, checksum, ...) outside of locks */
|
|
|
|
if (validate)
|
|
|
|
skb = validate_xmit_skb_list(skb, dev);
|
2009-11-10 09:14:14 +03:00
|
|
|
|
2016-04-12 09:45:52 +03:00
|
|
|
if (likely(skb)) {
|
2014-10-04 02:31:07 +04:00
|
|
|
HARD_TX_LOCK(dev, txq, smp_processor_id());
|
|
|
|
if (!netif_xmit_frozen_or_stopped(txq))
|
|
|
|
skb = dev_hard_start_xmit(skb, dev, txq, &ret);
|
2007-06-11 04:31:24 +04:00
|
|
|
|
2014-10-04 02:31:07 +04:00
|
|
|
HARD_TX_UNLOCK(dev, txq);
|
2016-04-12 09:45:52 +03:00
|
|
|
} else {
|
2016-06-09 17:45:11 +03:00
|
|
|
spin_lock(root_lock);
|
2016-04-12 09:45:52 +03:00
|
|
|
return qdisc_qlen(q);
|
2014-10-04 02:31:07 +04:00
|
|
|
}
|
2016-06-09 17:45:11 +03:00
|
|
|
spin_lock(root_lock);
|
2007-06-11 04:31:24 +04:00
|
|
|
|
2009-11-15 10:20:12 +03:00
|
|
|
if (dev_xmit_complete(ret)) {
|
|
|
|
/* Driver sent out skb successfully or skb was consumed */
|
[NET]: qdisc_restart - readability changes plus one bug fix.
New changes :
- Incorporated Peter Waskiewicz's comments.
- Re-added back one warning message (on driver returning wrong value).
Previous changes :
- Converted to use switch/case code which looks neater.
- "if (ret == NETDEV_TX_LOCKED && lockless)" is buggy, and the lockless
check should be removed, since driver will return NETDEV_TX_LOCKED only
if lockless is true and driver has to do the locking. In the original
code as well as the latest code, this code can result in a bug where
if LLTX is not set for a driver (lockless == 0) but the driver is written
wrongly to do a trylock (despite LLTX being set), the driver returns
LOCKED. But since lockless is zero, the packet is requeue'd instead of
calling collision code which will issue warning and free up the skb.
Instead this skb will be retried with this driver next time, and the same
result will ensue. Removing this check will catch these driver bugs instead
of hiding the problem. I am keeping this change to readability section
since :
a. it is confusing to check two things as it is; and
b. it is difficult to keep this check in the changed 'switch' code.
- Changed some names, like try_get_tx_pkt to dev_dequeue_skb (as that is
the work being done and easier to understand) and do_dev_requeue to
dev_requeue_skb, merged handle_dev_cpu_collision and tx_islocked to
dev_handle_collision (handle_dev_cpu_collision is a small routine with only
one caller, so there is no need to have two separate routines which also
results in getting rid of two macros, etc.
- Removed an XXX comment as it should never fail (I suspect this was related
to batch skb WIP, Jamal ?). Converted some functions to original coding
style of having the return values and the function name on same line, eg
prio2list.
Signed-off-by: Krishna Kumar <krkumar2@in.ibm.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2007-06-25 06:56:09 +04:00
|
|
|
ret = qdisc_qlen(q);
|
2009-11-15 10:20:12 +03:00
|
|
|
} else {
|
[NET]: qdisc_restart - readability changes plus one bug fix.
New changes :
- Incorporated Peter Waskiewicz's comments.
- Re-added back one warning message (on driver returning wrong value).
Previous changes :
- Converted to use switch/case code which looks neater.
- "if (ret == NETDEV_TX_LOCKED && lockless)" is buggy, and the lockless
check should be removed, since driver will return NETDEV_TX_LOCKED only
if lockless is true and driver has to do the locking. In the original
code as well as the latest code, this code can result in a bug where
if LLTX is not set for a driver (lockless == 0) but the driver is written
wrongly to do a trylock (despite LLTX being set), the driver returns
LOCKED. But since lockless is zero, the packet is requeue'd instead of
calling collision code which will issue warning and free up the skb.
Instead this skb will be retried with this driver next time, and the same
result will ensue. Removing this check will catch these driver bugs instead
of hiding the problem. I am keeping this change to readability section
since :
a. it is confusing to check two things as it is; and
b. it is difficult to keep this check in the changed 'switch' code.
- Changed some names, like try_get_tx_pkt to dev_dequeue_skb (as that is
the work being done and easier to understand) and do_dev_requeue to
dev_requeue_skb, merged handle_dev_cpu_collision and tx_islocked to
dev_handle_collision (handle_dev_cpu_collision is a small routine with only
one caller, so there is no need to have two separate routines which also
results in getting rid of two macros, etc.
- Removed an XXX comment as it should never fail (I suspect this was related
to batch skb WIP, Jamal ?). Converted some functions to original coding
style of having the return values and the function name on same line, eg
prio2list.
Signed-off-by: Krishna Kumar <krkumar2@in.ibm.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2007-06-25 06:56:09 +04:00
|
|
|
/* Driver returned NETDEV_TX_BUSY - requeue skb */
|
2012-05-14 01:56:26 +04:00
|
|
|
if (unlikely(ret != NETDEV_TX_BUSY))
|
|
|
|
net_warn_ratelimited("BUG %s code %d qlen %d\n",
|
|
|
|
dev->name, ret, q->q.qlen);
|
[NET]: qdisc_restart - readability changes plus one bug fix.
New changes :
- Incorporated Peter Waskiewicz's comments.
- Re-added back one warning message (on driver returning wrong value).
Previous changes :
- Converted to use switch/case code which looks neater.
- "if (ret == NETDEV_TX_LOCKED && lockless)" is buggy, and the lockless
check should be removed, since driver will return NETDEV_TX_LOCKED only
if lockless is true and driver has to do the locking. In the original
code as well as the latest code, this code can result in a bug where
if LLTX is not set for a driver (lockless == 0) but the driver is written
wrongly to do a trylock (despite LLTX being set), the driver returns
LOCKED. But since lockless is zero, the packet is requeue'd instead of
calling collision code which will issue warning and free up the skb.
Instead this skb will be retried with this driver next time, and the same
result will ensue. Removing this check will catch these driver bugs instead
of hiding the problem. I am keeping this change to readability section
since :
a. it is confusing to check two things as it is; and
b. it is difficult to keep this check in the changed 'switch' code.
- Changed some names, like try_get_tx_pkt to dev_dequeue_skb (as that is
the work being done and easier to understand) and do_dev_requeue to
dev_requeue_skb, merged handle_dev_cpu_collision and tx_islocked to
dev_handle_collision (handle_dev_cpu_collision is a small routine with only
one caller, so there is no need to have two separate routines which also
results in getting rid of two macros, etc.
- Removed an XXX comment as it should never fail (I suspect this was related
to batch skb WIP, Jamal ?). Converted some functions to original coding
style of having the return values and the function name on same line, eg
prio2list.
Signed-off-by: Krishna Kumar <krkumar2@in.ibm.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2007-06-25 06:56:09 +04:00
|
|
|
|
2008-07-16 13:15:04 +04:00
|
|
|
ret = dev_requeue_skb(skb, q);
|
[NET]: qdisc_restart - readability changes plus one bug fix.
New changes :
- Incorporated Peter Waskiewicz's comments.
- Re-added back one warning message (on driver returning wrong value).
Previous changes :
- Converted to use switch/case code which looks neater.
- "if (ret == NETDEV_TX_LOCKED && lockless)" is buggy, and the lockless
check should be removed, since driver will return NETDEV_TX_LOCKED only
if lockless is true and driver has to do the locking. In the original
code as well as the latest code, this code can result in a bug where
if LLTX is not set for a driver (lockless == 0) but the driver is written
wrongly to do a trylock (despite LLTX being set), the driver returns
LOCKED. But since lockless is zero, the packet is requeue'd instead of
calling collision code which will issue warning and free up the skb.
Instead this skb will be retried with this driver next time, and the same
result will ensue. Removing this check will catch these driver bugs instead
of hiding the problem. I am keeping this change to readability section
since :
a. it is confusing to check two things as it is; and
b. it is difficult to keep this check in the changed 'switch' code.
- Changed some names, like try_get_tx_pkt to dev_dequeue_skb (as that is
the work being done and easier to understand) and do_dev_requeue to
dev_requeue_skb, merged handle_dev_cpu_collision and tx_islocked to
dev_handle_collision (handle_dev_cpu_collision is a small routine with only
one caller, so there is no need to have two separate routines which also
results in getting rid of two macros, etc.
- Removed an XXX comment as it should never fail (I suspect this was related
to batch skb WIP, Jamal ?). Converted some functions to original coding
style of having the return values and the function name on same line, eg
prio2list.
Signed-off-by: Krishna Kumar <krkumar2@in.ibm.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2007-06-25 06:56:09 +04:00
|
|
|
}
|
2007-06-11 04:31:24 +04:00
|
|
|
|
2011-11-28 20:32:44 +04:00
|
|
|
if (ret && netif_xmit_frozen_or_stopped(txq))
|
2008-07-16 13:15:04 +04:00
|
|
|
ret = 0;
|
|
|
|
|
[NET]: qdisc_restart - readability changes plus one bug fix.
New changes :
- Incorporated Peter Waskiewicz's comments.
- Re-added back one warning message (on driver returning wrong value).
Previous changes :
- Converted to use switch/case code which looks neater.
- "if (ret == NETDEV_TX_LOCKED && lockless)" is buggy, and the lockless
check should be removed, since driver will return NETDEV_TX_LOCKED only
if lockless is true and driver has to do the locking. In the original
code as well as the latest code, this code can result in a bug where
if LLTX is not set for a driver (lockless == 0) but the driver is written
wrongly to do a trylock (despite LLTX being set), the driver returns
LOCKED. But since lockless is zero, the packet is requeue'd instead of
calling collision code which will issue warning and free up the skb.
Instead this skb will be retried with this driver next time, and the same
result will ensue. Removing this check will catch these driver bugs instead
of hiding the problem. I am keeping this change to readability section
since :
a. it is confusing to check two things as it is; and
b. it is difficult to keep this check in the changed 'switch' code.
- Changed some names, like try_get_tx_pkt to dev_dequeue_skb (as that is
the work being done and easier to understand) and do_dev_requeue to
dev_requeue_skb, merged handle_dev_cpu_collision and tx_islocked to
dev_handle_collision (handle_dev_cpu_collision is a small routine with only
one caller, so there is no need to have two separate routines which also
results in getting rid of two macros, etc.
- Removed an XXX comment as it should never fail (I suspect this was related
to batch skb WIP, Jamal ?). Converted some functions to original coding
style of having the return values and the function name on same line, eg
prio2list.
Signed-off-by: Krishna Kumar <krkumar2@in.ibm.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2007-06-25 06:56:09 +04:00
|
|
|
return ret;
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2009-08-06 05:44:21 +04:00
|
|
|
/*
|
|
|
|
* NOTE: Called under qdisc_lock(q) with locally disabled BH.
|
|
|
|
*
|
2016-06-06 19:37:15 +03:00
|
|
|
* running seqcount guarantees only one CPU can process
|
2009-08-06 05:44:21 +04:00
|
|
|
* this qdisc at a time. qdisc_lock(q) serializes queue accesses for
|
|
|
|
* this queue.
|
|
|
|
*
|
|
|
|
* netif_tx_lock serializes accesses to device driver.
|
|
|
|
*
|
|
|
|
* qdisc_lock(q) and netif_tx_lock are mutually exclusive,
|
|
|
|
* if one is grabbed, another must be free.
|
|
|
|
*
|
|
|
|
* Note, that this procedure can be called by a watchdog timer
|
|
|
|
*
|
|
|
|
* Returns to the caller:
|
|
|
|
* 0 - queue is empty or throttled.
|
|
|
|
* >0 - queue is not empty.
|
|
|
|
*
|
|
|
|
*/
|
net_sched: restore qdisc quota fairness limits after bulk dequeue
Restore the quota fairness between qdisc's, that we broke with commit
5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE").
Before that commit, the quota in __qdisc_run() were in packets as
dequeue_skb() would only dequeue a single packet, that assumption
broke with bulk dequeue.
We choose not to account for the number of packets inside the TSO/GSO
packets (accessable via "skb_gso_segs"). As the previous fairness
also had this "defect". Thus, GSO/TSO packets counts as a single
packet.
Further more, we choose to slack on accuracy, by allowing a bulk
dequeue try_bulk_dequeue_skb() to exceed the "packets" limit, only
limited by the BQL bytelimit. This is done because BQL prefers to get
its full budget for appropriate feedback from TX completion.
In future, we might consider reworking this further and, if it allows,
switch to a time-based model, as suggested by Eric. Right now, we only
restore old semantics.
Joint work with Eric, Hannes, Daniel and Jesper. Hannes wrote the
first patch in cooperation with Daniel and Jesper. Eric rewrote the
patch.
Fixes: 5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE")
Signed-off-by: Eric Dumazet <edumazet@google.com>
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-09 14:18:10 +04:00
|
|
|
static inline int qdisc_restart(struct Qdisc *q, int *packets)
|
2009-08-06 05:44:21 +04:00
|
|
|
{
|
|
|
|
struct netdev_queue *txq;
|
|
|
|
struct net_device *dev;
|
|
|
|
spinlock_t *root_lock;
|
|
|
|
struct sk_buff *skb;
|
2014-10-04 02:31:07 +04:00
|
|
|
bool validate;
|
2009-08-06 05:44:21 +04:00
|
|
|
|
|
|
|
/* Dequeue packet */
|
net_sched: restore qdisc quota fairness limits after bulk dequeue
Restore the quota fairness between qdisc's, that we broke with commit
5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE").
Before that commit, the quota in __qdisc_run() were in packets as
dequeue_skb() would only dequeue a single packet, that assumption
broke with bulk dequeue.
We choose not to account for the number of packets inside the TSO/GSO
packets (accessable via "skb_gso_segs"). As the previous fairness
also had this "defect". Thus, GSO/TSO packets counts as a single
packet.
Further more, we choose to slack on accuracy, by allowing a bulk
dequeue try_bulk_dequeue_skb() to exceed the "packets" limit, only
limited by the BQL bytelimit. This is done because BQL prefers to get
its full budget for appropriate feedback from TX completion.
In future, we might consider reworking this further and, if it allows,
switch to a time-based model, as suggested by Eric. Right now, we only
restore old semantics.
Joint work with Eric, Hannes, Daniel and Jesper. Hannes wrote the
first patch in cooperation with Daniel and Jesper. Eric rewrote the
patch.
Fixes: 5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE")
Signed-off-by: Eric Dumazet <edumazet@google.com>
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-09 14:18:10 +04:00
|
|
|
skb = dequeue_skb(q, &validate, packets);
|
2009-08-06 05:44:21 +04:00
|
|
|
if (unlikely(!skb))
|
|
|
|
return 0;
|
2014-08-27 13:11:27 +04:00
|
|
|
|
2009-08-06 05:44:21 +04:00
|
|
|
root_lock = qdisc_lock(q);
|
|
|
|
dev = qdisc_dev(q);
|
2014-08-27 13:11:27 +04:00
|
|
|
txq = skb_get_tx_queue(dev, skb);
|
2009-08-06 05:44:21 +04:00
|
|
|
|
2014-10-04 02:31:07 +04:00
|
|
|
return sch_direct_xmit(skb, q, dev, txq, root_lock, validate);
|
2009-08-06 05:44:21 +04:00
|
|
|
}
|
|
|
|
|
2008-07-16 13:15:04 +04:00
|
|
|
void __qdisc_run(struct Qdisc *q)
|
2006-06-20 10:57:59 +04:00
|
|
|
{
|
2016-12-29 23:37:21 +03:00
|
|
|
int quota = dev_tx_weight;
|
net_sched: restore qdisc quota fairness limits after bulk dequeue
Restore the quota fairness between qdisc's, that we broke with commit
5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE").
Before that commit, the quota in __qdisc_run() were in packets as
dequeue_skb() would only dequeue a single packet, that assumption
broke with bulk dequeue.
We choose not to account for the number of packets inside the TSO/GSO
packets (accessable via "skb_gso_segs"). As the previous fairness
also had this "defect". Thus, GSO/TSO packets counts as a single
packet.
Further more, we choose to slack on accuracy, by allowing a bulk
dequeue try_bulk_dequeue_skb() to exceed the "packets" limit, only
limited by the BQL bytelimit. This is done because BQL prefers to get
its full budget for appropriate feedback from TX completion.
In future, we might consider reworking this further and, if it allows,
switch to a time-based model, as suggested by Eric. Right now, we only
restore old semantics.
Joint work with Eric, Hannes, Daniel and Jesper. Hannes wrote the
first patch in cooperation with Daniel and Jesper. Eric rewrote the
patch.
Fixes: 5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE")
Signed-off-by: Eric Dumazet <edumazet@google.com>
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-09 14:18:10 +04:00
|
|
|
int packets;
|
2008-03-29 02:25:26 +03:00
|
|
|
|
net_sched: restore qdisc quota fairness limits after bulk dequeue
Restore the quota fairness between qdisc's, that we broke with commit
5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE").
Before that commit, the quota in __qdisc_run() were in packets as
dequeue_skb() would only dequeue a single packet, that assumption
broke with bulk dequeue.
We choose not to account for the number of packets inside the TSO/GSO
packets (accessable via "skb_gso_segs"). As the previous fairness
also had this "defect". Thus, GSO/TSO packets counts as a single
packet.
Further more, we choose to slack on accuracy, by allowing a bulk
dequeue try_bulk_dequeue_skb() to exceed the "packets" limit, only
limited by the BQL bytelimit. This is done because BQL prefers to get
its full budget for appropriate feedback from TX completion.
In future, we might consider reworking this further and, if it allows,
switch to a time-based model, as suggested by Eric. Right now, we only
restore old semantics.
Joint work with Eric, Hannes, Daniel and Jesper. Hannes wrote the
first patch in cooperation with Daniel and Jesper. Eric rewrote the
patch.
Fixes: 5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE")
Signed-off-by: Eric Dumazet <edumazet@google.com>
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-09 14:18:10 +04:00
|
|
|
while (qdisc_restart(q, &packets)) {
|
2008-03-29 02:25:26 +03:00
|
|
|
/*
|
2011-06-26 12:13:54 +04:00
|
|
|
* Ordered by possible occurrence: Postpone processing if
|
|
|
|
* 1. we've exceeded packet quota
|
|
|
|
* 2. another process needs the CPU;
|
2008-03-29 02:25:26 +03:00
|
|
|
*/
|
net_sched: restore qdisc quota fairness limits after bulk dequeue
Restore the quota fairness between qdisc's, that we broke with commit
5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE").
Before that commit, the quota in __qdisc_run() were in packets as
dequeue_skb() would only dequeue a single packet, that assumption
broke with bulk dequeue.
We choose not to account for the number of packets inside the TSO/GSO
packets (accessable via "skb_gso_segs"). As the previous fairness
also had this "defect". Thus, GSO/TSO packets counts as a single
packet.
Further more, we choose to slack on accuracy, by allowing a bulk
dequeue try_bulk_dequeue_skb() to exceed the "packets" limit, only
limited by the BQL bytelimit. This is done because BQL prefers to get
its full budget for appropriate feedback from TX completion.
In future, we might consider reworking this further and, if it allows,
switch to a time-based model, as suggested by Eric. Right now, we only
restore old semantics.
Joint work with Eric, Hannes, Daniel and Jesper. Hannes wrote the
first patch in cooperation with Daniel and Jesper. Eric rewrote the
patch.
Fixes: 5772e9a346 ("qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE")
Signed-off-by: Eric Dumazet <edumazet@google.com>
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Hannes Frederic Sowa <hannes@stressinduktion.org>
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-10-09 14:18:10 +04:00
|
|
|
quota -= packets;
|
|
|
|
if (quota <= 0 || need_resched()) {
|
2008-07-16 13:15:04 +04:00
|
|
|
__netif_schedule(q);
|
2007-05-10 15:55:14 +04:00
|
|
|
break;
|
2008-03-29 02:25:26 +03:00
|
|
|
}
|
|
|
|
}
|
2006-06-20 10:57:59 +04:00
|
|
|
|
2010-06-02 14:23:51 +04:00
|
|
|
qdisc_run_end(q);
|
2006-06-20 10:57:59 +04:00
|
|
|
}
|
|
|
|
|
2009-05-18 07:55:16 +04:00
|
|
|
unsigned long dev_trans_start(struct net_device *dev)
|
|
|
|
{
|
2013-08-04 00:07:47 +04:00
|
|
|
unsigned long val, res;
|
2009-05-18 07:55:16 +04:00
|
|
|
unsigned int i;
|
|
|
|
|
2013-08-04 00:07:47 +04:00
|
|
|
if (is_vlan_dev(dev))
|
|
|
|
dev = vlan_dev_real_dev(dev);
|
2017-12-06 18:50:28 +03:00
|
|
|
else if (netif_is_macvlan(dev))
|
|
|
|
dev = macvlan_dev_real_dev(dev);
|
2016-05-03 17:33:14 +03:00
|
|
|
res = netdev_get_tx_queue(dev, 0)->trans_start;
|
|
|
|
for (i = 1; i < dev->num_tx_queues; i++) {
|
2009-05-18 07:55:16 +04:00
|
|
|
val = netdev_get_tx_queue(dev, i)->trans_start;
|
|
|
|
if (val && time_after(val, res))
|
|
|
|
res = val;
|
|
|
|
}
|
2013-08-04 00:07:47 +04:00
|
|
|
|
2009-05-18 07:55:16 +04:00
|
|
|
return res;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(dev_trans_start);
|
|
|
|
|
2017-10-17 03:29:17 +03:00
|
|
|
static void dev_watchdog(struct timer_list *t)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
2017-10-17 03:29:17 +03:00
|
|
|
struct net_device *dev = from_timer(dev, t, watchdog_timer);
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2006-06-09 23:20:56 +04:00
|
|
|
netif_tx_lock(dev);
|
2008-07-17 11:34:19 +04:00
|
|
|
if (!qdisc_tx_is_noop(dev)) {
|
2005-04-17 02:20:36 +04:00
|
|
|
if (netif_device_present(dev) &&
|
|
|
|
netif_running(dev) &&
|
|
|
|
netif_carrier_ok(dev)) {
|
2009-05-18 07:55:16 +04:00
|
|
|
int some_queue_timedout = 0;
|
2008-07-17 11:34:19 +04:00
|
|
|
unsigned int i;
|
2009-05-18 07:55:16 +04:00
|
|
|
unsigned long trans_start;
|
2008-07-17 11:34:19 +04:00
|
|
|
|
|
|
|
for (i = 0; i < dev->num_tx_queues; i++) {
|
|
|
|
struct netdev_queue *txq;
|
|
|
|
|
|
|
|
txq = netdev_get_tx_queue(dev, i);
|
2016-05-03 17:33:14 +03:00
|
|
|
trans_start = txq->trans_start;
|
2011-11-28 20:32:44 +04:00
|
|
|
if (netif_xmit_stopped(txq) &&
|
2009-05-18 07:55:16 +04:00
|
|
|
time_after(jiffies, (trans_start +
|
|
|
|
dev->watchdog_timeo))) {
|
|
|
|
some_queue_timedout = 1;
|
2011-11-16 16:15:10 +04:00
|
|
|
txq->trans_timeout++;
|
2008-07-17 11:34:19 +04:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
2006-05-17 02:02:12 +04:00
|
|
|
|
2009-05-18 07:55:16 +04:00
|
|
|
if (some_queue_timedout) {
|
|
|
|
WARN_ONCE(1, KERN_INFO "NETDEV WATCHDOG: %s (%s): transmit queue %u timed out\n",
|
2011-06-07 03:41:33 +04:00
|
|
|
dev->name, netdev_drivername(dev), i);
|
2008-11-20 08:32:24 +03:00
|
|
|
dev->netdev_ops->ndo_tx_timeout(dev);
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
2008-07-17 11:34:19 +04:00
|
|
|
if (!mod_timer(&dev->watchdog_timer,
|
|
|
|
round_jiffies(jiffies +
|
|
|
|
dev->watchdog_timeo)))
|
2005-04-17 02:20:36 +04:00
|
|
|
dev_hold(dev);
|
|
|
|
}
|
|
|
|
}
|
2006-06-09 23:20:56 +04:00
|
|
|
netif_tx_unlock(dev);
|
2005-04-17 02:20:36 +04:00
|
|
|
|
|
|
|
dev_put(dev);
|
|
|
|
}
|
|
|
|
|
|
|
|
void __netdev_watchdog_up(struct net_device *dev)
|
|
|
|
{
|
2008-11-20 08:32:24 +03:00
|
|
|
if (dev->netdev_ops->ndo_tx_timeout) {
|
2005-04-17 02:20:36 +04:00
|
|
|
if (dev->watchdog_timeo <= 0)
|
|
|
|
dev->watchdog_timeo = 5*HZ;
|
2007-06-01 08:28:44 +04:00
|
|
|
if (!mod_timer(&dev->watchdog_timer,
|
|
|
|
round_jiffies(jiffies + dev->watchdog_timeo)))
|
2005-04-17 02:20:36 +04:00
|
|
|
dev_hold(dev);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void dev_watchdog_up(struct net_device *dev)
|
|
|
|
{
|
|
|
|
__netdev_watchdog_up(dev);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void dev_watchdog_down(struct net_device *dev)
|
|
|
|
{
|
2006-06-09 23:20:56 +04:00
|
|
|
netif_tx_lock_bh(dev);
|
2005-04-17 02:20:36 +04:00
|
|
|
if (del_timer(&dev->watchdog_timer))
|
2006-03-21 09:32:28 +03:00
|
|
|
dev_put(dev);
|
2006-06-09 23:20:56 +04:00
|
|
|
netif_tx_unlock_bh(dev);
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
[NET]: Make NAPI polling independent of struct net_device objects.
Several devices have multiple independant RX queues per net
device, and some have a single interrupt doorbell for several
queues.
In either case, it's easier to support layouts like that if the
structure representing the poll is independant from the net
device itself.
The signature of the ->poll() call back goes from:
int foo_poll(struct net_device *dev, int *budget)
to
int foo_poll(struct napi_struct *napi, int budget)
The caller is returned the number of RX packets processed (or
the number of "NAPI credits" consumed if you want to get
abstract). The callee no longer messes around bumping
dev->quota, *budget, etc. because that is all handled in the
caller upon return.
The napi_struct is to be embedded in the device driver private data
structures.
Furthermore, it is the driver's responsibility to disable all NAPI
instances in it's ->stop() device close handler. Since the
napi_struct is privatized into the driver's private data structures,
only the driver knows how to get at all of the napi_struct instances
it may have per-device.
With lots of help and suggestions from Rusty Russell, Roland Dreier,
Michael Chan, Jeff Garzik, and Jamal Hadi Salim.
Bug fixes from Thomas Graf, Roland Dreier, Peter Zijlstra,
Joseph Fannin, Scott Wood, Hans J. Koch, and Michael Chan.
[ Ported to current tree and all drivers converted. Integrated
Stephen's follow-on kerneldoc additions, and restored poll_list
handling to the old style to fix mutual exclusion issues. -DaveM ]
Signed-off-by: Stephen Hemminger <shemminger@linux-foundation.org>
Signed-off-by: David S. Miller <davem@davemloft.net>
2007-10-04 03:41:36 +04:00
|
|
|
/**
|
|
|
|
* netif_carrier_on - set carrier
|
|
|
|
* @dev: network device
|
|
|
|
*
|
|
|
|
* Device has detected that carrier.
|
|
|
|
*/
|
2005-08-12 02:32:53 +04:00
|
|
|
void netif_carrier_on(struct net_device *dev)
|
|
|
|
{
|
2007-10-18 10:26:43 +04:00
|
|
|
if (test_and_clear_bit(__LINK_STATE_NOCARRIER, &dev->state)) {
|
2008-11-20 02:33:54 +03:00
|
|
|
if (dev->reg_state == NETREG_UNINITIALIZED)
|
|
|
|
return;
|
2014-03-29 20:48:35 +04:00
|
|
|
atomic_inc(&dev->carrier_changes);
|
2005-08-12 02:32:53 +04:00
|
|
|
linkwatch_fire_event(dev);
|
2007-10-18 10:26:43 +04:00
|
|
|
if (netif_running(dev))
|
|
|
|
__netdev_watchdog_up(dev);
|
|
|
|
}
|
2005-08-12 02:32:53 +04:00
|
|
|
}
|
2008-01-23 09:10:23 +03:00
|
|
|
EXPORT_SYMBOL(netif_carrier_on);
|
2005-08-12 02:32:53 +04:00
|
|
|
|
[NET]: Make NAPI polling independent of struct net_device objects.
Several devices have multiple independant RX queues per net
device, and some have a single interrupt doorbell for several
queues.
In either case, it's easier to support layouts like that if the
structure representing the poll is independant from the net
device itself.
The signature of the ->poll() call back goes from:
int foo_poll(struct net_device *dev, int *budget)
to
int foo_poll(struct napi_struct *napi, int budget)
The caller is returned the number of RX packets processed (or
the number of "NAPI credits" consumed if you want to get
abstract). The callee no longer messes around bumping
dev->quota, *budget, etc. because that is all handled in the
caller upon return.
The napi_struct is to be embedded in the device driver private data
structures.
Furthermore, it is the driver's responsibility to disable all NAPI
instances in it's ->stop() device close handler. Since the
napi_struct is privatized into the driver's private data structures,
only the driver knows how to get at all of the napi_struct instances
it may have per-device.
With lots of help and suggestions from Rusty Russell, Roland Dreier,
Michael Chan, Jeff Garzik, and Jamal Hadi Salim.
Bug fixes from Thomas Graf, Roland Dreier, Peter Zijlstra,
Joseph Fannin, Scott Wood, Hans J. Koch, and Michael Chan.
[ Ported to current tree and all drivers converted. Integrated
Stephen's follow-on kerneldoc additions, and restored poll_list
handling to the old style to fix mutual exclusion issues. -DaveM ]
Signed-off-by: Stephen Hemminger <shemminger@linux-foundation.org>
Signed-off-by: David S. Miller <davem@davemloft.net>
2007-10-04 03:41:36 +04:00
|
|
|
/**
|
|
|
|
* netif_carrier_off - clear carrier
|
|
|
|
* @dev: network device
|
|
|
|
*
|
|
|
|
* Device has detected loss of carrier.
|
|
|
|
*/
|
2005-08-12 02:32:53 +04:00
|
|
|
void netif_carrier_off(struct net_device *dev)
|
|
|
|
{
|
2008-11-20 02:33:54 +03:00
|
|
|
if (!test_and_set_bit(__LINK_STATE_NOCARRIER, &dev->state)) {
|
|
|
|
if (dev->reg_state == NETREG_UNINITIALIZED)
|
|
|
|
return;
|
2014-03-29 20:48:35 +04:00
|
|
|
atomic_inc(&dev->carrier_changes);
|
2005-08-12 02:32:53 +04:00
|
|
|
linkwatch_fire_event(dev);
|
2008-11-20 02:33:54 +03:00
|
|
|
}
|
2005-08-12 02:32:53 +04:00
|
|
|
}
|
2008-01-23 09:10:23 +03:00
|
|
|
EXPORT_SYMBOL(netif_carrier_off);
|
2005-08-12 02:32:53 +04:00
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
/* "NOOP" scheduler: the best scheduler, recommended for all interfaces
|
|
|
|
under all circumstances. It is difficult to invent anything faster or
|
|
|
|
cheaper.
|
|
|
|
*/
|
|
|
|
|
2016-06-22 09:16:49 +03:00
|
|
|
static int noop_enqueue(struct sk_buff *skb, struct Qdisc *qdisc,
|
|
|
|
struct sk_buff **to_free)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
2016-06-22 09:16:49 +03:00
|
|
|
__qdisc_drop(skb, to_free);
|
2005-04-17 02:20:36 +04:00
|
|
|
return NET_XMIT_CN;
|
|
|
|
}
|
|
|
|
|
2013-12-10 16:55:31 +04:00
|
|
|
static struct sk_buff *noop_dequeue(struct Qdisc *qdisc)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2007-11-14 12:44:41 +03:00
|
|
|
struct Qdisc_ops noop_qdisc_ops __read_mostly = {
|
2005-04-17 02:20:36 +04:00
|
|
|
.id = "noop",
|
|
|
|
.priv_size = 0,
|
|
|
|
.enqueue = noop_enqueue,
|
|
|
|
.dequeue = noop_dequeue,
|
2008-10-31 10:45:27 +03:00
|
|
|
.peek = noop_dequeue,
|
2005-04-17 02:20:36 +04:00
|
|
|
.owner = THIS_MODULE,
|
|
|
|
};
|
|
|
|
|
2008-07-16 12:42:40 +04:00
|
|
|
static struct netdev_queue noop_netdev_queue = {
|
|
|
|
.qdisc = &noop_qdisc,
|
2008-10-20 10:37:47 +04:00
|
|
|
.qdisc_sleeping = &noop_qdisc,
|
2008-07-16 12:42:40 +04:00
|
|
|
};
|
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
struct Qdisc noop_qdisc = {
|
|
|
|
.enqueue = noop_enqueue,
|
|
|
|
.dequeue = noop_dequeue,
|
|
|
|
.flags = TCQ_F_BUILTIN,
|
2007-02-09 17:25:16 +03:00
|
|
|
.ops = &noop_qdisc_ops,
|
2008-07-17 11:53:03 +04:00
|
|
|
.q.lock = __SPIN_LOCK_UNLOCKED(noop_qdisc.q.lock),
|
2008-07-16 12:42:40 +04:00
|
|
|
.dev_queue = &noop_netdev_queue,
|
2016-06-06 19:37:15 +03:00
|
|
|
.running = SEQCNT_ZERO(noop_qdisc.running),
|
2010-10-15 23:22:34 +04:00
|
|
|
.busylock = __SPIN_LOCK_UNLOCKED(noop_qdisc.busylock),
|
2005-04-17 02:20:36 +04:00
|
|
|
};
|
2008-01-23 09:10:23 +03:00
|
|
|
EXPORT_SYMBOL(noop_qdisc);
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2015-08-27 22:21:38 +03:00
|
|
|
static int noqueue_init(struct Qdisc *qdisc, struct nlattr *opt)
|
|
|
|
{
|
|
|
|
/* register_qdisc() assigns a default of noop_enqueue if unset,
|
|
|
|
* but __dev_queue_xmit() treats noqueue only as such
|
|
|
|
* if this is NULL - so clear it here. */
|
|
|
|
qdisc->enqueue = NULL;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct Qdisc_ops noqueue_qdisc_ops __read_mostly = {
|
2005-04-17 02:20:36 +04:00
|
|
|
.id = "noqueue",
|
|
|
|
.priv_size = 0,
|
2015-08-27 22:21:38 +03:00
|
|
|
.init = noqueue_init,
|
2005-04-17 02:20:36 +04:00
|
|
|
.enqueue = noop_enqueue,
|
|
|
|
.dequeue = noop_dequeue,
|
2008-10-31 10:45:27 +03:00
|
|
|
.peek = noop_dequeue,
|
2005-04-17 02:20:36 +04:00
|
|
|
.owner = THIS_MODULE,
|
|
|
|
};
|
|
|
|
|
2011-01-19 22:26:56 +03:00
|
|
|
static const u8 prio2band[TC_PRIO_MAX + 1] = {
|
|
|
|
1, 2, 2, 2, 1, 2, 0, 0 , 1, 1, 1, 1, 1, 1, 1, 1
|
|
|
|
};
|
2008-07-21 20:56:13 +04:00
|
|
|
|
|
|
|
/* 3-band FIFO queue: old style, but should be a bit faster than
|
|
|
|
generic prio+fifo combination.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define PFIFO_FAST_BANDS 3
|
|
|
|
|
2009-08-19 01:55:59 +04:00
|
|
|
/*
|
|
|
|
* Private data for a pfifo_fast scheduler containing:
|
|
|
|
* - queues for the three band
|
|
|
|
* - bitmap indicating which of the bands contain skbs
|
|
|
|
*/
|
|
|
|
struct pfifo_fast_priv {
|
|
|
|
u32 bitmap;
|
2016-09-18 01:57:34 +03:00
|
|
|
struct qdisc_skb_head q[PFIFO_FAST_BANDS];
|
2009-08-19 01:55:59 +04:00
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Convert a bitmap to the first band number where an skb is queued, where:
|
|
|
|
* bitmap=0 means there are no skbs on any band.
|
|
|
|
* bitmap=1 means there is an skb on band 0.
|
|
|
|
* bitmap=7 means there are skbs on all 3 bands, etc.
|
|
|
|
*/
|
|
|
|
static const int bitmap2band[] = {-1, 0, 1, 0, 2, 0, 1, 0};
|
|
|
|
|
2016-09-18 01:57:34 +03:00
|
|
|
static inline struct qdisc_skb_head *band2list(struct pfifo_fast_priv *priv,
|
2009-08-19 01:55:59 +04:00
|
|
|
int band)
|
2008-07-21 20:56:13 +04:00
|
|
|
{
|
2009-08-19 01:55:59 +04:00
|
|
|
return priv->q + band;
|
2008-07-21 20:56:13 +04:00
|
|
|
}
|
|
|
|
|
2016-06-22 09:16:49 +03:00
|
|
|
static int pfifo_fast_enqueue(struct sk_buff *skb, struct Qdisc *qdisc,
|
|
|
|
struct sk_buff **to_free)
|
2005-06-19 09:58:35 +04:00
|
|
|
{
|
2016-09-18 01:57:31 +03:00
|
|
|
if (qdisc->q.qlen < qdisc_dev(qdisc)->tx_queue_len) {
|
2009-08-31 09:20:28 +04:00
|
|
|
int band = prio2band[skb->priority & TC_PRIO_MAX];
|
|
|
|
struct pfifo_fast_priv *priv = qdisc_priv(qdisc);
|
2016-09-18 01:57:34 +03:00
|
|
|
struct qdisc_skb_head *list = band2list(priv, band);
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2009-08-19 01:55:59 +04:00
|
|
|
priv->bitmap |= (1 << band);
|
2008-07-21 20:56:13 +04:00
|
|
|
qdisc->q.qlen++;
|
2005-06-19 09:58:15 +04:00
|
|
|
return __qdisc_enqueue_tail(skb, qdisc, list);
|
2008-07-21 20:56:13 +04:00
|
|
|
}
|
2005-06-19 09:58:15 +04:00
|
|
|
|
2016-06-22 09:16:49 +03:00
|
|
|
return qdisc_drop(skb, qdisc, to_free);
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2011-01-19 22:26:56 +03:00
|
|
|
static struct sk_buff *pfifo_fast_dequeue(struct Qdisc *qdisc)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
2009-08-19 01:55:59 +04:00
|
|
|
struct pfifo_fast_priv *priv = qdisc_priv(qdisc);
|
|
|
|
int band = bitmap2band[priv->bitmap];
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2009-08-19 01:55:59 +04:00
|
|
|
if (likely(band >= 0)) {
|
2016-09-18 01:57:34 +03:00
|
|
|
struct qdisc_skb_head *qh = band2list(priv, band);
|
|
|
|
struct sk_buff *skb = __qdisc_dequeue_head(qh);
|
2016-09-18 01:57:32 +03:00
|
|
|
|
|
|
|
if (likely(skb != NULL)) {
|
|
|
|
qdisc_qstats_backlog_dec(qdisc, skb);
|
|
|
|
qdisc_bstats_update(qdisc, skb);
|
|
|
|
}
|
2009-08-19 01:55:59 +04:00
|
|
|
|
|
|
|
qdisc->q.qlen--;
|
2016-09-18 01:57:34 +03:00
|
|
|
if (qh->qlen == 0)
|
2009-08-19 01:55:59 +04:00
|
|
|
priv->bitmap &= ~(1 << band);
|
|
|
|
|
|
|
|
return skb;
|
2008-07-21 20:56:13 +04:00
|
|
|
}
|
2005-06-19 09:58:53 +04:00
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2011-01-19 22:26:56 +03:00
|
|
|
static struct sk_buff *pfifo_fast_peek(struct Qdisc *qdisc)
|
2008-10-31 10:45:27 +03:00
|
|
|
{
|
2009-08-19 01:55:59 +04:00
|
|
|
struct pfifo_fast_priv *priv = qdisc_priv(qdisc);
|
|
|
|
int band = bitmap2band[priv->bitmap];
|
|
|
|
|
|
|
|
if (band >= 0) {
|
2016-09-18 01:57:34 +03:00
|
|
|
struct qdisc_skb_head *qh = band2list(priv, band);
|
2008-10-31 10:45:27 +03:00
|
|
|
|
2016-09-18 01:57:34 +03:00
|
|
|
return qh->head;
|
2008-10-31 10:45:27 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2011-01-19 22:26:56 +03:00
|
|
|
static void pfifo_fast_reset(struct Qdisc *qdisc)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
2008-07-21 20:56:13 +04:00
|
|
|
int prio;
|
2009-08-19 01:55:59 +04:00
|
|
|
struct pfifo_fast_priv *priv = qdisc_priv(qdisc);
|
2008-07-21 20:56:13 +04:00
|
|
|
|
|
|
|
for (prio = 0; prio < PFIFO_FAST_BANDS; prio++)
|
2016-06-14 06:21:50 +03:00
|
|
|
__qdisc_reset_queue(band2list(priv, prio));
|
2008-07-21 20:56:13 +04:00
|
|
|
|
2009-08-19 01:55:59 +04:00
|
|
|
priv->bitmap = 0;
|
2005-06-19 09:58:15 +04:00
|
|
|
qdisc->qstats.backlog = 0;
|
2008-07-21 20:56:13 +04:00
|
|
|
qdisc->q.qlen = 0;
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2008-07-21 20:56:13 +04:00
|
|
|
static int pfifo_fast_dump(struct Qdisc *qdisc, struct sk_buff *skb)
|
|
|
|
{
|
|
|
|
struct tc_prio_qopt opt = { .bands = PFIFO_FAST_BANDS };
|
|
|
|
|
2011-01-19 22:26:56 +03:00
|
|
|
memcpy(&opt.priomap, prio2band, TC_PRIO_MAX + 1);
|
2012-03-29 13:11:39 +04:00
|
|
|
if (nla_put(skb, TCA_OPTIONS, sizeof(opt), &opt))
|
|
|
|
goto nla_put_failure;
|
2008-07-21 20:56:13 +04:00
|
|
|
return skb->len;
|
|
|
|
|
|
|
|
nla_put_failure:
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int pfifo_fast_init(struct Qdisc *qdisc, struct nlattr *opt)
|
|
|
|
{
|
|
|
|
int prio;
|
2009-08-19 01:55:59 +04:00
|
|
|
struct pfifo_fast_priv *priv = qdisc_priv(qdisc);
|
2008-07-21 20:56:13 +04:00
|
|
|
|
|
|
|
for (prio = 0; prio < PFIFO_FAST_BANDS; prio++)
|
2016-09-18 01:57:34 +03:00
|
|
|
qdisc_skb_head_init(band2list(priv, prio));
|
2008-07-21 20:56:13 +04:00
|
|
|
|
2011-01-22 03:26:09 +03:00
|
|
|
/* Can by-pass the queue discipline */
|
|
|
|
qdisc->flags |= TCQ_F_CAN_BYPASS;
|
2008-07-21 20:56:13 +04:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2009-09-06 12:58:51 +04:00
|
|
|
struct Qdisc_ops pfifo_fast_ops __read_mostly = {
|
2008-07-21 20:56:13 +04:00
|
|
|
.id = "pfifo_fast",
|
2009-08-19 01:55:59 +04:00
|
|
|
.priv_size = sizeof(struct pfifo_fast_priv),
|
2008-07-21 20:56:13 +04:00
|
|
|
.enqueue = pfifo_fast_enqueue,
|
|
|
|
.dequeue = pfifo_fast_dequeue,
|
2008-10-31 10:45:27 +03:00
|
|
|
.peek = pfifo_fast_peek,
|
2008-07-21 20:56:13 +04:00
|
|
|
.init = pfifo_fast_init,
|
|
|
|
.reset = pfifo_fast_reset,
|
|
|
|
.dump = pfifo_fast_dump,
|
2005-04-17 02:20:36 +04:00
|
|
|
.owner = THIS_MODULE,
|
|
|
|
};
|
2016-03-02 19:21:43 +03:00
|
|
|
EXPORT_SYMBOL(pfifo_fast_ops);
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2012-09-05 05:02:56 +04:00
|
|
|
static struct lock_class_key qdisc_tx_busylock;
|
2016-06-06 19:37:15 +03:00
|
|
|
static struct lock_class_key qdisc_running_key;
|
2012-09-05 05:02:56 +04:00
|
|
|
|
2008-07-09 04:06:30 +04:00
|
|
|
struct Qdisc *qdisc_alloc(struct netdev_queue *dev_queue,
|
2013-08-31 21:15:50 +04:00
|
|
|
const struct Qdisc_ops *ops)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
|
|
|
void *p;
|
|
|
|
struct Qdisc *sch;
|
2011-03-03 22:10:02 +03:00
|
|
|
unsigned int size = QDISC_ALIGN(sizeof(*sch)) + ops->priv_size;
|
2005-07-06 01:15:09 +04:00
|
|
|
int err = -ENOBUFS;
|
2017-10-17 04:01:23 +03:00
|
|
|
struct net_device *dev;
|
|
|
|
|
|
|
|
if (!dev_queue) {
|
|
|
|
err = -EINVAL;
|
|
|
|
goto errout;
|
|
|
|
}
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2017-10-17 04:01:23 +03:00
|
|
|
dev = dev_queue->dev;
|
2010-11-29 11:14:37 +03:00
|
|
|
p = kzalloc_node(size, GFP_KERNEL,
|
|
|
|
netdev_queue_numa_node_read(dev_queue));
|
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
if (!p)
|
2005-07-06 01:15:09 +04:00
|
|
|
goto errout;
|
|
|
|
sch = (struct Qdisc *) QDISC_ALIGN((unsigned long) p);
|
2011-03-03 22:10:02 +03:00
|
|
|
/* if we got non aligned memory, ask more and do alignment ourself */
|
|
|
|
if (sch != p) {
|
|
|
|
kfree(p);
|
|
|
|
p = kzalloc_node(size + QDISC_ALIGNTO - 1, GFP_KERNEL,
|
|
|
|
netdev_queue_numa_node_read(dev_queue));
|
|
|
|
if (!p)
|
|
|
|
goto errout;
|
|
|
|
sch = (struct Qdisc *) QDISC_ALIGN((unsigned long) p);
|
|
|
|
sch->padded = (char *) sch - (char *) p;
|
|
|
|
}
|
2016-09-18 01:57:34 +03:00
|
|
|
qdisc_skb_head_init(&sch->q);
|
|
|
|
spin_lock_init(&sch->q.lock);
|
2012-09-05 05:02:56 +04:00
|
|
|
|
2010-06-02 16:09:29 +04:00
|
|
|
spin_lock_init(&sch->busylock);
|
2012-09-05 05:02:56 +04:00
|
|
|
lockdep_set_class(&sch->busylock,
|
|
|
|
dev->qdisc_tx_busylock ?: &qdisc_tx_busylock);
|
|
|
|
|
2016-06-06 19:37:15 +03:00
|
|
|
seqcount_init(&sch->running);
|
|
|
|
lockdep_set_class(&sch->running,
|
|
|
|
dev->qdisc_running_key ?: &qdisc_running_key);
|
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
sch->ops = ops;
|
|
|
|
sch->enqueue = ops->enqueue;
|
|
|
|
sch->dequeue = ops->dequeue;
|
2008-07-09 03:55:56 +04:00
|
|
|
sch->dev_queue = dev_queue;
|
2012-09-05 05:02:56 +04:00
|
|
|
dev_hold(dev);
|
2017-07-04 15:53:07 +03:00
|
|
|
refcount_set(&sch->refcnt, 1);
|
2005-07-06 01:15:09 +04:00
|
|
|
|
|
|
|
return sch;
|
|
|
|
errout:
|
2008-06-28 06:51:35 +04:00
|
|
|
return ERR_PTR(err);
|
2005-07-06 01:15:09 +04:00
|
|
|
}
|
|
|
|
|
2010-10-16 17:04:08 +04:00
|
|
|
struct Qdisc *qdisc_create_dflt(struct netdev_queue *dev_queue,
|
2013-08-31 21:15:50 +04:00
|
|
|
const struct Qdisc_ops *ops,
|
|
|
|
unsigned int parentid)
|
2005-07-06 01:15:09 +04:00
|
|
|
{
|
|
|
|
struct Qdisc *sch;
|
2007-02-09 17:25:16 +03:00
|
|
|
|
2013-08-28 03:19:08 +04:00
|
|
|
if (!try_module_get(ops->owner))
|
2016-08-24 19:39:02 +03:00
|
|
|
return NULL;
|
2013-08-28 03:19:08 +04:00
|
|
|
|
2008-07-09 04:06:30 +04:00
|
|
|
sch = qdisc_alloc(dev_queue, ops);
|
2016-08-24 19:39:02 +03:00
|
|
|
if (IS_ERR(sch)) {
|
|
|
|
module_put(ops->owner);
|
|
|
|
return NULL;
|
|
|
|
}
|
2006-11-30 04:35:18 +03:00
|
|
|
sch->parent = parentid;
|
2005-07-06 01:15:09 +04:00
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
if (!ops->init || ops->init(sch, NULL) == 0)
|
|
|
|
return sch;
|
|
|
|
|
2005-08-23 21:12:44 +04:00
|
|
|
qdisc_destroy(sch);
|
2005-04-17 02:20:36 +04:00
|
|
|
return NULL;
|
|
|
|
}
|
2008-01-23 09:10:23 +03:00
|
|
|
EXPORT_SYMBOL(qdisc_create_dflt);
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2008-08-03 07:02:43 +04:00
|
|
|
/* Under qdisc_lock(qdisc) and BH! */
|
2005-04-17 02:20:36 +04:00
|
|
|
|
|
|
|
void qdisc_reset(struct Qdisc *qdisc)
|
|
|
|
{
|
2007-11-14 12:44:41 +03:00
|
|
|
const struct Qdisc_ops *ops = qdisc->ops;
|
2005-04-17 02:20:36 +04:00
|
|
|
|
|
|
|
if (ops->reset)
|
|
|
|
ops->reset(qdisc);
|
2008-11-03 13:52:50 +03:00
|
|
|
|
net_sched: generalize bulk dequeue
When qdisc bulk dequeue was added in linux-3.18 (commit
5772e9a3463b "qdisc: bulk dequeue support for qdiscs
with TCQ_F_ONETXQUEUE"), it was constrained to some
specific qdiscs.
With some extra care, we can extend this to all qdiscs,
so that typical traffic shaping solutions can benefit from
small batches (8 packets in this patch).
For example, HTB is often used on some multi queue device.
And bonding/team are multi queue devices...
Idea is to bulk-dequeue packets mapping to the same transmit queue.
This brings between 35 and 80 % performance increase in HTB setup
under pressure on a bonding setup :
1) NUMA node contention : 610,000 pps -> 1,110,000 pps
2) No node contention : 1,380,000 pps -> 1,930,000 pps
Now we should work to add batches on the enqueue() side ;)
Signed-off-by: Eric Dumazet <edumazet@google.com>
Cc: John Fastabend <john.r.fastabend@intel.com>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Hannes Frederic Sowa <hannes@stressinduktion.org>
Cc: Florian Westphal <fw@strlen.de>
Cc: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-06-22 09:16:52 +03:00
|
|
|
kfree_skb(qdisc->skb_bad_txq);
|
|
|
|
qdisc->skb_bad_txq = NULL;
|
|
|
|
|
2009-08-06 05:44:21 +04:00
|
|
|
if (qdisc->gso_skb) {
|
2014-09-03 14:12:50 +04:00
|
|
|
kfree_skb_list(qdisc->gso_skb);
|
2009-08-06 05:44:21 +04:00
|
|
|
qdisc->gso_skb = NULL;
|
|
|
|
}
|
net_sched: generalize bulk dequeue
When qdisc bulk dequeue was added in linux-3.18 (commit
5772e9a3463b "qdisc: bulk dequeue support for qdiscs
with TCQ_F_ONETXQUEUE"), it was constrained to some
specific qdiscs.
With some extra care, we can extend this to all qdiscs,
so that typical traffic shaping solutions can benefit from
small batches (8 packets in this patch).
For example, HTB is often used on some multi queue device.
And bonding/team are multi queue devices...
Idea is to bulk-dequeue packets mapping to the same transmit queue.
This brings between 35 and 80 % performance increase in HTB setup
under pressure on a bonding setup :
1) NUMA node contention : 610,000 pps -> 1,110,000 pps
2) No node contention : 1,380,000 pps -> 1,930,000 pps
Now we should work to add batches on the enqueue() side ;)
Signed-off-by: Eric Dumazet <edumazet@google.com>
Cc: John Fastabend <john.r.fastabend@intel.com>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Hannes Frederic Sowa <hannes@stressinduktion.org>
Cc: Florian Westphal <fw@strlen.de>
Cc: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-06-22 09:16:52 +03:00
|
|
|
qdisc->q.qlen = 0;
|
2017-09-20 15:45:36 +03:00
|
|
|
qdisc->qstats.backlog = 0;
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
2008-01-23 09:10:23 +03:00
|
|
|
EXPORT_SYMBOL(qdisc_reset);
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2017-09-19 23:15:42 +03:00
|
|
|
static void qdisc_free(struct Qdisc *qdisc)
|
2010-03-31 11:06:04 +04:00
|
|
|
{
|
2016-01-05 20:11:36 +03:00
|
|
|
if (qdisc_is_percpu_stats(qdisc)) {
|
2014-09-28 22:52:56 +04:00
|
|
|
free_percpu(qdisc->cpu_bstats);
|
2016-01-05 20:11:36 +03:00
|
|
|
free_percpu(qdisc->cpu_qstats);
|
|
|
|
}
|
2014-09-28 22:52:56 +04:00
|
|
|
|
2010-03-31 11:06:04 +04:00
|
|
|
kfree((char *) qdisc - qdisc->padded);
|
|
|
|
}
|
|
|
|
|
2008-08-18 09:31:26 +04:00
|
|
|
void qdisc_destroy(struct Qdisc *qdisc)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
2008-07-17 11:47:45 +04:00
|
|
|
const struct Qdisc_ops *ops = qdisc->ops;
|
|
|
|
|
2008-08-18 09:31:26 +04:00
|
|
|
if (qdisc->flags & TCQ_F_BUILTIN ||
|
2017-07-04 15:53:07 +03:00
|
|
|
!refcount_dec_and_test(&qdisc->refcnt))
|
2008-08-18 09:31:26 +04:00
|
|
|
return;
|
|
|
|
|
2008-07-21 05:13:01 +04:00
|
|
|
#ifdef CONFIG_NET_SCHED
|
2016-08-10 12:05:15 +03:00
|
|
|
qdisc_hash_del(qdisc);
|
2008-08-22 14:24:05 +04:00
|
|
|
|
2011-01-20 06:48:19 +03:00
|
|
|
qdisc_put_stab(rtnl_dereference(qdisc->stab));
|
2008-07-21 05:13:01 +04:00
|
|
|
#endif
|
2016-12-04 20:48:16 +03:00
|
|
|
gen_kill_estimator(&qdisc->rate_est);
|
2008-07-17 11:47:45 +04:00
|
|
|
if (ops->reset)
|
|
|
|
ops->reset(qdisc);
|
|
|
|
if (ops->destroy)
|
|
|
|
ops->destroy(qdisc);
|
|
|
|
|
|
|
|
module_put(ops->owner);
|
|
|
|
dev_put(qdisc_dev(qdisc));
|
|
|
|
|
2014-09-03 14:12:50 +04:00
|
|
|
kfree_skb_list(qdisc->gso_skb);
|
net_sched: generalize bulk dequeue
When qdisc bulk dequeue was added in linux-3.18 (commit
5772e9a3463b "qdisc: bulk dequeue support for qdiscs
with TCQ_F_ONETXQUEUE"), it was constrained to some
specific qdiscs.
With some extra care, we can extend this to all qdiscs,
so that typical traffic shaping solutions can benefit from
small batches (8 packets in this patch).
For example, HTB is often used on some multi queue device.
And bonding/team are multi queue devices...
Idea is to bulk-dequeue packets mapping to the same transmit queue.
This brings between 35 and 80 % performance increase in HTB setup
under pressure on a bonding setup :
1) NUMA node contention : 610,000 pps -> 1,110,000 pps
2) No node contention : 1,380,000 pps -> 1,930,000 pps
Now we should work to add batches on the enqueue() side ;)
Signed-off-by: Eric Dumazet <edumazet@google.com>
Cc: John Fastabend <john.r.fastabend@intel.com>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Hannes Frederic Sowa <hannes@stressinduktion.org>
Cc: Florian Westphal <fw@strlen.de>
Cc: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-06-22 09:16:52 +03:00
|
|
|
kfree_skb(qdisc->skb_bad_txq);
|
2017-09-19 23:15:42 +03:00
|
|
|
qdisc_free(qdisc);
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
2008-01-23 09:10:23 +03:00
|
|
|
EXPORT_SYMBOL(qdisc_destroy);
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2009-09-04 10:41:20 +04:00
|
|
|
/* Attach toplevel qdisc to device queue. */
|
|
|
|
struct Qdisc *dev_graft_qdisc(struct netdev_queue *dev_queue,
|
|
|
|
struct Qdisc *qdisc)
|
|
|
|
{
|
|
|
|
struct Qdisc *oqdisc = dev_queue->qdisc_sleeping;
|
|
|
|
spinlock_t *root_lock;
|
|
|
|
|
|
|
|
root_lock = qdisc_lock(oqdisc);
|
|
|
|
spin_lock_bh(root_lock);
|
|
|
|
|
|
|
|
/* Prune old scheduler */
|
2017-07-04 15:53:07 +03:00
|
|
|
if (oqdisc && refcount_read(&oqdisc->refcnt) <= 1)
|
2009-09-04 10:41:20 +04:00
|
|
|
qdisc_reset(oqdisc);
|
|
|
|
|
|
|
|
/* ... and graft new one */
|
|
|
|
if (qdisc == NULL)
|
|
|
|
qdisc = &noop_qdisc;
|
|
|
|
dev_queue->qdisc_sleeping = qdisc;
|
|
|
|
rcu_assign_pointer(dev_queue->qdisc, &noop_qdisc);
|
|
|
|
|
|
|
|
spin_unlock_bh(root_lock);
|
|
|
|
|
|
|
|
return oqdisc;
|
|
|
|
}
|
2011-01-17 11:06:09 +03:00
|
|
|
EXPORT_SYMBOL(dev_graft_qdisc);
|
2009-09-04 10:41:20 +04:00
|
|
|
|
2008-07-17 11:34:19 +04:00
|
|
|
static void attach_one_default_qdisc(struct net_device *dev,
|
|
|
|
struct netdev_queue *dev_queue,
|
|
|
|
void *_unused)
|
|
|
|
{
|
2015-08-27 22:21:39 +03:00
|
|
|
struct Qdisc *qdisc;
|
|
|
|
const struct Qdisc_ops *ops = default_qdisc_ops;
|
2008-07-17 11:34:19 +04:00
|
|
|
|
2015-08-27 22:21:39 +03:00
|
|
|
if (dev->priv_flags & IFF_NO_QUEUE)
|
|
|
|
ops = &noqueue_qdisc_ops;
|
|
|
|
|
|
|
|
qdisc = qdisc_create_dflt(dev_queue, ops, TC_H_ROOT);
|
|
|
|
if (!qdisc) {
|
|
|
|
netdev_info(dev, "activation failed\n");
|
|
|
|
return;
|
2008-07-17 11:34:19 +04:00
|
|
|
}
|
2015-08-27 22:21:39 +03:00
|
|
|
if (!netif_is_multiqueue(dev))
|
2015-12-02 07:08:51 +03:00
|
|
|
qdisc->flags |= TCQ_F_ONETXQUEUE | TCQ_F_NOPARENT;
|
2008-07-17 11:34:19 +04:00
|
|
|
dev_queue->qdisc_sleeping = qdisc;
|
|
|
|
}
|
|
|
|
|
2009-09-06 12:58:51 +04:00
|
|
|
static void attach_default_qdiscs(struct net_device *dev)
|
|
|
|
{
|
|
|
|
struct netdev_queue *txq;
|
|
|
|
struct Qdisc *qdisc;
|
|
|
|
|
|
|
|
txq = netdev_get_tx_queue(dev, 0);
|
|
|
|
|
2015-08-13 20:01:07 +03:00
|
|
|
if (!netif_is_multiqueue(dev) ||
|
|
|
|
dev->priv_flags & IFF_NO_QUEUE) {
|
2009-09-06 12:58:51 +04:00
|
|
|
netdev_for_each_tx_queue(dev, attach_one_default_qdisc, NULL);
|
|
|
|
dev->qdisc = txq->qdisc_sleeping;
|
2017-08-25 07:12:28 +03:00
|
|
|
qdisc_refcount_inc(dev->qdisc);
|
2009-09-06 12:58:51 +04:00
|
|
|
} else {
|
2010-10-16 17:04:08 +04:00
|
|
|
qdisc = qdisc_create_dflt(txq, &mq_qdisc_ops, TC_H_ROOT);
|
2009-09-06 12:58:51 +04:00
|
|
|
if (qdisc) {
|
|
|
|
dev->qdisc = qdisc;
|
2013-12-13 03:41:56 +04:00
|
|
|
qdisc->ops->attach(qdisc);
|
2009-09-06 12:58:51 +04:00
|
|
|
}
|
|
|
|
}
|
2016-08-10 12:05:15 +03:00
|
|
|
#ifdef CONFIG_NET_SCHED
|
2017-04-05 04:51:30 +03:00
|
|
|
if (dev->qdisc != &noop_qdisc)
|
2017-03-08 18:03:32 +03:00
|
|
|
qdisc_hash_add(dev->qdisc, false);
|
2016-08-10 12:05:15 +03:00
|
|
|
#endif
|
2009-09-06 12:58:51 +04:00
|
|
|
}
|
|
|
|
|
2008-07-17 11:34:19 +04:00
|
|
|
static void transition_one_qdisc(struct net_device *dev,
|
|
|
|
struct netdev_queue *dev_queue,
|
|
|
|
void *_need_watchdog)
|
|
|
|
{
|
2008-07-17 11:53:03 +04:00
|
|
|
struct Qdisc *new_qdisc = dev_queue->qdisc_sleeping;
|
2008-07-17 11:34:19 +04:00
|
|
|
int *need_watchdog_p = _need_watchdog;
|
|
|
|
|
2008-08-18 08:51:03 +04:00
|
|
|
if (!(new_qdisc->flags & TCQ_F_BUILTIN))
|
|
|
|
clear_bit(__QDISC_STATE_DEACTIVATED, &new_qdisc->state);
|
|
|
|
|
2008-07-17 11:53:03 +04:00
|
|
|
rcu_assign_pointer(dev_queue->qdisc, new_qdisc);
|
2015-08-27 22:21:39 +03:00
|
|
|
if (need_watchdog_p) {
|
2009-05-18 07:55:16 +04:00
|
|
|
dev_queue->trans_start = 0;
|
2008-07-17 11:34:19 +04:00
|
|
|
*need_watchdog_p = 1;
|
2009-05-18 07:55:16 +04:00
|
|
|
}
|
2008-07-17 11:34:19 +04:00
|
|
|
}
|
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
void dev_activate(struct net_device *dev)
|
|
|
|
{
|
2008-07-17 11:34:19 +04:00
|
|
|
int need_watchdog;
|
2008-07-09 04:42:10 +04:00
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
/* No queueing discipline is attached to device;
|
2013-08-28 03:19:08 +04:00
|
|
|
* create default one for devices, which need queueing
|
|
|
|
* and noqueue_qdisc for virtual interfaces
|
2005-04-17 02:20:36 +04:00
|
|
|
*/
|
|
|
|
|
2009-09-06 12:58:51 +04:00
|
|
|
if (dev->qdisc == &noop_qdisc)
|
|
|
|
attach_default_qdiscs(dev);
|
2009-09-04 10:41:18 +04:00
|
|
|
|
2005-05-04 03:18:52 +04:00
|
|
|
if (!netif_carrier_ok(dev))
|
|
|
|
/* Delay activation until next carrier-on event */
|
|
|
|
return;
|
|
|
|
|
2008-07-17 11:34:19 +04:00
|
|
|
need_watchdog = 0;
|
|
|
|
netdev_for_each_tx_queue(dev, transition_one_qdisc, &need_watchdog);
|
2010-10-02 10:11:55 +04:00
|
|
|
if (dev_ingress_queue(dev))
|
|
|
|
transition_one_qdisc(dev, dev_ingress_queue(dev), NULL);
|
2008-07-17 11:34:19 +04:00
|
|
|
|
|
|
|
if (need_watchdog) {
|
2016-05-03 17:33:13 +03:00
|
|
|
netif_trans_update(dev);
|
2005-04-17 02:20:36 +04:00
|
|
|
dev_watchdog_up(dev);
|
|
|
|
}
|
2008-07-09 04:42:10 +04:00
|
|
|
}
|
2011-01-17 11:06:09 +03:00
|
|
|
EXPORT_SYMBOL(dev_activate);
|
2008-07-09 04:42:10 +04:00
|
|
|
|
2008-07-17 11:34:19 +04:00
|
|
|
static void dev_deactivate_queue(struct net_device *dev,
|
|
|
|
struct netdev_queue *dev_queue,
|
|
|
|
void *_qdisc_default)
|
2008-07-09 04:42:10 +04:00
|
|
|
{
|
2008-07-17 11:34:19 +04:00
|
|
|
struct Qdisc *qdisc_default = _qdisc_default;
|
2008-07-09 10:10:33 +04:00
|
|
|
struct Qdisc *qdisc;
|
|
|
|
|
2014-09-13 07:04:52 +04:00
|
|
|
qdisc = rtnl_dereference(dev_queue->qdisc);
|
2008-07-09 04:42:10 +04:00
|
|
|
if (qdisc) {
|
2008-07-17 11:53:03 +04:00
|
|
|
spin_lock_bh(qdisc_lock(qdisc));
|
|
|
|
|
2008-08-18 08:51:03 +04:00
|
|
|
if (!(qdisc->flags & TCQ_F_BUILTIN))
|
|
|
|
set_bit(__QDISC_STATE_DEACTIVATED, &qdisc->state);
|
|
|
|
|
2008-08-27 13:22:07 +04:00
|
|
|
rcu_assign_pointer(dev_queue->qdisc, qdisc_default);
|
2008-07-09 04:42:10 +04:00
|
|
|
qdisc_reset(qdisc);
|
2008-07-16 07:14:35 +04:00
|
|
|
|
2008-07-17 11:53:03 +04:00
|
|
|
spin_unlock_bh(qdisc_lock(qdisc));
|
2008-07-09 04:42:10 +04:00
|
|
|
}
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2008-08-18 08:58:07 +04:00
|
|
|
static bool some_qdisc_is_busy(struct net_device *dev)
|
2008-07-17 11:34:19 +04:00
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
for (i = 0; i < dev->num_tx_queues; i++) {
|
|
|
|
struct netdev_queue *dev_queue;
|
2008-07-16 12:42:40 +04:00
|
|
|
spinlock_t *root_lock;
|
2008-07-16 11:56:32 +04:00
|
|
|
struct Qdisc *q;
|
2008-07-17 11:34:19 +04:00
|
|
|
int val;
|
|
|
|
|
|
|
|
dev_queue = netdev_get_tx_queue(dev, i);
|
2008-08-14 02:18:38 +04:00
|
|
|
q = dev_queue->qdisc_sleeping;
|
2008-08-03 07:02:43 +04:00
|
|
|
root_lock = qdisc_lock(q);
|
2008-07-17 11:34:19 +04:00
|
|
|
|
2008-08-18 08:58:07 +04:00
|
|
|
spin_lock_bh(root_lock);
|
2008-07-17 11:34:19 +04:00
|
|
|
|
2010-06-02 14:23:51 +04:00
|
|
|
val = (qdisc_is_running(q) ||
|
2008-08-14 02:18:38 +04:00
|
|
|
test_bit(__QDISC_STATE_SCHED, &q->state));
|
2008-07-17 11:34:19 +04:00
|
|
|
|
2008-08-18 08:58:07 +04:00
|
|
|
spin_unlock_bh(root_lock);
|
2008-07-17 11:34:19 +04:00
|
|
|
|
|
|
|
if (val)
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2011-05-20 03:42:09 +04:00
|
|
|
/**
|
|
|
|
* dev_deactivate_many - deactivate transmissions on several devices
|
|
|
|
* @head: list of devices to deactivate
|
|
|
|
*
|
|
|
|
* This function returns only when all outstanding transmissions
|
|
|
|
* have completed, unless all devices are in dismantle phase.
|
|
|
|
*/
|
2010-12-13 15:44:07 +03:00
|
|
|
void dev_deactivate_many(struct list_head *head)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
2010-12-13 15:44:07 +03:00
|
|
|
struct net_device *dev;
|
2011-05-20 03:42:09 +04:00
|
|
|
bool sync_needed = false;
|
2007-05-11 01:12:47 +04:00
|
|
|
|
2013-10-06 06:26:05 +04:00
|
|
|
list_for_each_entry(dev, head, close_list) {
|
2010-12-13 15:44:07 +03:00
|
|
|
netdev_for_each_tx_queue(dev, dev_deactivate_queue,
|
|
|
|
&noop_qdisc);
|
|
|
|
if (dev_ingress_queue(dev))
|
|
|
|
dev_deactivate_queue(dev, dev_ingress_queue(dev),
|
|
|
|
&noop_qdisc);
|
|
|
|
|
|
|
|
dev_watchdog_down(dev);
|
2011-05-20 03:42:09 +04:00
|
|
|
sync_needed |= !dev->dismantle;
|
2010-12-13 15:44:07 +03:00
|
|
|
}
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2011-05-20 03:42:09 +04:00
|
|
|
/* Wait for outstanding qdisc-less dev_queue_xmit calls.
|
|
|
|
* This is avoided if all devices are in dismantle phase :
|
|
|
|
* Caller will call synchronize_net() for us
|
|
|
|
*/
|
|
|
|
if (sync_needed)
|
|
|
|
synchronize_net();
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2006-06-22 13:28:18 +04:00
|
|
|
/* Wait for outstanding qdisc_run calls. */
|
2013-10-06 06:26:05 +04:00
|
|
|
list_for_each_entry(dev, head, close_list)
|
2010-12-13 15:44:07 +03:00
|
|
|
while (some_qdisc_is_busy(dev))
|
|
|
|
yield();
|
|
|
|
}
|
|
|
|
|
|
|
|
void dev_deactivate(struct net_device *dev)
|
|
|
|
{
|
|
|
|
LIST_HEAD(single);
|
|
|
|
|
2013-10-06 06:26:05 +04:00
|
|
|
list_add(&dev->close_list, &single);
|
2010-12-13 15:44:07 +03:00
|
|
|
dev_deactivate_many(&single);
|
2011-02-20 22:49:45 +03:00
|
|
|
list_del(&single);
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
2011-01-17 11:06:09 +03:00
|
|
|
EXPORT_SYMBOL(dev_deactivate);
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2008-07-09 04:42:10 +04:00
|
|
|
static void dev_init_scheduler_queue(struct net_device *dev,
|
|
|
|
struct netdev_queue *dev_queue,
|
2008-07-17 11:34:19 +04:00
|
|
|
void *_qdisc)
|
2008-07-09 04:42:10 +04:00
|
|
|
{
|
2008-07-17 11:34:19 +04:00
|
|
|
struct Qdisc *qdisc = _qdisc;
|
|
|
|
|
2014-09-13 07:04:52 +04:00
|
|
|
rcu_assign_pointer(dev_queue->qdisc, qdisc);
|
2008-07-09 04:42:10 +04:00
|
|
|
dev_queue->qdisc_sleeping = qdisc;
|
|
|
|
}
|
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
void dev_init_scheduler(struct net_device *dev)
|
|
|
|
{
|
2009-09-04 10:41:18 +04:00
|
|
|
dev->qdisc = &noop_qdisc;
|
2008-07-17 11:34:19 +04:00
|
|
|
netdev_for_each_tx_queue(dev, dev_init_scheduler_queue, &noop_qdisc);
|
2010-10-02 10:11:55 +04:00
|
|
|
if (dev_ingress_queue(dev))
|
|
|
|
dev_init_scheduler_queue(dev, dev_ingress_queue(dev), &noop_qdisc);
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2017-10-17 03:29:17 +03:00
|
|
|
timer_setup(&dev->watchdog_timer, dev_watchdog, 0);
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2008-07-17 11:34:19 +04:00
|
|
|
static void shutdown_scheduler_queue(struct net_device *dev,
|
|
|
|
struct netdev_queue *dev_queue,
|
|
|
|
void *_qdisc_default)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
2008-07-09 04:42:10 +04:00
|
|
|
struct Qdisc *qdisc = dev_queue->qdisc_sleeping;
|
2008-07-17 11:34:19 +04:00
|
|
|
struct Qdisc *qdisc_default = _qdisc_default;
|
2008-07-09 04:42:10 +04:00
|
|
|
|
|
|
|
if (qdisc) {
|
2008-08-27 13:22:07 +04:00
|
|
|
rcu_assign_pointer(dev_queue->qdisc, qdisc_default);
|
2008-07-09 04:42:10 +04:00
|
|
|
dev_queue->qdisc_sleeping = qdisc_default;
|
2005-04-17 02:20:36 +04:00
|
|
|
|
|
|
|
qdisc_destroy(qdisc);
|
2007-02-09 17:25:16 +03:00
|
|
|
}
|
2008-07-09 04:42:10 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
void dev_shutdown(struct net_device *dev)
|
|
|
|
{
|
2008-07-17 11:34:19 +04:00
|
|
|
netdev_for_each_tx_queue(dev, shutdown_scheduler_queue, &noop_qdisc);
|
2010-10-02 10:11:55 +04:00
|
|
|
if (dev_ingress_queue(dev))
|
|
|
|
shutdown_scheduler_queue(dev, dev_ingress_queue(dev), &noop_qdisc);
|
2009-09-04 10:41:18 +04:00
|
|
|
qdisc_destroy(dev->qdisc);
|
|
|
|
dev->qdisc = &noop_qdisc;
|
|
|
|
|
2008-07-26 08:43:18 +04:00
|
|
|
WARN_ON(timer_pending(&dev->watchdog_timer));
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
2013-02-12 04:12:03 +04:00
|
|
|
|
2013-06-02 17:55:05 +04:00
|
|
|
void psched_ratecfg_precompute(struct psched_ratecfg *r,
|
2013-09-19 20:10:03 +04:00
|
|
|
const struct tc_ratespec *conf,
|
|
|
|
u64 rate64)
|
2013-02-12 04:12:03 +04:00
|
|
|
{
|
2013-06-02 17:55:05 +04:00
|
|
|
memset(r, 0, sizeof(*r));
|
|
|
|
r->overhead = conf->overhead;
|
2013-09-19 20:10:03 +04:00
|
|
|
r->rate_bytes_ps = max_t(u64, conf->rate, rate64);
|
2013-08-15 01:47:11 +04:00
|
|
|
r->linklayer = (conf->linklayer & TC_LINKLAYER_MASK);
|
2013-02-12 04:12:03 +04:00
|
|
|
r->mult = 1;
|
|
|
|
/*
|
2013-06-07 00:56:19 +04:00
|
|
|
* The deal here is to replace a divide by a reciprocal one
|
|
|
|
* in fast path (a reciprocal divide is a multiply and a shift)
|
|
|
|
*
|
|
|
|
* Normal formula would be :
|
|
|
|
* time_in_ns = (NSEC_PER_SEC * len) / rate_bps
|
|
|
|
*
|
|
|
|
* We compute mult/shift to use instead :
|
|
|
|
* time_in_ns = (len * mult) >> shift;
|
|
|
|
*
|
|
|
|
* We try to get the highest possible mult value for accuracy,
|
|
|
|
* but have to make sure no overflows will ever happen.
|
2013-02-12 04:12:03 +04:00
|
|
|
*/
|
2013-06-07 00:56:19 +04:00
|
|
|
if (r->rate_bytes_ps > 0) {
|
|
|
|
u64 factor = NSEC_PER_SEC;
|
|
|
|
|
|
|
|
for (;;) {
|
|
|
|
r->mult = div64_u64(factor, r->rate_bytes_ps);
|
|
|
|
if (r->mult & (1U << 31) || factor & (1ULL << 63))
|
2013-02-12 04:12:03 +04:00
|
|
|
break;
|
2013-06-07 00:56:19 +04:00
|
|
|
factor <<= 1;
|
|
|
|
r->shift++;
|
2013-02-12 04:12:03 +04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(psched_ratecfg_precompute);
|
2017-11-03 13:46:25 +03:00
|
|
|
|
|
|
|
static void mini_qdisc_rcu_func(struct rcu_head *head)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
void mini_qdisc_pair_swap(struct mini_Qdisc_pair *miniqp,
|
|
|
|
struct tcf_proto *tp_head)
|
|
|
|
{
|
|
|
|
struct mini_Qdisc *miniq_old = rtnl_dereference(*miniqp->p_miniq);
|
|
|
|
struct mini_Qdisc *miniq;
|
|
|
|
|
|
|
|
if (!tp_head) {
|
|
|
|
RCU_INIT_POINTER(*miniqp->p_miniq, NULL);
|
2017-12-21 10:26:24 +03:00
|
|
|
/* Wait for flying RCU callback before it is freed. */
|
|
|
|
rcu_barrier_bh();
|
2017-11-03 13:46:25 +03:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
miniq = !miniq_old || miniq_old == &miniqp->miniq2 ?
|
|
|
|
&miniqp->miniq1 : &miniqp->miniq2;
|
|
|
|
|
|
|
|
/* We need to make sure that readers won't see the miniq
|
|
|
|
* we are about to modify. So wait until previous call_rcu_bh callback
|
|
|
|
* is done.
|
|
|
|
*/
|
|
|
|
rcu_barrier_bh();
|
|
|
|
miniq->filter_list = tp_head;
|
|
|
|
rcu_assign_pointer(*miniqp->p_miniq, miniq);
|
|
|
|
|
|
|
|
if (miniq_old)
|
2017-12-21 10:26:24 +03:00
|
|
|
/* This is counterpart of the rcu barriers above. We need to
|
2017-11-03 13:46:25 +03:00
|
|
|
* block potential new user of miniq_old until all readers
|
|
|
|
* are not seeing it.
|
|
|
|
*/
|
|
|
|
call_rcu_bh(&miniq_old->rcu, mini_qdisc_rcu_func);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(mini_qdisc_pair_swap);
|
|
|
|
|
|
|
|
void mini_qdisc_pair_init(struct mini_Qdisc_pair *miniqp, struct Qdisc *qdisc,
|
|
|
|
struct mini_Qdisc __rcu **p_miniq)
|
|
|
|
{
|
|
|
|
miniqp->miniq1.cpu_bstats = qdisc->cpu_bstats;
|
|
|
|
miniqp->miniq1.cpu_qstats = qdisc->cpu_qstats;
|
|
|
|
miniqp->miniq2.cpu_bstats = qdisc->cpu_bstats;
|
|
|
|
miniqp->miniq2.cpu_qstats = qdisc->cpu_qstats;
|
|
|
|
miniqp->p_miniq = p_miniq;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(mini_qdisc_pair_init);
|