434 строки
17 KiB
Plaintext
434 строки
17 KiB
Plaintext
Linux Base Driver for 10 Gigabit Intel(R) Network Connection
|
||
=============================================================
|
||
|
||
October 9, 2007
|
||
|
||
|
||
Contents
|
||
========
|
||
|
||
- In This Release
|
||
- Identifying Your Adapter
|
||
- Building and Installation
|
||
- Command Line Parameters
|
||
- Improving Performance
|
||
- Additional Configurations
|
||
- Known Issues/Troubleshooting
|
||
- Support
|
||
|
||
|
||
|
||
In This Release
|
||
===============
|
||
|
||
This file describes the ixgb Linux Base Driver for the 10 Gigabit Intel(R)
|
||
Network Connection. This driver includes support for Itanium(R)2-based
|
||
systems.
|
||
|
||
For questions related to hardware requirements, refer to the documentation
|
||
supplied with your 10 Gigabit adapter. All hardware requirements listed apply
|
||
to use with Linux.
|
||
|
||
The following features are available in this kernel:
|
||
- Native VLANs
|
||
- Channel Bonding (teaming)
|
||
- SNMP
|
||
|
||
Channel Bonding documentation can be found in the Linux kernel source:
|
||
/Documentation/networking/bonding.txt
|
||
|
||
The driver information previously displayed in the /proc filesystem is not
|
||
supported in this release. Alternatively, you can use ethtool (version 1.6
|
||
or later), lspci, and ifconfig to obtain the same information.
|
||
|
||
Instructions on updating ethtool can be found in the section "Additional
|
||
Configurations" later in this document.
|
||
|
||
|
||
Identifying Your Adapter
|
||
========================
|
||
|
||
The following Intel network adapters are compatible with the drivers in this
|
||
release:
|
||
|
||
Controller Adapter Name Physical Layer
|
||
---------- ------------ --------------
|
||
82597EX Intel(R) PRO/10GbE LR/SR/CX4 10G Base-LR (1310 nm optical fiber)
|
||
Server Adapters 10G Base-SR (850 nm optical fiber)
|
||
10G Base-CX4(twin-axial copper cabling)
|
||
|
||
For more information on how to identify your adapter, go to the Adapter &
|
||
Driver ID Guide at:
|
||
|
||
http://support.intel.com/support/network/sb/CS-012904.htm
|
||
|
||
|
||
Building and Installation
|
||
=========================
|
||
|
||
select m for "Intel(R) PRO/10GbE support" located at:
|
||
Location:
|
||
-> Device Drivers
|
||
-> Network device support (NETDEVICES [=y])
|
||
-> Ethernet (10000 Mbit) (NETDEV_10000 [=y])
|
||
1. make modules && make modules_install
|
||
|
||
2. Load the module:
|
||
|
||
modprobe ixgb <parameter>=<value>
|
||
|
||
The insmod command can be used if the full
|
||
path to the driver module is specified. For example:
|
||
|
||
insmod /lib/modules/<KERNEL VERSION>/kernel/drivers/net/ixgb/ixgb.ko
|
||
|
||
With 2.6 based kernels also make sure that older ixgb drivers are
|
||
removed from the kernel, before loading the new module:
|
||
|
||
rmmod ixgb; modprobe ixgb
|
||
|
||
3. Assign an IP address to the interface by entering the following, where
|
||
x is the interface number:
|
||
|
||
ifconfig ethx <IP_address>
|
||
|
||
4. Verify that the interface works. Enter the following, where <IP_address>
|
||
is the IP address for another machine on the same subnet as the interface
|
||
that is being tested:
|
||
|
||
ping <IP_address>
|
||
|
||
|
||
Command Line Parameters
|
||
=======================
|
||
|
||
If the driver is built as a module, the following optional parameters are
|
||
used by entering them on the command line with the modprobe command using
|
||
this syntax:
|
||
|
||
modprobe ixgb [<option>=<VAL1>,<VAL2>,...]
|
||
|
||
For example, with two 10GbE PCI adapters, entering:
|
||
|
||
modprobe ixgb TxDescriptors=80,128
|
||
|
||
loads the ixgb driver with 80 TX resources for the first adapter and 128 TX
|
||
resources for the second adapter.
|
||
|
||
The default value for each parameter is generally the recommended setting,
|
||
unless otherwise noted.
|
||
|
||
FlowControl
|
||
Valid Range: 0-3 (0=none, 1=Rx only, 2=Tx only, 3=Rx&Tx)
|
||
Default: Read from the EEPROM
|
||
If EEPROM is not detected, default is 1
|
||
This parameter controls the automatic generation(Tx) and response(Rx) to
|
||
Ethernet PAUSE frames. There are hardware bugs associated with enabling
|
||
Tx flow control so beware.
|
||
|
||
RxDescriptors
|
||
Valid Range: 64-512
|
||
Default Value: 512
|
||
This value is the number of receive descriptors allocated by the driver.
|
||
Increasing this value allows the driver to buffer more incoming packets.
|
||
Each descriptor is 16 bytes. A receive buffer is also allocated for
|
||
each descriptor and can be either 2048, 4056, 8192, or 16384 bytes,
|
||
depending on the MTU setting. When the MTU size is 1500 or less, the
|
||
receive buffer size is 2048 bytes. When the MTU is greater than 1500 the
|
||
receive buffer size will be either 4056, 8192, or 16384 bytes. The
|
||
maximum MTU size is 16114.
|
||
|
||
RxIntDelay
|
||
Valid Range: 0-65535 (0=off)
|
||
Default Value: 72
|
||
This value delays the generation of receive interrupts in units of
|
||
0.8192 microseconds. Receive interrupt reduction can improve CPU
|
||
efficiency if properly tuned for specific network traffic. Increasing
|
||
this value adds extra latency to frame reception and can end up
|
||
decreasing the throughput of TCP traffic. If the system is reporting
|
||
dropped receives, this value may be set too high, causing the driver to
|
||
run out of available receive descriptors.
|
||
|
||
TxDescriptors
|
||
Valid Range: 64-4096
|
||
Default Value: 256
|
||
This value is the number of transmit descriptors allocated by the driver.
|
||
Increasing this value allows the driver to queue more transmits. Each
|
||
descriptor is 16 bytes.
|
||
|
||
XsumRX
|
||
Valid Range: 0-1
|
||
Default Value: 1
|
||
A value of '1' indicates that the driver should enable IP checksum
|
||
offload for received packets (both UDP and TCP) to the adapter hardware.
|
||
|
||
|
||
Improving Performance
|
||
=====================
|
||
|
||
With the 10 Gigabit server adapters, the default Linux configuration will
|
||
very likely limit the total available throughput artificially. There is a set
|
||
of configuration changes that, when applied together, will increase the ability
|
||
of Linux to transmit and receive data. The following enhancements were
|
||
originally acquired from settings published at http://www.spec.org/web99/ for
|
||
various submitted results using Linux.
|
||
|
||
NOTE: These changes are only suggestions, and serve as a starting point for
|
||
tuning your network performance.
|
||
|
||
The changes are made in three major ways, listed in order of greatest effect:
|
||
- Use ifconfig to modify the mtu (maximum transmission unit) and the txqueuelen
|
||
parameter.
|
||
- Use sysctl to modify /proc parameters (essentially kernel tuning)
|
||
- Use setpci to modify the MMRBC field in PCI-X configuration space to increase
|
||
transmit burst lengths on the bus.
|
||
|
||
NOTE: setpci modifies the adapter's configuration registers to allow it to read
|
||
up to 4k bytes at a time (for transmits). However, for some systems the
|
||
behavior after modifying this register may be undefined (possibly errors of
|
||
some kind). A power-cycle, hard reset or explicitly setting the e6 register
|
||
back to 22 (setpci -d 8086:1a48 e6.b=22) may be required to get back to a
|
||
stable configuration.
|
||
|
||
- COPY these lines and paste them into ixgb_perf.sh:
|
||
#!/bin/bash
|
||
echo "configuring network performance , edit this file to change the interface
|
||
or device ID of 10GbE card"
|
||
# set mmrbc to 4k reads, modify only Intel 10GbE device IDs
|
||
# replace 1a48 with appropriate 10GbE device's ID installed on the system,
|
||
# if needed.
|
||
setpci -d 8086:1a48 e6.b=2e
|
||
# set the MTU (max transmission unit) - it requires your switch and clients
|
||
# to change as well.
|
||
# set the txqueuelen
|
||
# your ixgb adapter should be loaded as eth1 for this to work, change if needed
|
||
ifconfig eth1 mtu 9000 txqueuelen 1000 up
|
||
# call the sysctl utility to modify /proc/sys entries
|
||
sysctl -p ./sysctl_ixgb.conf
|
||
- END ixgb_perf.sh
|
||
|
||
- COPY these lines and paste them into sysctl_ixgb.conf:
|
||
# some of the defaults may be different for your kernel
|
||
# call this file with sysctl -p <this file>
|
||
# these are just suggested values that worked well to increase throughput in
|
||
# several network benchmark tests, your mileage may vary
|
||
|
||
### IPV4 specific settings
|
||
# turn TCP timestamp support off, default 1, reduces CPU use
|
||
net.ipv4.tcp_timestamps = 0
|
||
# turn SACK support off, default on
|
||
# on systems with a VERY fast bus -> memory interface this is the big gainer
|
||
net.ipv4.tcp_sack = 0
|
||
# set min/default/max TCP read buffer, default 4096 87380 174760
|
||
net.ipv4.tcp_rmem = 10000000 10000000 10000000
|
||
# set min/pressure/max TCP write buffer, default 4096 16384 131072
|
||
net.ipv4.tcp_wmem = 10000000 10000000 10000000
|
||
# set min/pressure/max TCP buffer space, default 31744 32256 32768
|
||
net.ipv4.tcp_mem = 10000000 10000000 10000000
|
||
|
||
### CORE settings (mostly for socket and UDP effect)
|
||
# set maximum receive socket buffer size, default 131071
|
||
net.core.rmem_max = 524287
|
||
# set maximum send socket buffer size, default 131071
|
||
net.core.wmem_max = 524287
|
||
# set default receive socket buffer size, default 65535
|
||
net.core.rmem_default = 524287
|
||
# set default send socket buffer size, default 65535
|
||
net.core.wmem_default = 524287
|
||
# set maximum amount of option memory buffers, default 10240
|
||
net.core.optmem_max = 524287
|
||
# set number of unprocessed input packets before kernel starts dropping them; default 300
|
||
net.core.netdev_max_backlog = 300000
|
||
- END sysctl_ixgb.conf
|
||
|
||
Edit the ixgb_perf.sh script if necessary to change eth1 to whatever interface
|
||
your ixgb driver is using and/or replace '1a48' with appropriate 10GbE device's
|
||
ID installed on the system.
|
||
|
||
NOTE: Unless these scripts are added to the boot process, these changes will
|
||
only last only until the next system reboot.
|
||
|
||
|
||
Resolving Slow UDP Traffic
|
||
--------------------------
|
||
If your server does not seem to be able to receive UDP traffic as fast as it
|
||
can receive TCP traffic, it could be because Linux, by default, does not set
|
||
the network stack buffers as large as they need to be to support high UDP
|
||
transfer rates. One way to alleviate this problem is to allow more memory to
|
||
be used by the IP stack to store incoming data.
|
||
|
||
For instance, use the commands:
|
||
sysctl -w net.core.rmem_max=262143
|
||
and
|
||
sysctl -w net.core.rmem_default=262143
|
||
to increase the read buffer memory max and default to 262143 (256k - 1) from
|
||
defaults of max=131071 (128k - 1) and default=65535 (64k - 1). These variables
|
||
will increase the amount of memory used by the network stack for receives, and
|
||
can be increased significantly more if necessary for your application.
|
||
|
||
|
||
Additional Configurations
|
||
=========================
|
||
|
||
Configuring the Driver on Different Distributions
|
||
-------------------------------------------------
|
||
Configuring a network driver to load properly when the system is started is
|
||
distribution dependent. Typically, the configuration process involves adding
|
||
an alias line to /etc/modprobe.conf as well as editing other system startup
|
||
scripts and/or configuration files. Many popular Linux distributions ship
|
||
with tools to make these changes for you. To learn the proper way to
|
||
configure a network device for your system, refer to your distribution
|
||
documentation. If during this process you are asked for the driver or module
|
||
name, the name for the Linux Base Driver for the Intel 10GbE Family of
|
||
Adapters is ixgb.
|
||
|
||
Viewing Link Messages
|
||
---------------------
|
||
Link messages will not be displayed to the console if the distribution is
|
||
restricting system messages. In order to see network driver link messages on
|
||
your console, set dmesg to eight by entering the following:
|
||
|
||
dmesg -n 8
|
||
|
||
NOTE: This setting is not saved across reboots.
|
||
|
||
|
||
Jumbo Frames
|
||
------------
|
||
The driver supports Jumbo Frames for all adapters. Jumbo Frames support is
|
||
enabled by changing the MTU to a value larger than the default of 1500.
|
||
The maximum value for the MTU is 16114. Use the ifconfig command to
|
||
increase the MTU size. For example:
|
||
|
||
ifconfig ethx mtu 9000 up
|
||
|
||
The maximum MTU setting for Jumbo Frames is 16114. This value coincides
|
||
with the maximum Jumbo Frames size of 16128.
|
||
|
||
|
||
Ethtool
|
||
-------
|
||
The driver utilizes the ethtool interface for driver configuration and
|
||
diagnostics, as well as displaying statistical information. The ethtool
|
||
version 1.6 or later is required for this functionality.
|
||
|
||
The latest release of ethtool can be found from
|
||
http://ftp.kernel.org/pub/software/network/ethtool/
|
||
|
||
NOTE: The ethtool version 1.6 only supports a limited set of ethtool options.
|
||
Support for a more complete ethtool feature set can be enabled by
|
||
upgrading to the latest version.
|
||
|
||
|
||
NAPI
|
||
----
|
||
|
||
NAPI (Rx polling mode) is supported in the ixgb driver. NAPI is enabled
|
||
or disabled based on the configuration of the kernel. see CONFIG_IXGB_NAPI
|
||
|
||
See www.cyberus.ca/~hadi/usenix-paper.tgz for more information on NAPI.
|
||
|
||
|
||
Known Issues/Troubleshooting
|
||
============================
|
||
|
||
NOTE: After installing the driver, if your Intel Network Connection is not
|
||
working, verify in the "In This Release" section of the readme that you have
|
||
installed the correct driver.
|
||
|
||
Intel(R) PRO/10GbE CX4 Server Adapter Cable Interoperability Issue with
|
||
Fujitsu XENPAK Module in SmartBits Chassis
|
||
---------------------------------------------------------------------
|
||
Excessive CRC errors may be observed if the Intel(R) PRO/10GbE CX4
|
||
Server adapter is connected to a Fujitsu XENPAK CX4 module in a SmartBits
|
||
chassis using 15 m/24AWG cable assemblies manufactured by Fujitsu or Leoni.
|
||
The CRC errors may be received either by the Intel(R) PRO/10GbE CX4
|
||
Server adapter or the SmartBits. If this situation occurs using a different
|
||
cable assembly may resolve the issue.
|
||
|
||
CX4 Server Adapter Cable Interoperability Issues with HP Procurve 3400cl
|
||
Switch Port
|
||
------------------------------------------------------------------------
|
||
Excessive CRC errors may be observed if the Intel(R) PRO/10GbE CX4 Server
|
||
adapter is connected to an HP Procurve 3400cl switch port using short cables
|
||
(1 m or shorter). If this situation occurs, using a longer cable may resolve
|
||
the issue.
|
||
|
||
Excessive CRC errors may be observed using Fujitsu 24AWG cable assemblies that
|
||
Are 10 m or longer or where using a Leoni 15 m/24AWG cable assembly. The CRC
|
||
errors may be received either by the CX4 Server adapter or at the switch. If
|
||
this situation occurs, using a different cable assembly may resolve the issue.
|
||
|
||
|
||
Jumbo Frames System Requirement
|
||
-------------------------------
|
||
Memory allocation failures have been observed on Linux systems with 64 MB
|
||
of RAM or less that are running Jumbo Frames. If you are using Jumbo
|
||
Frames, your system may require more than the advertised minimum
|
||
requirement of 64 MB of system memory.
|
||
|
||
|
||
Performance Degradation with Jumbo Frames
|
||
-----------------------------------------
|
||
Degradation in throughput performance may be observed in some Jumbo frames
|
||
environments. If this is observed, increasing the application's socket buffer
|
||
size and/or increasing the /proc/sys/net/ipv4/tcp_*mem entry values may help.
|
||
See the specific application manual and /usr/src/linux*/Documentation/
|
||
networking/ip-sysctl.txt for more details.
|
||
|
||
|
||
Allocating Rx Buffers when Using Jumbo Frames
|
||
---------------------------------------------
|
||
Allocating Rx buffers when using Jumbo Frames on 2.6.x kernels may fail if
|
||
the available memory is heavily fragmented. This issue may be seen with PCI-X
|
||
adapters or with packet split disabled. This can be reduced or eliminated
|
||
by changing the amount of available memory for receive buffer allocation, by
|
||
increasing /proc/sys/vm/min_free_kbytes.
|
||
|
||
|
||
Multiple Interfaces on Same Ethernet Broadcast Network
|
||
------------------------------------------------------
|
||
Due to the default ARP behavior on Linux, it is not possible to have
|
||
one system on two IP networks in the same Ethernet broadcast domain
|
||
(non-partitioned switch) behave as expected. All Ethernet interfaces
|
||
will respond to IP traffic for any IP address assigned to the system.
|
||
This results in unbalanced receive traffic.
|
||
|
||
If you have multiple interfaces in a server, do either of the following:
|
||
|
||
- Turn on ARP filtering by entering:
|
||
echo 1 > /proc/sys/net/ipv4/conf/all/arp_filter
|
||
|
||
- Install the interfaces in separate broadcast domains - either in
|
||
different switches or in a switch partitioned to VLANs.
|
||
|
||
|
||
UDP Stress Test Dropped Packet Issue
|
||
--------------------------------------
|
||
Under small packets UDP stress test with 10GbE driver, the Linux system
|
||
may drop UDP packets due to the fullness of socket buffers. You may want
|
||
to change the driver's Flow Control variables to the minimum value for
|
||
controlling packet reception.
|
||
|
||
|
||
Tx Hangs Possible Under Stress
|
||
------------------------------
|
||
Under stress conditions, if TX hangs occur, turning off TSO
|
||
"ethtool -K eth0 tso off" may resolve the problem.
|
||
|
||
|
||
Support
|
||
=======
|
||
|
||
For general information, go to the Intel support website at:
|
||
|
||
http://support.intel.com
|
||
|
||
or the Intel Wired Networking project hosted by Sourceforge at:
|
||
|
||
http://sourceforge.net/projects/e1000
|
||
|
||
If an issue is identified with the released source code on the supported
|
||
kernel with a supported adapter, email the specific information related
|
||
to the issue to e1000-devel@lists.sf.net
|