2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
|
|
|
|
HOWTO for the linux packet generator
|
|
|
|
------------------------------------
|
|
|
|
|
2015-02-24 10:31:52 +08:00
|
|
|
Enable CONFIG_NET_PKTGEN to compile and build pktgen.o either in-kernel
|
|
|
|
or as a module. A module is preferred; insmod pktgen if needed. Once
|
|
|
|
running, pktgen creates a thread for each CPU with affinity to that CPU.
|
|
|
|
Monitoring and controlling is done via /proc. It is easiest to select a
|
|
|
|
suitable sample script and configure that.
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
On a dual CPU:
|
|
|
|
|
|
|
|
ps aux | grep pkt
|
|
|
|
root 129 0.3 0.0 0 0 ? SW 2003 523:20 [pktgen/0]
|
|
|
|
root 130 0.3 0.0 0 0 ? SW 2003 509:50 [pktgen/1]
|
|
|
|
|
|
|
|
|
2006-10-04 04:50:39 +08:00
|
|
|
For monitoring and control pktgen creates:
|
2005-04-17 06:20:36 +08:00
|
|
|
/proc/net/pktgen/pgctrl
|
|
|
|
/proc/net/pktgen/kpktgend_X
|
|
|
|
/proc/net/pktgen/ethX
|
|
|
|
|
|
|
|
|
pktgen: document tuning for max NIC performance
Using pktgen I'm seeing the ixgbe driver "push-back", due TX ring
running full. Thus, the TX ring is artificially limiting pktgen.
(Diagnose via "ethtool -S", look for "tx_restart_queue" or "tx_busy"
counters.)
Using ixgbe, the real reason behind the TX ring running full, is due
to TX ring not being cleaned up fast enough. The ixgbe driver combines
TX+RX ring cleanups, and the cleanup interval is affected by the
ethtool --coalesce setting of parameter "rx-usecs".
Do not increase the default NIC TX ring buffer or default cleanup
interval. Instead simply document that pktgen needs special NIC
tuning for maximum packet per sec performance.
Performance results with pktgen with clone_skb=100000.
TX ring size 512 (default), adjusting "rx-usecs":
(Single CPU performance, E5-2630, ixgbe)
- 3935002 pps - rx-usecs: 1 (irqs: 9346)
- 5132350 pps - rx-usecs: 10 (irqs: 99157)
- 5375111 pps - rx-usecs: 20 (irqs: 50154)
- 5454050 pps - rx-usecs: 30 (irqs: 33872)
- 5496320 pps - rx-usecs: 40 (irqs: 26197)
- 5502510 pps - rx-usecs: 50 (irqs: 21527)
TX ring size adjusting (ethtool -G), "rx-usecs==1" (default):
- 3935002 pps - tx-size: 512
- 5354401 pps - tx-size: 768
- 5356847 pps - tx-size: 1024
- 5327595 pps - tx-size: 1536
- 5356779 pps - tx-size: 2048
- 5353438 pps - tx-size: 4096
Notice after commit 6f25cd47d (pktgen: fix xmit test for BQL enabled
devices) pktgen uses netif_xmit_frozen_or_drv_stopped() and ignores
the BQL "stack" pause (QUEUE_STATE_STACK_XOFF) flag. This allow us to put
more pressure on the TX ring buffers.
It is the ixgbe_maybe_stop_tx() call that stops the transmits, and
pktgen respecting this in the call to netif_xmit_frozen_or_drv_stopped(txq).
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-06-26 19:16:27 +08:00
|
|
|
Tuning NIC for max performance
|
|
|
|
==============================
|
|
|
|
|
2015-02-24 10:31:52 +08:00
|
|
|
The default NIC settings are (likely) not tuned for pktgen's artificial
|
pktgen: document tuning for max NIC performance
Using pktgen I'm seeing the ixgbe driver "push-back", due TX ring
running full. Thus, the TX ring is artificially limiting pktgen.
(Diagnose via "ethtool -S", look for "tx_restart_queue" or "tx_busy"
counters.)
Using ixgbe, the real reason behind the TX ring running full, is due
to TX ring not being cleaned up fast enough. The ixgbe driver combines
TX+RX ring cleanups, and the cleanup interval is affected by the
ethtool --coalesce setting of parameter "rx-usecs".
Do not increase the default NIC TX ring buffer or default cleanup
interval. Instead simply document that pktgen needs special NIC
tuning for maximum packet per sec performance.
Performance results with pktgen with clone_skb=100000.
TX ring size 512 (default), adjusting "rx-usecs":
(Single CPU performance, E5-2630, ixgbe)
- 3935002 pps - rx-usecs: 1 (irqs: 9346)
- 5132350 pps - rx-usecs: 10 (irqs: 99157)
- 5375111 pps - rx-usecs: 20 (irqs: 50154)
- 5454050 pps - rx-usecs: 30 (irqs: 33872)
- 5496320 pps - rx-usecs: 40 (irqs: 26197)
- 5502510 pps - rx-usecs: 50 (irqs: 21527)
TX ring size adjusting (ethtool -G), "rx-usecs==1" (default):
- 3935002 pps - tx-size: 512
- 5354401 pps - tx-size: 768
- 5356847 pps - tx-size: 1024
- 5327595 pps - tx-size: 1536
- 5356779 pps - tx-size: 2048
- 5353438 pps - tx-size: 4096
Notice after commit 6f25cd47d (pktgen: fix xmit test for BQL enabled
devices) pktgen uses netif_xmit_frozen_or_drv_stopped() and ignores
the BQL "stack" pause (QUEUE_STATE_STACK_XOFF) flag. This allow us to put
more pressure on the TX ring buffers.
It is the ixgbe_maybe_stop_tx() call that stops the transmits, and
pktgen respecting this in the call to netif_xmit_frozen_or_drv_stopped(txq).
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-06-26 19:16:27 +08:00
|
|
|
overload type of benchmarking, as this could hurt the normal use-case.
|
|
|
|
|
|
|
|
Specifically increasing the TX ring buffer in the NIC:
|
|
|
|
# ethtool -G ethX tx 1024
|
|
|
|
|
|
|
|
A larger TX ring can improve pktgen's performance, while it can hurt
|
|
|
|
in the general case, 1) because the TX ring buffer might get larger
|
2015-02-24 10:31:52 +08:00
|
|
|
than the CPU's L1/L2 cache, 2) because it allows more queueing in the
|
pktgen: document tuning for max NIC performance
Using pktgen I'm seeing the ixgbe driver "push-back", due TX ring
running full. Thus, the TX ring is artificially limiting pktgen.
(Diagnose via "ethtool -S", look for "tx_restart_queue" or "tx_busy"
counters.)
Using ixgbe, the real reason behind the TX ring running full, is due
to TX ring not being cleaned up fast enough. The ixgbe driver combines
TX+RX ring cleanups, and the cleanup interval is affected by the
ethtool --coalesce setting of parameter "rx-usecs".
Do not increase the default NIC TX ring buffer or default cleanup
interval. Instead simply document that pktgen needs special NIC
tuning for maximum packet per sec performance.
Performance results with pktgen with clone_skb=100000.
TX ring size 512 (default), adjusting "rx-usecs":
(Single CPU performance, E5-2630, ixgbe)
- 3935002 pps - rx-usecs: 1 (irqs: 9346)
- 5132350 pps - rx-usecs: 10 (irqs: 99157)
- 5375111 pps - rx-usecs: 20 (irqs: 50154)
- 5454050 pps - rx-usecs: 30 (irqs: 33872)
- 5496320 pps - rx-usecs: 40 (irqs: 26197)
- 5502510 pps - rx-usecs: 50 (irqs: 21527)
TX ring size adjusting (ethtool -G), "rx-usecs==1" (default):
- 3935002 pps - tx-size: 512
- 5354401 pps - tx-size: 768
- 5356847 pps - tx-size: 1024
- 5327595 pps - tx-size: 1536
- 5356779 pps - tx-size: 2048
- 5353438 pps - tx-size: 4096
Notice after commit 6f25cd47d (pktgen: fix xmit test for BQL enabled
devices) pktgen uses netif_xmit_frozen_or_drv_stopped() and ignores
the BQL "stack" pause (QUEUE_STATE_STACK_XOFF) flag. This allow us to put
more pressure on the TX ring buffers.
It is the ixgbe_maybe_stop_tx() call that stops the transmits, and
pktgen respecting this in the call to netif_xmit_frozen_or_drv_stopped(txq).
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-06-26 19:16:27 +08:00
|
|
|
NIC HW layer (which is bad for bufferbloat).
|
|
|
|
|
2015-02-24 10:31:52 +08:00
|
|
|
One should hesitate to conclude that packets/descriptors in the HW
|
pktgen: document tuning for max NIC performance
Using pktgen I'm seeing the ixgbe driver "push-back", due TX ring
running full. Thus, the TX ring is artificially limiting pktgen.
(Diagnose via "ethtool -S", look for "tx_restart_queue" or "tx_busy"
counters.)
Using ixgbe, the real reason behind the TX ring running full, is due
to TX ring not being cleaned up fast enough. The ixgbe driver combines
TX+RX ring cleanups, and the cleanup interval is affected by the
ethtool --coalesce setting of parameter "rx-usecs".
Do not increase the default NIC TX ring buffer or default cleanup
interval. Instead simply document that pktgen needs special NIC
tuning for maximum packet per sec performance.
Performance results with pktgen with clone_skb=100000.
TX ring size 512 (default), adjusting "rx-usecs":
(Single CPU performance, E5-2630, ixgbe)
- 3935002 pps - rx-usecs: 1 (irqs: 9346)
- 5132350 pps - rx-usecs: 10 (irqs: 99157)
- 5375111 pps - rx-usecs: 20 (irqs: 50154)
- 5454050 pps - rx-usecs: 30 (irqs: 33872)
- 5496320 pps - rx-usecs: 40 (irqs: 26197)
- 5502510 pps - rx-usecs: 50 (irqs: 21527)
TX ring size adjusting (ethtool -G), "rx-usecs==1" (default):
- 3935002 pps - tx-size: 512
- 5354401 pps - tx-size: 768
- 5356847 pps - tx-size: 1024
- 5327595 pps - tx-size: 1536
- 5356779 pps - tx-size: 2048
- 5353438 pps - tx-size: 4096
Notice after commit 6f25cd47d (pktgen: fix xmit test for BQL enabled
devices) pktgen uses netif_xmit_frozen_or_drv_stopped() and ignores
the BQL "stack" pause (QUEUE_STATE_STACK_XOFF) flag. This allow us to put
more pressure on the TX ring buffers.
It is the ixgbe_maybe_stop_tx() call that stops the transmits, and
pktgen respecting this in the call to netif_xmit_frozen_or_drv_stopped(txq).
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-06-26 19:16:27 +08:00
|
|
|
TX ring cause delay. Drivers usually delay cleaning up the
|
2015-02-24 10:31:52 +08:00
|
|
|
ring-buffers for various performance reasons, and packets stalling
|
|
|
|
the TX ring might just be waiting for cleanup.
|
pktgen: document tuning for max NIC performance
Using pktgen I'm seeing the ixgbe driver "push-back", due TX ring
running full. Thus, the TX ring is artificially limiting pktgen.
(Diagnose via "ethtool -S", look for "tx_restart_queue" or "tx_busy"
counters.)
Using ixgbe, the real reason behind the TX ring running full, is due
to TX ring not being cleaned up fast enough. The ixgbe driver combines
TX+RX ring cleanups, and the cleanup interval is affected by the
ethtool --coalesce setting of parameter "rx-usecs".
Do not increase the default NIC TX ring buffer or default cleanup
interval. Instead simply document that pktgen needs special NIC
tuning for maximum packet per sec performance.
Performance results with pktgen with clone_skb=100000.
TX ring size 512 (default), adjusting "rx-usecs":
(Single CPU performance, E5-2630, ixgbe)
- 3935002 pps - rx-usecs: 1 (irqs: 9346)
- 5132350 pps - rx-usecs: 10 (irqs: 99157)
- 5375111 pps - rx-usecs: 20 (irqs: 50154)
- 5454050 pps - rx-usecs: 30 (irqs: 33872)
- 5496320 pps - rx-usecs: 40 (irqs: 26197)
- 5502510 pps - rx-usecs: 50 (irqs: 21527)
TX ring size adjusting (ethtool -G), "rx-usecs==1" (default):
- 3935002 pps - tx-size: 512
- 5354401 pps - tx-size: 768
- 5356847 pps - tx-size: 1024
- 5327595 pps - tx-size: 1536
- 5356779 pps - tx-size: 2048
- 5353438 pps - tx-size: 4096
Notice after commit 6f25cd47d (pktgen: fix xmit test for BQL enabled
devices) pktgen uses netif_xmit_frozen_or_drv_stopped() and ignores
the BQL "stack" pause (QUEUE_STATE_STACK_XOFF) flag. This allow us to put
more pressure on the TX ring buffers.
It is the ixgbe_maybe_stop_tx() call that stops the transmits, and
pktgen respecting this in the call to netif_xmit_frozen_or_drv_stopped(txq).
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-06-26 19:16:27 +08:00
|
|
|
|
2015-02-24 10:31:52 +08:00
|
|
|
This cleanup issue is specifically the case for the driver ixgbe
|
|
|
|
(Intel 82599 chip). This driver (ixgbe) combines TX+RX ring cleanups,
|
pktgen: document tuning for max NIC performance
Using pktgen I'm seeing the ixgbe driver "push-back", due TX ring
running full. Thus, the TX ring is artificially limiting pktgen.
(Diagnose via "ethtool -S", look for "tx_restart_queue" or "tx_busy"
counters.)
Using ixgbe, the real reason behind the TX ring running full, is due
to TX ring not being cleaned up fast enough. The ixgbe driver combines
TX+RX ring cleanups, and the cleanup interval is affected by the
ethtool --coalesce setting of parameter "rx-usecs".
Do not increase the default NIC TX ring buffer or default cleanup
interval. Instead simply document that pktgen needs special NIC
tuning for maximum packet per sec performance.
Performance results with pktgen with clone_skb=100000.
TX ring size 512 (default), adjusting "rx-usecs":
(Single CPU performance, E5-2630, ixgbe)
- 3935002 pps - rx-usecs: 1 (irqs: 9346)
- 5132350 pps - rx-usecs: 10 (irqs: 99157)
- 5375111 pps - rx-usecs: 20 (irqs: 50154)
- 5454050 pps - rx-usecs: 30 (irqs: 33872)
- 5496320 pps - rx-usecs: 40 (irqs: 26197)
- 5502510 pps - rx-usecs: 50 (irqs: 21527)
TX ring size adjusting (ethtool -G), "rx-usecs==1" (default):
- 3935002 pps - tx-size: 512
- 5354401 pps - tx-size: 768
- 5356847 pps - tx-size: 1024
- 5327595 pps - tx-size: 1536
- 5356779 pps - tx-size: 2048
- 5353438 pps - tx-size: 4096
Notice after commit 6f25cd47d (pktgen: fix xmit test for BQL enabled
devices) pktgen uses netif_xmit_frozen_or_drv_stopped() and ignores
the BQL "stack" pause (QUEUE_STATE_STACK_XOFF) flag. This allow us to put
more pressure on the TX ring buffers.
It is the ixgbe_maybe_stop_tx() call that stops the transmits, and
pktgen respecting this in the call to netif_xmit_frozen_or_drv_stopped(txq).
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-06-26 19:16:27 +08:00
|
|
|
and the cleanup interval is affected by the ethtool --coalesce setting
|
|
|
|
of parameter "rx-usecs".
|
|
|
|
|
2015-02-24 10:31:52 +08:00
|
|
|
For ixgbe use e.g. "30" resulting in approx 33K interrupts/sec (1/30*10^6):
|
pktgen: document tuning for max NIC performance
Using pktgen I'm seeing the ixgbe driver "push-back", due TX ring
running full. Thus, the TX ring is artificially limiting pktgen.
(Diagnose via "ethtool -S", look for "tx_restart_queue" or "tx_busy"
counters.)
Using ixgbe, the real reason behind the TX ring running full, is due
to TX ring not being cleaned up fast enough. The ixgbe driver combines
TX+RX ring cleanups, and the cleanup interval is affected by the
ethtool --coalesce setting of parameter "rx-usecs".
Do not increase the default NIC TX ring buffer or default cleanup
interval. Instead simply document that pktgen needs special NIC
tuning for maximum packet per sec performance.
Performance results with pktgen with clone_skb=100000.
TX ring size 512 (default), adjusting "rx-usecs":
(Single CPU performance, E5-2630, ixgbe)
- 3935002 pps - rx-usecs: 1 (irqs: 9346)
- 5132350 pps - rx-usecs: 10 (irqs: 99157)
- 5375111 pps - rx-usecs: 20 (irqs: 50154)
- 5454050 pps - rx-usecs: 30 (irqs: 33872)
- 5496320 pps - rx-usecs: 40 (irqs: 26197)
- 5502510 pps - rx-usecs: 50 (irqs: 21527)
TX ring size adjusting (ethtool -G), "rx-usecs==1" (default):
- 3935002 pps - tx-size: 512
- 5354401 pps - tx-size: 768
- 5356847 pps - tx-size: 1024
- 5327595 pps - tx-size: 1536
- 5356779 pps - tx-size: 2048
- 5353438 pps - tx-size: 4096
Notice after commit 6f25cd47d (pktgen: fix xmit test for BQL enabled
devices) pktgen uses netif_xmit_frozen_or_drv_stopped() and ignores
the BQL "stack" pause (QUEUE_STATE_STACK_XOFF) flag. This allow us to put
more pressure on the TX ring buffers.
It is the ixgbe_maybe_stop_tx() call that stops the transmits, and
pktgen respecting this in the call to netif_xmit_frozen_or_drv_stopped(txq).
Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2014-06-26 19:16:27 +08:00
|
|
|
# ethtool -C ethX rx-usecs 30
|
|
|
|
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
Viewing threads
|
|
|
|
===============
|
|
|
|
/proc/net/pktgen/kpktgend_0
|
|
|
|
Name: kpktgend_0 max_before_softirq: 10000
|
|
|
|
Running:
|
|
|
|
Stopped: eth1
|
|
|
|
Result: OK: max_before_softirq=10000
|
|
|
|
|
2015-02-24 10:31:52 +08:00
|
|
|
Most important are the devices assigned to the thread. Note that a
|
|
|
|
device can only belong to one thread.
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
|
|
|
|
Viewing devices
|
|
|
|
===============
|
|
|
|
|
2015-02-24 10:31:52 +08:00
|
|
|
The Params section holds configured information. The Current section
|
|
|
|
holds running statistics. The Result is printed after a run or after
|
|
|
|
interruption. Example:
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
/proc/net/pktgen/eth1
|
|
|
|
|
|
|
|
Params: count 10000000 min_pkt_size: 60 max_pkt_size: 60
|
|
|
|
frags: 0 delay: 0 clone_skb: 1000000 ifname: eth1
|
|
|
|
flows: 0 flowlen: 0
|
|
|
|
dst_min: 10.10.11.2 dst_max:
|
|
|
|
src_min: src_max:
|
|
|
|
src_mac: 00:00:00:00:00:00 dst_mac: 00:04:23:AC:FD:82
|
|
|
|
udp_src_min: 9 udp_src_max: 9 udp_dst_min: 9 udp_dst_max: 9
|
|
|
|
src_mac_count: 0 dst_mac_count: 0
|
|
|
|
Flags:
|
|
|
|
Current:
|
|
|
|
pkts-sofar: 10000000 errors: 39664
|
|
|
|
started: 1103053986245187us stopped: 1103053999346329us idle: 880401us
|
|
|
|
seq_num: 10000011 cur_dst_mac_offset: 0 cur_src_mac_offset: 0
|
|
|
|
cur_saddr: 0x10a0a0a cur_daddr: 0x20b0a0a
|
|
|
|
cur_udp_dst: 9 cur_udp_src: 9
|
|
|
|
flows: 0
|
|
|
|
Result: OK: 13101142(c12220741+d880401) usec, 10000000 (60byte,0frags)
|
|
|
|
763292pps 390Mb/sec (390805504bps) errors: 39664
|
|
|
|
|
2006-11-30 12:21:10 +08:00
|
|
|
Configuring threads and devices
|
|
|
|
================================
|
2015-02-24 10:31:52 +08:00
|
|
|
This is done via the /proc interface, and most easily done via pgset in
|
|
|
|
the scripts.
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
Examples:
|
|
|
|
|
|
|
|
pgset "clone_skb 1" sets the number of copies of the same packet
|
|
|
|
pgset "clone_skb 0" use single SKB for all transmits
|
2014-10-01 08:53:21 +08:00
|
|
|
pgset "burst 8" uses xmit_more API to queue 8 copies of the same
|
|
|
|
packet and update HW tx queue tail pointer once.
|
|
|
|
"burst 1" is the default
|
2005-04-17 06:20:36 +08:00
|
|
|
pgset "pkt_size 9014" sets packet size to 9014
|
|
|
|
pgset "frags 5" packet will consist of 5 fragments
|
|
|
|
pgset "count 200000" sets number of packets to send, set to zero
|
2006-07-01 00:28:43 +08:00
|
|
|
for continuous sends until explicitly stopped.
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
pgset "delay 5000" adds delay to hard_start_xmit(). nanoseconds
|
|
|
|
|
|
|
|
pgset "dst 10.0.0.1" sets IP destination address
|
|
|
|
(BEWARE! This generator is very aggressive!)
|
|
|
|
|
|
|
|
pgset "dst_min 10.0.0.1" Same as dst
|
|
|
|
pgset "dst_max 10.0.0.254" Set the maximum destination IP.
|
|
|
|
pgset "src_min 10.0.0.1" Set the minimum (or only) source IP.
|
|
|
|
pgset "src_max 10.0.0.254" Set the maximum source IP.
|
|
|
|
pgset "dst6 fec0::1" IPV6 destination address
|
|
|
|
pgset "src6 fec0::2" IPV6 source address
|
|
|
|
pgset "dstmac 00:00:00:00:00:00" sets MAC destination address
|
|
|
|
pgset "srcmac 00:00:00:00:00:00" sets MAC source address
|
|
|
|
|
2009-10-03 04:24:59 +08:00
|
|
|
pgset "queue_map_min 0" Sets the min value of tx queue interval
|
|
|
|
pgset "queue_map_max 7" Sets the max value of tx queue interval, for multiqueue devices
|
|
|
|
To select queue 1 of a given device,
|
|
|
|
use queue_map_min=1 and queue_map_max=1
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
pgset "src_mac_count 1" Sets the number of MACs we'll range through.
|
|
|
|
The 'minimum' MAC is what you set with srcmac.
|
|
|
|
|
|
|
|
pgset "dst_mac_count 1" Sets the number of MACs we'll range through.
|
|
|
|
The 'minimum' MAC is what you set with dstmac.
|
|
|
|
|
|
|
|
pgset "flag [name]" Set a flag to determine behaviour. Current flags
|
2014-02-22 04:38:36 +08:00
|
|
|
are: IPSRC_RND # IP source is random (between min/max)
|
|
|
|
IPDST_RND # IP destination is random
|
|
|
|
UDPSRC_RND, UDPDST_RND,
|
|
|
|
MACSRC_RND, MACDST_RND
|
|
|
|
TXSIZE_RND, IPV6,
|
2006-09-28 07:33:05 +08:00
|
|
|
MPLS_RND, VID_RND, SVID_RND
|
2014-02-22 04:38:36 +08:00
|
|
|
FLOW_SEQ,
|
2009-10-03 04:24:59 +08:00
|
|
|
QUEUE_MAP_RND # queue map random
|
|
|
|
QUEUE_MAP_CPU # queue map mirrors smp_processor_id()
|
2014-02-22 04:38:36 +08:00
|
|
|
UDPCSUM,
|
|
|
|
IPSEC # IPsec encapsulation (needs CONFIG_XFRM)
|
|
|
|
NODE_ALLOC # node specific memory allocation
|
2009-10-03 04:24:59 +08:00
|
|
|
|
2014-01-03 11:18:34 +08:00
|
|
|
pgset spi SPI_VALUE Set specific SA used to transform packet.
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
pgset "udp_src_min 9" set UDP source port min, If < udp_src_max, then
|
|
|
|
cycle through the port range.
|
|
|
|
|
|
|
|
pgset "udp_src_max 9" set UDP source port max.
|
|
|
|
pgset "udp_dst_min 9" set UDP destination port min, If < udp_dst_max, then
|
|
|
|
cycle through the port range.
|
|
|
|
pgset "udp_dst_max 9" set UDP destination port max.
|
|
|
|
|
2006-03-23 17:10:26 +08:00
|
|
|
pgset "mpls 0001000a,0002000a,0000000a" set MPLS labels (in this example
|
|
|
|
outer label=16,middle label=32,
|
|
|
|
inner label=0 (IPv4 NULL)) Note that
|
|
|
|
there must be no spaces between the
|
|
|
|
arguments. Leading zeros are required.
|
|
|
|
Do not set the bottom of stack bit,
|
2006-11-30 11:55:36 +08:00
|
|
|
that's done automatically. If you do
|
2006-03-23 17:10:26 +08:00
|
|
|
set the bottom of stack bit, that
|
|
|
|
indicates that you want to randomly
|
|
|
|
generate that address and the flag
|
|
|
|
MPLS_RND will be turned on. You
|
|
|
|
can have any mix of random and fixed
|
|
|
|
labels in the label stack.
|
|
|
|
|
|
|
|
pgset "mpls 0" turn off mpls (or any invalid argument works too!)
|
|
|
|
|
2006-09-28 07:33:05 +08:00
|
|
|
pgset "vlan_id 77" set VLAN ID 0-4095
|
|
|
|
pgset "vlan_p 3" set priority bit 0-7 (default 0)
|
|
|
|
pgset "vlan_cfi 0" set canonical format identifier 0-1 (default 0)
|
|
|
|
|
|
|
|
pgset "svlan_id 22" set SVLAN ID 0-4095
|
|
|
|
pgset "svlan_p 3" set priority bit 0-7 (default 0)
|
|
|
|
pgset "svlan_cfi 0" set canonical format identifier 0-1 (default 0)
|
|
|
|
|
|
|
|
pgset "vlan_id 9999" > 4095 remove vlan and svlan tags
|
|
|
|
pgset "svlan 9999" > 4095 remove svlan tag
|
|
|
|
|
|
|
|
|
|
|
|
pgset "tos XX" set former IPv4 TOS field (e.g. "tos 28" for AF11 no ECN, default 00)
|
|
|
|
pgset "traffic_class XX" set former IPv6 TRAFFIC CLASS (e.g. "traffic_class B8" for EF no ECN, default 00)
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
pgset stop aborts injection. Also, ^C aborts generator.
|
|
|
|
|
2010-06-10 06:49:57 +08:00
|
|
|
pgset "rate 300M" set rate to 300 Mb/s
|
|
|
|
pgset "ratep 1000000" set rate to 1Mpps
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
Example scripts
|
|
|
|
===============
|
|
|
|
|
2015-02-24 10:31:52 +08:00
|
|
|
A collection of small tutorial scripts for pktgen is in the examples
|
|
|
|
directory:
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
pktgen.conf-1-1 # 1 CPU 1 dev
|
|
|
|
pktgen.conf-1-2 # 1 CPU 2 dev
|
|
|
|
pktgen.conf-2-1 # 2 CPU's 1 dev
|
|
|
|
pktgen.conf-2-2 # 2 CPU's 2 dev
|
|
|
|
pktgen.conf-1-1-rdos # 1 CPU 1 dev w. route DoS
|
|
|
|
pktgen.conf-1-1-ip6 # 1 CPU 1 dev ipv6
|
|
|
|
pktgen.conf-1-1-ip6-rdos # 1 CPU 1 dev ipv6 w. route DoS
|
|
|
|
pktgen.conf-1-1-flows # 1 CPU 1 dev multiple flows.
|
|
|
|
|
2015-02-24 10:31:52 +08:00
|
|
|
Run in shell: ./pktgen.conf-X-Y
|
|
|
|
This does all the setup including sending.
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
|
|
|
|
Interrupt affinity
|
|
|
|
===================
|
2015-02-24 10:31:52 +08:00
|
|
|
Note that when adding devices to a specific CPU it is a good idea to
|
|
|
|
also assign /proc/irq/XX/smp_affinity so that the TX interrupts are bound
|
|
|
|
to the same CPU. This reduces cache bouncing when freeing skbs.
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2014-01-03 11:18:34 +08:00
|
|
|
Enable IPsec
|
|
|
|
============
|
2015-02-24 10:31:52 +08:00
|
|
|
Default IPsec transformation with ESP encapsulation plus transport mode
|
|
|
|
can be enabled by simply setting:
|
2014-01-03 11:18:34 +08:00
|
|
|
|
|
|
|
pgset "flag IPSEC"
|
|
|
|
pgset "flows 1"
|
|
|
|
|
|
|
|
To avoid breaking existing testbed scripts for using AH type and tunnel mode,
|
2015-02-24 10:31:52 +08:00
|
|
|
you can use "pgset spi SPI_VALUE" to specify which transformation mode
|
2014-01-03 11:18:34 +08:00
|
|
|
to employ.
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
Current commands and configuration options
|
|
|
|
==========================================
|
|
|
|
|
|
|
|
** Pgcontrol commands:
|
|
|
|
|
|
|
|
start
|
|
|
|
stop
|
|
|
|
|
|
|
|
** Thread commands:
|
|
|
|
|
|
|
|
add_device
|
|
|
|
rem_device_all
|
|
|
|
max_before_softirq
|
|
|
|
|
|
|
|
|
|
|
|
** Device commands:
|
|
|
|
|
|
|
|
count
|
|
|
|
clone_skb
|
|
|
|
debug
|
|
|
|
|
|
|
|
frags
|
|
|
|
delay
|
|
|
|
|
|
|
|
src_mac_count
|
|
|
|
dst_mac_count
|
|
|
|
|
|
|
|
pkt_size
|
|
|
|
min_pkt_size
|
|
|
|
max_pkt_size
|
|
|
|
|
2006-03-23 17:10:26 +08:00
|
|
|
mpls
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
udp_src_min
|
|
|
|
udp_src_max
|
|
|
|
|
|
|
|
udp_dst_min
|
|
|
|
udp_dst_max
|
|
|
|
|
|
|
|
flag
|
|
|
|
IPSRC_RND
|
|
|
|
IPDST_RND
|
|
|
|
UDPSRC_RND
|
|
|
|
UDPDST_RND
|
|
|
|
MACSRC_RND
|
|
|
|
MACDST_RND
|
2014-02-22 04:38:36 +08:00
|
|
|
TXSIZE_RND
|
|
|
|
IPV6
|
|
|
|
MPLS_RND
|
|
|
|
VID_RND
|
|
|
|
SVID_RND
|
|
|
|
FLOW_SEQ
|
|
|
|
QUEUE_MAP_RND
|
|
|
|
QUEUE_MAP_CPU
|
|
|
|
UDPCSUM
|
2014-01-03 11:18:34 +08:00
|
|
|
IPSEC
|
2014-02-22 04:38:36 +08:00
|
|
|
NODE_ALLOC
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
dst_min
|
|
|
|
dst_max
|
|
|
|
|
|
|
|
src_min
|
|
|
|
src_max
|
|
|
|
|
|
|
|
dst_mac
|
|
|
|
src_mac
|
|
|
|
|
|
|
|
clear_counters
|
|
|
|
|
|
|
|
dst6
|
|
|
|
src6
|
|
|
|
|
|
|
|
flows
|
|
|
|
flowlen
|
|
|
|
|
2010-06-10 06:49:57 +08:00
|
|
|
rate
|
|
|
|
ratep
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
References:
|
|
|
|
ftp://robur.slu.se/pub/Linux/net-development/pktgen-testing/
|
|
|
|
ftp://robur.slu.se/pub/Linux/net-development/pktgen-testing/examples/
|
|
|
|
|
|
|
|
Paper from Linux-Kongress in Erlangen 2004.
|
|
|
|
ftp://robur.slu.se/pub/Linux/net-development/pktgen-testing/pktgen_paper.pdf
|
|
|
|
|
|
|
|
Thanks to:
|
|
|
|
Grant Grundler for testing on IA-64 and parisc, Harald Welte, Lennert Buytenhek
|
|
|
|
Stephen Hemminger, Andi Kleen, Dave Miller and many others.
|
|
|
|
|
|
|
|
|
2006-03-23 17:10:26 +08:00
|
|
|
Good luck with the linux net-development.
|