DPDK documentation

Getting Started Guide for Linux

Introduction

This document contains instructions for installing and configuring the Data Plane Development Kit (DPDK) software. It is designed to get customers up and running quickly. The document describes how to compile and run a DPDK application in a Linux application (linuxapp) environment, without going deeply into detail.

Documentation Roadmap

The following is a list of DPDK documents in the suggested reading order:

  • Release Notes: Provides release-specific information, including supported features, limitations, fixed issues, known issues and so on. Also, provides the answers to frequently asked questions in FAQ format.

  • Getting Started Guide (this document): Describes how to install and configure the DPDK; designed to get users up and running quickly with the software.

  • Programmer’s Guide: Describes:

    • The software architecture and how to use it (through examples), specifically in a Linux application (linuxapp) environment
    • The content of the DPDK, the build system (including the commands that can be used in the root DPDK Makefile to build the development kit and an application) and guidelines for porting an application
    • Optimizations used in the software and those that should be considered for new development

    A glossary of terms is also provided.

  • API Reference: Provides detailed information about DPDK functions, data structures and other programming constructs.

  • Sample Applications User Guide: Describes a set of sample applications. Each chapter describes a sample application that showcases specific functionality and provides instructions on how to compile, run and use the sample application.

System Requirements

This chapter describes the packages required to compile the DPDK.

Note

If the DPDK is being used on an Intel® Communications Chipset 89xx Series platform, please consult the Intel® Communications Chipset 89xx Series Software for Linux Getting Started Guide.

BIOS Setting Prerequisite on x86

For the majority of platforms, no special BIOS settings are needed to use basic DPDK functionality. However, for additional HPET timer and power management functionality, and high performance of small packets, BIOS setting changes may be needed. Consult the section on Enabling Additional Functionality for more information on the required changes.

Note

If UEFI secure boot is enabled, the Linux kernel may disallow the use of UIO on the system. Therefore, devices for use by DPDK should be bound to the vfio-pci kernel module rather than igb_uio or uio_pci_generic. For more details see Binding and Unbinding Network Ports to/from the Kernel Modules.

Compilation of the DPDK

Required Tools and Libraries:

Note

Testing has been performed using Fedora 18. The setup commands and installed packages needed on other systems may be different. For details on other Linux distributions and the versions tested, please consult the DPDK Release Notes.

  • GNU make.

  • coreutils: cmp, sed, grep, arch, etc.

  • gcc: versions 4.9 or later is recommended for all platforms. On some distributions, some specific compiler flags and linker flags are enabled by default and affect performance (-fstack-protector, for example). Please refer to the documentation of your distribution and to gcc -dumpspecs.

  • libc headers, often packaged as gcc-multilib (glibc-devel.i686 / libc6-dev-i386; glibc-devel.x86_64 / libc6-dev for 64-bit compilation on Intel architecture; glibc-devel.ppc64 for 64 bit IBM Power architecture;)

  • Linux kernel headers or sources required to build kernel modules. (kernel - devel.x86_64; kernel - devel.ppc64)

  • Additional packages required for 32-bit compilation on 64-bit systems are:

    • glibc.i686, libgcc.i686, libstdc++.i686 and glibc-devel.i686 for Intel i686/x86_64;
    • glibc.ppc64, libgcc.ppc64, libstdc++.ppc64 and glibc-devel.ppc64 for IBM ppc_64;

    Note

    x86_x32 ABI is currently supported with distribution packages only on Ubuntu higher than 13.10 or recent Debian distribution. The only supported compiler is gcc 4.9+.

  • libnuma-devel - library for handling NUMA (Non Uniform Memory Access).

  • Python, version 2.7+ or 3.2+, to use various helper scripts included in the DPDK package.

Optional Tools:

  • Intel® C++ Compiler (icc). For installation, additional libraries may be required. See the icc Installation Guide found in the Documentation directory under the compiler installation.
  • IBM® Advance ToolChain for Powerlinux. This is a set of open source development tools and runtime libraries which allows users to take leading edge advantage of IBM’s latest POWER hardware features on Linux. To install it, see the IBM official installation document.
  • libpcap headers and libraries (libpcap-devel) to compile and use the libpcap-based poll-mode driver. This driver is disabled by default and can be enabled by setting CONFIG_RTE_LIBRTE_PMD_PCAP=y in the build time config file.
  • libarchive headers and library are needed for some unit tests using tar to get their resources.

Running DPDK Applications

To run an DPDK application, some customization may be required on the target machine.

System Software

Required:

  • Kernel version >= 2.6.34

    The kernel version in use can be checked using the command:

    uname -r
    
  • glibc >= 2.7 (for features related to cpuset)

    The version can be checked using the ldd --version command.

  • Kernel configuration

    In the Fedora OS and other common distributions, such as Ubuntu, or Red Hat Enterprise Linux, the vendor supplied kernel configurations can be used to run most DPDK applications.

    For other kernel builds, options which should be enabled for DPDK include:

Use of Hugepages in the Linux Environment

Hugepage support is required for the large memory pool allocation used for packet buffers (the HUGETLBFS option must be enabled in the running kernel as indicated the previous section). By using hugepage allocations, performance is increased since fewer pages are needed, and therefore less Translation Lookaside Buffers (TLBs, high speed translation caches), which reduce the time it takes to translate a virtual page address to a physical page address. Without hugepages, high TLB miss rates would occur with the standard 4k page size, slowing performance.

Reserving Hugepages for DPDK Use

The allocation of hugepages should be done at boot time or as soon as possible after system boot to prevent memory from being fragmented in physical memory. To reserve hugepages at boot time, a parameter is passed to the Linux kernel on the kernel command line.

For 2 MB pages, just pass the hugepages option to the kernel. For example, to reserve 1024 pages of 2 MB, use:

hugepages=1024

For other hugepage sizes, for example 1G pages, the size must be specified explicitly and can also be optionally set as the default hugepage size for the system. For example, to reserve 4G of hugepage memory in the form of four 1G pages, the following options should be passed to the kernel:

default_hugepagesz=1G hugepagesz=1G hugepages=4

Note

The hugepage sizes that a CPU supports can be determined from the CPU flags on Intel architecture. If pse exists, 2M hugepages are supported; if pdpe1gb exists, 1G hugepages are supported. On IBM Power architecture, the supported hugepage sizes are 16MB and 16GB.

Note

For 64-bit applications, it is recommended to use 1 GB hugepages if the platform supports them.

In the case of a dual-socket NUMA system, the number of hugepages reserved at boot time is generally divided equally between the two sockets (on the assumption that sufficient memory is present on both sockets).

See the Documentation/kernel-parameters.txt file in your Linux source tree for further details of these and other kernel options.

Alternative:

For 2 MB pages, there is also the option of allocating hugepages after the system has booted. This is done by echoing the number of hugepages required to a nr_hugepages file in the /sys/devices/ directory. For a single-node system, the command to use is as follows (assuming that 1024 pages are required):

echo 1024 > /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages

On a NUMA machine, pages should be allocated explicitly on separate nodes:

echo 1024 > /sys/devices/system/node/node0/hugepages/hugepages-2048kB/nr_hugepages
echo 1024 > /sys/devices/system/node/node1/hugepages/hugepages-2048kB/nr_hugepages

Note

For 1G pages, it is not possible to reserve the hugepage memory after the system has booted.

On IBM POWER system, the nr_overcommit_hugepages should be set to the same value as nr_hugepages. For example, if the required page number is 128, the following commands are used:

echo 128 > /sys/kernel/mm/hugepages/hugepages-16384kB/nr_hugepages
echo 128 > /sys/kernel/mm/hugepages/hugepages-16384kB/nr_overcommit_hugepages
Using Hugepages with the DPDK

Once the hugepage memory is reserved, to make the memory available for DPDK use, perform the following steps:

mkdir /mnt/huge
mount -t hugetlbfs nodev /mnt/huge

The mount point can be made permanent across reboots, by adding the following line to the /etc/fstab file:

nodev /mnt/huge hugetlbfs defaults 0 0

For 1GB pages, the page size must be specified as a mount option:

nodev /mnt/huge_1GB hugetlbfs pagesize=1GB 0 0

Compiling the DPDK Target from Source

Note

Parts of this process can also be done using the setup script described in the Quick Start Setup Script section of this document.

Install the DPDK and Browse Sources

First, uncompress the archive and move to the uncompressed DPDK source directory:

tar xJf dpdk-<version>.tar.xz
cd dpdk-<version>

The DPDK is composed of several directories:

  • lib: Source code of DPDK libraries
  • drivers: Source code of DPDK poll-mode drivers
  • app: Source code of DPDK applications (automatic tests)
  • examples: Source code of DPDK application examples
  • config, buildtools, mk: Framework-related makefiles, scripts and configuration

Installation of DPDK Target Environments

The format of a DPDK target is:

ARCH-MACHINE-EXECENV-TOOLCHAIN

where:

  • ARCH can be: i686, x86_64, ppc_64, arm64
  • MACHINE can be: native, power8, armv8a
  • EXECENV can be: linuxapp, bsdapp
  • TOOLCHAIN can be: gcc, icc

The targets to be installed depend on the 32-bit and/or 64-bit packages and compilers installed on the host. Available targets can be found in the DPDK/config directory. The defconfig_ prefix should not be used.

Note

Configuration files are provided with the RTE_MACHINE optimization level set. Within the configuration files, the RTE_MACHINE configuration value is set to native, which means that the compiled software is tuned for the platform on which it is built. For more information on this setting, and its possible values, see the DPDK Programmers Guide.

When using the Intel® C++ Compiler (icc), one of the following commands should be invoked for 64-bit or 32-bit use respectively. Notice that the shell scripts update the $PATH variable and therefore should not be performed in the same session. Also, verify the compiler’s installation directory since the path may be different:

source /opt/intel/bin/iccvars.sh intel64
source /opt/intel/bin/iccvars.sh ia32

To install and make targets, use the make install T=<target> command in the top-level DPDK directory.

For example, to compile a 64-bit target using icc, run:

make install T=x86_64-native-linuxapp-icc

To compile a 32-bit build using gcc, the make command should be:

make install T=i686-native-linuxapp-gcc

To prepare a target without building it, for example, if the configuration changes need to be made before compilation, use the make config T=<target> command:

make config T=x86_64-native-linuxapp-gcc

Warning

Any kernel modules to be used, e.g. igb_uio, kni, must be compiled with the same kernel as the one running on the target. If the DPDK is not being built on the target machine, the RTE_KERNELDIR environment variable should be used to point the compilation at a copy of the kernel version to be used on the target machine.

Once the target environment is created, the user may move to the target environment directory and continue to make code changes and re-compile. The user may also make modifications to the compile-time DPDK configuration by editing the .config file in the build directory. (This is a build-local copy of the defconfig file from the top- level config directory).

cd x86_64-native-linuxapp-gcc
vi .config
make

In addition, the make clean command can be used to remove any existing compiled files for a subsequent full, clean rebuild of the code.

Browsing the Installed DPDK Environment Target

Once a target is created it contains all libraries, including poll-mode drivers, and header files for the DPDK environment that are required to build customer applications. In addition, the test and testpmd applications are built under the build/app directory, which may be used for testing. A kmod directory is also present that contains kernel modules which may be loaded if needed.

Linux Drivers

Different PMDs may require different kernel drivers in order to work properly. Depends on the PMD being used, a corresponding kernel driver should be load and bind to the network ports.

UIO

A small kernel module to set up the device, map device memory to user-space and register interrupts. In many cases, the standard uio_pci_generic module included in the Linux kernel can provide the uio capability. This module can be loaded using the command:

sudo modprobe uio_pci_generic

Note

uio_pci_generic module doesn’t support the creation of virtual functions.

As an alternative to the uio_pci_generic, the DPDK also includes the igb_uio module which can be found in the kmod subdirectory referred to above. It can be loaded as shown below:

sudo modprobe uio
sudo insmod kmod/igb_uio.ko

Note

For some devices which lack support for legacy interrupts, e.g. virtual function (VF) devices, the igb_uio module may be needed in place of uio_pci_generic.

Note

If UEFI secure boot is enabled, the Linux kernel may disallow the use of UIO on the system. Therefore, devices for use by DPDK should be bound to the vfio-pci kernel module rather than igb_uio or uio_pci_generic. For more details see Binding and Unbinding Network Ports to/from the Kernel Modules below.

Since DPDK release 1.7 onward provides VFIO support, use of UIO is optional for platforms that support using VFIO.

VFIO

A more robust and secure driver in compare to the UIO, relying on IOMMU protection. To make use of VFIO, the vfio-pci module must be loaded:

sudo modprobe vfio-pci

Note that in order to use VFIO, your kernel must support it. VFIO kernel modules have been included in the Linux kernel since version 3.6.0 and are usually present by default, however please consult your distributions documentation to make sure that is the case.

Also, to use VFIO, both kernel and BIOS must support and be configured to use IO virtualization (such as Intel® VT-d).

Note

vfio-pci module doesn’t support the creation of virtual functions.

For proper operation of VFIO when running DPDK applications as a non-privileged user, correct permissions should also be set up. This can be done by using the DPDK setup script (called dpdk-setup.sh and located in the usertools directory).

Note

VFIO can be used without IOMMU. While this is just as unsafe as using UIO, it does make it possible for the user to keep the degree of device access and programming that VFIO has, in situations where IOMMU is not available.

Bifurcated Driver

PMDs which use the bifurcated driver co-exists with the device kernel driver. On such model the NIC is controlled by the kernel, while the data path is performed by the PMD directly on top of the device.

Such model has the following benefits:

  • It is secure and robust, as the memory management and isolation is done by the kernel.
  • It enables the user to use legacy linux tools such as ethtool or ifconfig while running DPDK application on the same network ports.
  • It enables the DPDK application to filter only part of the traffic, While the rest will be directed and handled by the kernel driver.

More about the bifurcated driver can be found in Mellanox Bifurcated DPDK PMD.

Binding and Unbinding Network Ports to/from the Kernel Modules

Note

PMDs Which use the bifurcated driver should not be unbind from their kernel drivers. this section is for PMDs which use the UIO or VFIO drivers.

As of release 1.4, DPDK applications no longer automatically unbind all supported network ports from the kernel driver in use. Instead, in case the PMD being used use the UIO or VFIO drivers, all ports that are to be used by an DPDK application must be bound to the uio_pci_generic, igb_uio or vfio-pci module before the application is run. For such PMDs, any network ports under Linux* control will be ignored and cannot be used by the application.

To bind ports to the uio_pci_generic, igb_uio or vfio-pci module for DPDK use, and then subsequently return ports to Linux* control, a utility script called dpdk-devbind.py is provided in the usertools subdirectory. This utility can be used to provide a view of the current state of the network ports on the system, and to bind and unbind those ports from the different kernel modules, including the uio and vfio modules. The following are some examples of how the script can be used. A full description of the script and its parameters can be obtained by calling the script with the --help or --usage options. Note that the uio or vfio kernel modules to be used, should be loaded into the kernel before running the dpdk-devbind.py script.

Warning

Due to the way VFIO works, there are certain limitations to which devices can be used with VFIO. Mainly it comes down to how IOMMU groups work. Any Virtual Function device can be used with VFIO on its own, but physical devices will require either all ports bound to VFIO, or some of them bound to VFIO while others not being bound to anything at all.

If your device is behind a PCI-to-PCI bridge, the bridge will then be part of the IOMMU group in which your device is in. Therefore, the bridge driver should also be unbound from the bridge PCI device for VFIO to work with devices behind the bridge.

Warning

While any user can run the dpdk-devbind.py script to view the status of the network ports, binding or unbinding network ports requires root privileges.

To see the status of all network ports on the system:

./usertools/dpdk-devbind.py --status

Network devices using DPDK-compatible driver
============================================
0000:82:00.0 '82599EB 10-GbE NIC' drv=uio_pci_generic unused=ixgbe
0000:82:00.1 '82599EB 10-GbE NIC' drv=uio_pci_generic unused=ixgbe

Network devices using kernel driver
===================================
0000:04:00.0 'I350 1-GbE NIC' if=em0  drv=igb unused=uio_pci_generic *Active*
0000:04:00.1 'I350 1-GbE NIC' if=eth1 drv=igb unused=uio_pci_generic
0000:04:00.2 'I350 1-GbE NIC' if=eth2 drv=igb unused=uio_pci_generic
0000:04:00.3 'I350 1-GbE NIC' if=eth3 drv=igb unused=uio_pci_generic

Other network devices
=====================
<none>

To bind device eth1,``04:00.1``, to the uio_pci_generic driver:

./usertools/dpdk-devbind.py --bind=uio_pci_generic 04:00.1

or, alternatively,

./usertools/dpdk-devbind.py --bind=uio_pci_generic eth1

To restore device 82:00.0 to its original kernel binding:

./usertools/dpdk-devbind.py --bind=ixgbe 82:00.0

Compiling and Running Sample Applications

The chapter describes how to compile and run applications in an DPDK environment. It also provides a pointer to where sample applications are stored.

Note

Parts of this process can also be done using the setup script described the Quick Start Setup Script section of this document.

Compiling a Sample Application

Once an DPDK target environment directory has been created (such as x86_64-native-linuxapp-gcc), it contains all libraries and header files required to build an application.

When compiling an application in the Linux* environment on the DPDK, the following variables must be exported:

  • RTE_SDK - Points to the DPDK installation directory.
  • RTE_TARGET - Points to the DPDK target environment directory.

The following is an example of creating the helloworld application, which runs in the DPDK Linux environment. This example may be found in the ${RTE_SDK}/examples directory.

The directory contains the main.c file. This file, when combined with the libraries in the DPDK target environment, calls the various functions to initialize the DPDK environment, then launches an entry point (dispatch application) for each core to be utilized. By default, the binary is generated in the build directory.

cd examples/helloworld/
export RTE_SDK=$HOME/DPDK
export RTE_TARGET=x86_64-native-linuxapp-gcc

make
    CC main.o
    LD helloworld
    INSTALL-APP helloworld
    INSTALL-MAP helloworld.map

ls build/app
    helloworld helloworld.map

Note

In the above example, helloworld was in the directory structure of the DPDK. However, it could have been located outside the directory structure to keep the DPDK structure intact. In the following case, the helloworld application is copied to a new directory as a new starting point.

export RTE_SDK=/home/user/DPDK
cp -r $(RTE_SDK)/examples/helloworld my_rte_app
cd my_rte_app/
export RTE_TARGET=x86_64-native-linuxapp-gcc

make
  CC main.o
  LD helloworld
  INSTALL-APP helloworld
  INSTALL-MAP helloworld.map

Running a Sample Application

Warning

Before running the application make sure:

  • Hugepages setup is done.
  • Any kernel driver being used is loaded.
  • In case needed, ports being used by the application should be bound to the corresponding kernel driver.

refer to Linux Drivers for more details.

The application is linked with the DPDK target environment’s Environmental Abstraction Layer (EAL) library, which provides some options that are generic to every DPDK application.

The following is the list of options that can be given to the EAL:

./rte-app [-c COREMASK | -l CORELIST] [-n NUM] [-b <domain:bus:devid.func>] \
          [--socket-mem=MB,...] [-d LIB.so|DIR] [-m MB] [-r NUM] [-v] [--file-prefix] \
          [--proc-type <primary|secondary|auto>]

The EAL options are as follows:

  • -c COREMASK or -l CORELIST: An hexadecimal bit mask of the cores to run on. Note that core numbering can change between platforms and should be determined beforehand. The corelist is a set of core numbers instead of a bitmap core mask.
  • -n NUM: Number of memory channels per processor socket.
  • -b <domain:bus:devid.func>: Blacklisting of ports; prevent EAL from using specified PCI device (multiple -b options are allowed).
  • --use-device: use the specified Ethernet device(s) only. Use comma-separate [domain:]bus:devid.func values. Cannot be used with -b option.
  • --socket-mem: Memory to allocate from hugepages on specific sockets.
  • -d: Add a driver or driver directory to be loaded. The application should use this option to load the pmd drivers that are built as shared libraries.
  • -m MB: Memory to allocate from hugepages, regardless of processor socket. It is recommended that --socket-mem be used instead of this option.
  • -r NUM: Number of memory ranks.
  • -v: Display version information on startup.
  • --huge-dir: The directory where hugetlbfs is mounted.
  • mbuf-pool-ops-name: Pool ops name for mbuf to use.
  • --file-prefix: The prefix text used for hugepage filenames.
  • --proc-type: The type of process instance.
  • --vmware-tsc-map: Use VMware TSC map instead of native RDTSC.
  • --base-virtaddr: Specify base virtual address.
  • --vfio-intr: Specify interrupt type to be used by VFIO (has no effect if VFIO is not used).

The -c or -l and option is mandatory; the others are optional.

Copy the DPDK application binary to your target, then run the application as follows (assuming the platform has four memory channels per processor socket, and that cores 0-3 are present and are to be used for running the application):

./helloworld -l 0-3 -n 4

Note

The --proc-type and --file-prefix EAL options are used for running multiple DPDK processes. See the “Multi-process Sample Application” chapter in the DPDK Sample Applications User Guide and the DPDK Programmers Guide for more details.

Logical Core Use by Applications

The coremask (-c 0x0f) or corelist (-l 0-3) parameter is always mandatory for DPDK applications. Each bit of the mask corresponds to the equivalent logical core number as reported by Linux. The preferred corelist option is a cleaner method to define cores to be used. Since these logical core numbers, and their mapping to specific cores on specific NUMA sockets, can vary from platform to platform, it is recommended that the core layout for each platform be considered when choosing the coremask/corelist to use in each case.

On initialization of the EAL layer by an DPDK application, the logical cores to be used and their socket location are displayed. This information can also be determined for all cores on the system by examining the /proc/cpuinfo file, for example, by running cat /proc/cpuinfo. The physical id attribute listed for each processor indicates the CPU socket to which it belongs. This can be useful when using other processors to understand the mapping of the logical cores to the sockets.

Note

A more graphical view of the logical core layout may be obtained using the lstopo Linux utility. On Fedora Linux, this may be installed and run using the following command:

sudo yum install hwloc
./lstopo

Warning

The logical core layout can change between different board layouts and should be checked before selecting an application coremask/corelist.

Hugepage Memory Use by Applications

When running an application, it is recommended to use the same amount of memory as that allocated for hugepages. This is done automatically by the DPDK application at startup, if no -m or --socket-mem parameter is passed to it when run.

If more memory is requested by explicitly passing a -m or --socket-mem value, the application fails. However, the application itself can also fail if the user requests less memory than the reserved amount of hugepage-memory, particularly if using the -m option. The reason is as follows. Suppose the system has 1024 reserved 2 MB pages in socket 0 and 1024 in socket 1. If the user requests 128 MB of memory, the 64 pages may not match the constraints:

  • The hugepage memory by be given to the application by the kernel in socket 1 only. In this case, if the application attempts to create an object, such as a ring or memory pool in socket 0, it fails. To avoid this issue, it is recommended that the --socket-mem option be used instead of the -m option.
  • These pages can be located anywhere in physical memory, and, although the DPDK EAL will attempt to allocate memory in contiguous blocks, it is possible that the pages will not be contiguous. In this case, the application is not able to allocate big memory pools.

The socket-mem option can be used to request specific amounts of memory for specific sockets. This is accomplished by supplying the --socket-mem flag followed by amounts of memory requested on each socket, for example, supply --socket-mem=0,512 to try and reserve 512 MB for socket 1 only. Similarly, on a four socket system, to allocate 1 GB memory on each of sockets 0 and 2 only, the parameter --socket-mem=1024,0,1024 can be used. No memory will be reserved on any CPU socket that is not explicitly referenced, for example, socket 3 in this case. If the DPDK cannot allocate enough memory on each socket, the EAL initialization fails.

Additional Sample Applications

Additional sample applications are included in the ${RTE_SDK}/examples directory. These sample applications may be built and run in a manner similar to that described in earlier sections in this manual. In addition, see the DPDK Sample Applications User Guide for a description of the application, specific instructions on compilation and execution and some explanation of the code.

Additional Test Applications

In addition, there are two other applications that are built when the libraries are created. The source files for these are in the DPDK/app directory and are called test and testpmd. Once the libraries are created, they can be found in the build/app directory.

  • The test application provides a variety of specific tests for the various functions in the DPDK.
  • The testpmd application provides a number of different packet throughput tests and examples of features such as how to use the Flow Director found in the Intel® 82599 10 Gigabit Ethernet Controller.

Enabling Additional Functionality

High Precision Event Timer (HPET) Functionality

BIOS Support

The High Precision Timer (HPET) must be enabled in the platform BIOS if the HPET is to be used. Otherwise, the Time Stamp Counter (TSC) is used by default. The BIOS is typically accessed by pressing F2 while the platform is starting up. The user can then navigate to the HPET option. On the Crystal Forest platform BIOS, the path is: Advanced -> PCH-IO Configuration -> High Precision Timer -> (Change from Disabled to Enabled if necessary).

On a system that has already booted, the following command can be issued to check if HPET is enabled:

grep hpet /proc/timer_list

If no entries are returned, HPET must be enabled in the BIOS (as per the instructions above) and the system rebooted.

Linux Kernel Support

The DPDK makes use of the platform HPET timer by mapping the timer counter into the process address space, and as such, requires that the HPET_MMAP kernel configuration option be enabled.

Warning

On Fedora, and other common distributions such as Ubuntu, the HPET_MMAP kernel option is not enabled by default. To recompile the Linux kernel with this option enabled, please consult the distributions documentation for the relevant instructions.

Enabling HPET in the DPDK

By default, HPET support is disabled in the DPDK build configuration files. To use HPET, the CONFIG_RTE_LIBEAL_USE_HPET setting should be changed to y, which will enable the HPET settings at compile time.

For an application to use the rte_get_hpet_cycles() and rte_get_hpet_hz() API calls, and optionally to make the HPET the default time source for the rte_timer library, the new rte_eal_hpet_init() API call should be called at application initialization. This API call will ensure that the HPET is accessible, returning an error to the application if it is not, for example, if HPET_MMAP is not enabled in the kernel. The application can then determine what action to take, if any, if the HPET is not available at run-time.

Note

For applications that require timing APIs, but not the HPET timer specifically, it is recommended that the rte_get_timer_cycles() and rte_get_timer_hz() API calls be used instead of the HPET-specific APIs. These generic APIs can work with either TSC or HPET time sources, depending on what is requested by an application call to rte_eal_hpet_init(), if any, and on what is available on the system at runtime.

Running DPDK Applications Without Root Privileges

Note

The instructions below will allow running DPDK as non-root with older Linux kernel versions. However, since version 4.0, the kernel does not allow unprivileged processes to read the physical address information from the pagemaps file, making it impossible for those processes to use HW devices which require physical addresses

Although applications using the DPDK use network ports and other hardware resources directly, with a number of small permission adjustments it is possible to run these applications as a user other than “root”. To do so, the ownership, or permissions, on the following Linux file system objects should be adjusted to ensure that the Linux user account being used to run the DPDK application has access to them:

  • All directories which serve as hugepage mount points, for example, /mnt/huge

  • The userspace-io device files in /dev, for example, /dev/uio0, /dev/uio1, and so on

  • The userspace-io sysfs config and resource files, for example for uio0:

    /sys/class/uio/uio0/device/config
    /sys/class/uio/uio0/device/resource*
    
  • If the HPET is to be used, /dev/hpet

Note

On some Linux installations, /dev/hugepages is also a hugepage mount point created by default.

Power Management and Power Saving Functionality

Enhanced Intel SpeedStep® Technology must be enabled in the platform BIOS if the power management feature of DPDK is to be used. Otherwise, the sys file folder /sys/devices/system/cpu/cpu0/cpufreq will not exist, and the CPU frequency- based power management cannot be used. Consult the relevant BIOS documentation to determine how these settings can be accessed.

For example, on some Intel reference platform BIOS variants, the path to Enhanced Intel SpeedStep® Technology is:

Advanced
  -> Processor Configuration
  -> Enhanced Intel SpeedStep® Tech

In addition, C3 and C6 should be enabled as well for power management. The path of C3 and C6 on the same platform BIOS is:

Advanced
  -> Processor Configuration
  -> Processor C3 Advanced
  -> Processor Configuration
  -> Processor C6

Using Linux Core Isolation to Reduce Context Switches

While the threads used by an DPDK application are pinned to logical cores on the system, it is possible for the Linux scheduler to run other tasks on those cores also. To help prevent additional workloads from running on those cores, it is possible to use the isolcpus Linux kernel parameter to isolate them from the general Linux scheduler.

For example, if DPDK applications are to run on logical cores 2, 4 and 6, the following should be added to the kernel parameter list:

isolcpus=2,4,6

Loading the DPDK KNI Kernel Module

To run the DPDK Kernel NIC Interface (KNI) sample application, an extra kernel module (the kni module) must be loaded into the running kernel. The module is found in the kmod sub-directory of the DPDK target directory. Similar to the loading of the igb_uio module, this module should be loaded using the insmod command as shown below (assuming that the current directory is the DPDK target directory):

insmod kmod/rte_kni.ko

Note

See the “Kernel NIC Interface Sample Application” chapter in the DPDK Sample Applications User Guide for more details.

Using Linux IOMMU Pass-Through to Run DPDK with Intel® VT-d

To enable Intel® VT-d in a Linux kernel, a number of kernel configuration options must be set. These include:

  • IOMMU_SUPPORT
  • IOMMU_API
  • INTEL_IOMMU

In addition, to run the DPDK with Intel® VT-d, the iommu=pt kernel parameter must be used when using igb_uio driver. This results in pass-through of the DMAR (DMA Remapping) lookup in the host. Also, if INTEL_IOMMU_DEFAULT_ON is not set in the kernel, the intel_iommu=on kernel parameter must be used too. This ensures that the Intel IOMMU is being initialized as expected.

Please note that while using iommu=pt is compulsory for igb_uio driver, the vfio-pci driver can actually work with both iommu=pt and iommu=on.

Quick Start Setup Script

The dpdk-setup.sh script, found in the usertools subdirectory, allows the user to perform the following tasks:

  • Build the DPDK libraries
  • Insert and remove the DPDK IGB_UIO kernel module
  • Insert and remove VFIO kernel modules
  • Insert and remove the DPDK KNI kernel module
  • Create and delete hugepages for NUMA and non-NUMA cases
  • View network port status and reserve ports for DPDK application use
  • Set up permissions for using VFIO as a non-privileged user
  • Run the test and testpmd applications
  • Look at hugepages in the meminfo
  • List hugepages in /mnt/huge
  • Remove built DPDK libraries

Once these steps have been completed for one of the EAL targets, the user may compile their own application that links in the EAL libraries to create the DPDK image.

Script Organization

The dpdk-setup.sh script is logically organized into a series of steps that a user performs in sequence. Each step provides a number of options that guide the user to completing the desired task. The following is a brief synopsis of each step.

Step 1: Build DPDK Libraries

Initially, the user must select a DPDK target to choose the correct target type and compiler options to use when building the libraries.

The user must have all libraries, modules, updates and compilers installed in the system prior to this, as described in the earlier chapters in this Getting Started Guide.

Step 2: Setup Environment

The user configures the Linux* environment to support the running of DPDK applications. Hugepages can be set up for NUMA or non-NUMA systems. Any existing hugepages will be removed. The DPDK kernel module that is needed can also be inserted in this step, and network ports may be bound to this module for DPDK application use.

Step 3: Run an Application

The user may run the test application once the other steps have been performed. The test application allows the user to run a series of functional tests for the DPDK. The testpmd application, which supports the receiving and sending of packets, can also be run.

Step 4: Examining the System

This step provides some tools for examining the status of hugepage mappings.

Step 5: System Cleanup

The final step has options for restoring the system to its original state.

Use Cases

The following are some example of how to use the dpdk-setup.sh script. The script should be run using the source command. Some options in the script prompt the user for further data before proceeding.

Warning

The dpdk-setup.sh script should be run with root privileges.

source usertools/dpdk-setup.sh

------------------------------------------------------------------------

RTE_SDK exported as /home/user/rte

------------------------------------------------------------------------

Step 1: Select the DPDK environment to build

------------------------------------------------------------------------

[1] i686-native-linuxapp-gcc

[2] i686-native-linuxapp-icc

[3] ppc_64-power8-linuxapp-gcc

[4] x86_64-native-bsdapp-clang

[5] x86_64-native-bsdapp-gcc

[6] x86_64-native-linuxapp-clang

[7] x86_64-native-linuxapp-gcc

[8] x86_64-native-linuxapp-icc

------------------------------------------------------------------------

Step 2: Setup linuxapp environment

------------------------------------------------------------------------

[11] Insert IGB UIO module

[12] Insert VFIO module

[13] Insert KNI module

[14] Setup hugepage mappings for non-NUMA systems

[15] Setup hugepage mappings for NUMA systems

[16] Display current Ethernet device settings

[17] Bind Ethernet device to IGB UIO module

[18] Bind Ethernet device to VFIO module

[19] Setup VFIO permissions

------------------------------------------------------------------------

Step 3: Run test application for linuxapp environment

------------------------------------------------------------------------

[20] Run test application ($RTE_TARGET/app/test)

[21] Run testpmd application in interactive mode ($RTE_TARGET/app/testpmd)

------------------------------------------------------------------------

Step 4: Other tools

------------------------------------------------------------------------

[22] List hugepage info from /proc/meminfo

------------------------------------------------------------------------

Step 5: Uninstall and system cleanup

------------------------------------------------------------------------

[23] Uninstall all targets

[24] Unbind NICs from IGB UIO driver

[25] Remove IGB UIO module

[26] Remove VFIO module

[27] Remove KNI module

[28] Remove hugepage mappings

[29] Exit Script

Option:

The following selection demonstrates the creation of the x86_64-native-linuxapp-gcc DPDK library.

Option: 9

================== Installing x86_64-native-linuxapp-gcc

Configuration done
== Build lib
...
Build complete
RTE_TARGET exported as x86_64-native-linuxapp-gcc

The following selection demonstrates the starting of the DPDK UIO driver.

Option: 25

Unloading any existing DPDK UIO module
Loading DPDK UIO module

The following selection demonstrates the creation of hugepages in a NUMA system. 1024 2 MByte pages are assigned to each node. The result is that the application should use -m 4096 for starting the application to access both memory areas (this is done automatically if the -m option is not provided).

Note

If prompts are displayed to remove temporary files, type ‘y’.

Option: 15

Removing currently reserved hugepages
mounting /mnt/huge and removing directory
Input the number of 2MB pages for each node
Example: to have 128MB of hugepages available per node,
enter '64' to reserve 64 * 2MB pages on each node
Number of pages for node0: 1024
Number of pages for node1: 1024
Reserving hugepages
Creating /mnt/huge and mounting as hugetlbfs

The following selection demonstrates the launch of the test application to run on a single core.

Option: 20

Enter hex bitmask of cores to execute test app on
Example: to execute app on cores 0 to 7, enter 0xff
bitmask: 0x01
Launching app
EAL: coremask set to 1
EAL: Detected lcore 0 on socket 0
...
EAL: Master core 0 is ready (tid=1b2ad720)
RTE>>

Applications

Once the user has run the dpdk-setup.sh script, built one of the EAL targets and set up hugepages (if using one of the Linux EAL targets), the user can then move on to building and running their application or one of the examples provided.

The examples in the /examples directory provide a good starting point to gain an understanding of the operation of the DPDK. The following command sequence shows how the helloworld sample application is built and run. As recommended in Section 4.2.1 , “Logical Core Use by Applications”, the logical core layout of the platform should be determined when selecting a core mask to use for an application.

cd helloworld/
make
  CC main.o
  LD helloworld
  INSTALL-APP helloworld
  INSTALL-MAP helloworld.map

sudo ./build/app/helloworld -l 0-3 -n 3
[sudo] password for rte:

EAL: coremask set to f
EAL: Detected lcore 0 as core 0 on socket 0
EAL: Detected lcore 1 as core 0 on socket 1
EAL: Detected lcore 2 as core 1 on socket 0
EAL: Detected lcore 3 as core 1 on socket 1
EAL: Setting up hugepage memory...
EAL: Ask a virtual area of 0x200000 bytes
EAL: Virtual area found at 0x7f0add800000 (size = 0x200000)
EAL: Ask a virtual area of 0x3d400000 bytes
EAL: Virtual area found at 0x7f0aa0200000 (size = 0x3d400000)
EAL: Ask a virtual area of 0x400000 bytes
EAL: Virtual area found at 0x7f0a9fc00000 (size = 0x400000)
EAL: Ask a virtual area of 0x400000 bytes
EAL: Virtual area found at 0x7f0a9f600000 (size = 0x400000)
EAL: Ask a virtual area of 0x400000 bytes
EAL: Virtual area found at 0x7f0a9f000000 (size = 0x400000)
EAL: Ask a virtual area of 0x800000 bytes
EAL: Virtual area found at 0x7f0a9e600000 (size = 0x800000)
EAL: Ask a virtual area of 0x800000 bytes
EAL: Virtual area found at 0x7f0a9dc00000 (size = 0x800000)
EAL: Ask a virtual area of 0x400000 bytes
EAL: Virtual area found at 0x7f0a9d600000 (size = 0x400000)
EAL: Ask a virtual area of 0x400000 bytes
EAL: Virtual area found at 0x7f0a9d000000 (size = 0x400000)
EAL: Ask a virtual area of 0x400000 bytes
EAL: Virtual area found at 0x7f0a9ca00000 (size = 0x400000)
EAL: Ask a virtual area of 0x200000 bytes
EAL: Virtual area found at 0x7f0a9c600000 (size = 0x200000)
EAL: Ask a virtual area of 0x200000 bytes
EAL: Virtual area found at 0x7f0a9c200000 (size = 0x200000)
EAL: Ask a virtual area of 0x3fc00000 bytes
EAL: Virtual area found at 0x7f0a5c400000 (size = 0x3fc00000)
EAL: Ask a virtual area of 0x200000 bytes
EAL: Virtual area found at 0x7f0a5c000000 (size = 0x200000)
EAL: Requesting 1024 pages of size 2MB from socket 0
EAL: Requesting 1024 pages of size 2MB from socket 1
EAL: Master core 0 is ready (tid=de25b700)
EAL: Core 1 is ready (tid=5b7fe700)
EAL: Core 3 is ready (tid=5a7fc700)
EAL: Core 2 is ready (tid=5affd700)
hello from core 1
hello from core 2
hello from core 3
hello from core 0

How to get best performance with NICs on Intel platforms

This document is a step-by-step guide for getting high performance from DPDK applications on Intel platforms.

Hardware and Memory Requirements

For best performance use an Intel Xeon class server system such as Ivy Bridge, Haswell or newer.

Ensure that each memory channel has at least one memory DIMM inserted, and that the memory size for each is at least 4GB. Note: this has one of the most direct effects on performance.

You can check the memory configuration using dmidecode as follows:

dmidecode -t memory | grep Locator

Locator: DIMM_A1
Bank Locator: NODE 1
Locator: DIMM_A2
Bank Locator: NODE 1
Locator: DIMM_B1
Bank Locator: NODE 1
Locator: DIMM_B2
Bank Locator: NODE 1
...
Locator: DIMM_G1
Bank Locator: NODE 2
Locator: DIMM_G2
Bank Locator: NODE 2
Locator: DIMM_H1
Bank Locator: NODE 2
Locator: DIMM_H2
Bank Locator: NODE 2

The sample output above shows a total of 8 channels, from A to H, where each channel has 2 DIMMs.

You can also use dmidecode to determine the memory frequency:

dmidecode -t memory | grep Speed

Speed: 2133 MHz
Configured Clock Speed: 2134 MHz
Speed: Unknown
Configured Clock Speed: Unknown
Speed: 2133 MHz
Configured Clock Speed: 2134 MHz
Speed: Unknown
...
Speed: 2133 MHz
Configured Clock Speed: 2134 MHz
Speed: Unknown
Configured Clock Speed: Unknown
Speed: 2133 MHz
Configured Clock Speed: 2134 MHz
Speed: Unknown
Configured Clock Speed: Unknown

The output shows a speed of 2133 MHz (DDR4) and Unknown (not existing). This aligns with the previous output which showed that each channel has one memory bar.

Network Interface Card Requirements

Use a DPDK supported high end NIC such as the Intel XL710 40GbE.

Make sure each NIC has been flashed the latest version of NVM/firmware.

Use PCIe Gen3 slots, such as Gen3 x8 or Gen3 x16 because PCIe Gen2 slots don’t provide enough bandwidth for 2 x 10GbE and above. You can use lspci to check the speed of a PCI slot using something like the following:

lspci -s 03:00.1 -vv | grep LnkSta

LnkSta: Speed 8GT/s, Width x8, TrErr- Train- SlotClk+ DLActive- ...
LnkSta2: Current De-emphasis Level: -6dB, EqualizationComplete+ ...

When inserting NICs into PCI slots always check the caption, such as CPU0 or CPU1 to indicate which socket it is connected to.

Care should be take with NUMA. If you are using 2 or more ports from different NICs, it is best to ensure that these NICs are on the same CPU socket. An example of how to determine this is shown further below.

BIOS Settings

The following are some recommendations on BIOS settings. Different platforms will have different BIOS naming so the following is mainly for reference:

  1. Before starting consider resetting all BIOS settings to their default.
  2. Disable all power saving options such as: Power performance tuning, CPU P-State, CPU C3 Report and CPU C6 Report.
  3. Select Performance as the CPU Power and Performance policy.
  4. Disable Turbo Boost to ensure the performance scaling increases with the number of cores.
  5. Set memory frequency to the highest available number, NOT auto.
  6. Disable all virtualization options when you test the physical function of the NIC, and turn on VT-d if you wants to use VFIO.
Linux boot command line

The following are some recommendations on GRUB boot settings:

  1. Use the default grub file as a starting point.

  2. Reserve 1G huge pages via grub configurations. For example to reserve 8 huge pages of 1G size:

    default_hugepagesz=1G hugepagesz=1G hugepages=8
    
  3. Isolate CPU cores which will be used for DPDK. For example:

    isolcpus=2,3,4,5,6,7,8
    
  4. If it wants to use VFIO, use the following additional grub parameters:

    iommu=pt intel_iommu=on
    

Configurations before running DPDK

  1. Build the DPDK target and reserve huge pages. See the earlier section on Use of Hugepages in the Linux Environment for more details.

    The following shell commands may help with building and configuration:

    # Build DPDK target.
    cd dpdk_folder
    make install T=x86_64-native-linuxapp-gcc -j
    
    # Get the hugepage size.
    awk '/Hugepagesize/ {print $2}' /proc/meminfo
    
    # Get the total huge page numbers.
    awk '/HugePages_Total/ {print $2} ' /proc/meminfo
    
    # Unmount the hugepages.
    umount `awk '/hugetlbfs/ {print $2}' /proc/mounts`
    
    # Create the hugepage mount folder.
    mkdir -p /mnt/huge
    
    # Mount to the specific folder.
    mount -t hugetlbfs nodev /mnt/huge
    
  2. Check the CPU layout using using the DPDK cpu_layout utility:

    cd dpdk_folder
    
    usertools/cpu_layout.py
    

    Or run lscpu to check the the cores on each socket.

  3. Check your NIC id and related socket id:

    # List all the NICs with PCI address and device IDs.
    lspci -nn | grep Eth
    

    For example suppose your output was as follows:

    82:00.0 Ethernet [0200]: Intel XL710 for 40GbE QSFP+ [8086:1583]
    82:00.1 Ethernet [0200]: Intel XL710 for 40GbE QSFP+ [8086:1583]
    85:00.0 Ethernet [0200]: Intel XL710 for 40GbE QSFP+ [8086:1583]
    85:00.1 Ethernet [0200]: Intel XL710 for 40GbE QSFP+ [8086:1583]
    

    Check the PCI device related numa node id:

    cat /sys/bus/pci/devices/0000\:xx\:00.x/numa_node
    

    Usually 0x:00.x is on socket 0 and 8x:00.x is on socket 1. Note: To get the best performance, ensure that the core and NICs are in the same socket. In the example above 85:00.0 is on socket 1 and should be used by cores on socket 1 for the best performance.

4. Check which kernel drivers needs to be loaded and whether there is a need to unbind the network ports from their kernel drivers. More details about DPDK setup and Linux kernel requirements see Compiling the DPDK Target from Source and Linux Drivers.

Getting Started Guide for FreeBSD

Introduction

This document contains instructions for installing and configuring the Data Plane Development Kit (DPDK) software. It is designed to get customers up and running quickly and describes how to compile and run a DPDK application in a FreeBSD application (bsdapp) environment, without going deeply into detail.

For a comprehensive guide to installing and using FreeBSD, the following handbook is available from the FreeBSD Documentation Project: FreeBSD Handbook.

Note

The DPDK is now available as part of the FreeBSD ports collection. Installing via the ports collection infrastructure is now the recommended way to install the DPDK on FreeBSD, and is documented in the next chapter, Installing DPDK from the Ports Collection.

Documentation Roadmap

The following is a list of DPDK documents in the suggested reading order:

  • Release Notes : Provides release-specific information, including supported features, limitations, fixed issues, known issues and so on. Also, provides the answers to frequently asked questions in FAQ format.

  • Getting Started Guide (this document): Describes how to install and configure the DPDK; designed to get users up and running quickly with the software.

  • Programmer’s Guide: Describes:

    • The software architecture and how to use it (through examples), specifically in a Linux* application (linuxapp) environment
    • The content of the DPDK, the build system (including the commands that can be used in the root DPDK Makefile to build the development kit and an application) and guidelines for porting an application
    • Optimizations used in the software and those that should be considered for new development

    A glossary of terms is also provided.

  • API Reference: Provides detailed information about DPDK functions, data structures and other programming constructs.

  • Sample Applications User Guide: Describes a set of sample applications. Each chapter describes a sample application that showcases specific functionality and provides instructions on how to compile, run and use the sample application.

Installing DPDK from the Ports Collection

The easiest way to get up and running with the DPDK on FreeBSD is to install it from the ports collection. Details of getting and using the ports collection are documented in the FreeBSD Handbook.

Note

Testing has been performed using FreeBSD 10.0-RELEASE (x86_64) and requires the installation of the kernel sources, which should be included during the installation of FreeBSD.

Installing the DPDK FreeBSD Port

On a system with the ports collection installed in /usr/ports, the DPDK can be installed using the commands:

cd /usr/ports/net/dpdk

make install

After the installation of the DPDK port, instructions will be printed on how to install the kernel modules required to use the DPDK. A more complete version of these instructions can be found in the sections Loading the DPDK contigmem Module and Loading the DPDK nic_uio Module. Normally, lines like those below would be added to the file /boot/loader.conf.

# Reserve 2 x 1G blocks of contiguous memory using contigmem driver:
hw.contigmem.num_buffers=2
hw.contigmem.buffer_size=1073741824
contigmem_load="YES"

# Identify NIC devices for DPDK apps to use and load nic_uio driver:
hw.nic_uio.bdfs="2:0:0,2:0:1"
nic_uio_load="YES"

Compiling and Running the Example Applications

When the DPDK has been installed from the ports collection it installs its example applications in /usr/local/share/dpdk/examples - also accessible via symlink as /usr/local/share/examples/dpdk. These examples can be compiled and run as described in Compiling and Running Sample Applications. In this case, the required environmental variables should be set as below:

  • RTE_SDK=/usr/local/share/dpdk
  • RTE_TARGET=x86_64-native-bsdapp-clang

Note

To install a copy of the DPDK compiled using gcc, please download the official DPDK package from http://dpdk.org/ and install manually using the instructions given in the next chapter, Compiling the DPDK Target from Source

An example application can therefore be copied to a user’s home directory and compiled and run as below:

export RTE_SDK=/usr/local/share/dpdk

export RTE_TARGET=x86_64-native-bsdapp-clang

cp -r /usr/local/share/dpdk/examples/helloworld .

cd helloworld/

gmake
  CC main.o
  LD helloworld
  INSTALL-APP helloworld
  INSTALL-MAP helloworld.map

sudo ./build/helloworld -l 0-3 -n 2

EAL: Contigmem driver has 2 buffers, each of size 1GB
EAL: Sysctl reports 8 cpus
EAL: Detected lcore 0
EAL: Detected lcore 1
EAL: Detected lcore 2
EAL: Detected lcore 3
EAL: Support maximum 64 logical core(s) by configuration.
EAL: Detected 4 lcore(s)
EAL: Setting up physically contiguous memory...
EAL: Mapped memory segment 1 @ 0x802400000: len 1073741824
EAL: Mapped memory segment 2 @ 0x842400000: len 1073741824
EAL: WARNING: clock_gettime cannot use CLOCK_MONOTONIC_RAW and HPET
     is not available - clock timings may be less accurate.
EAL: TSC frequency is ~3569023 KHz
EAL: PCI scan found 24 devices
EAL: Master core 0 is ready (tid=0x802006400)
EAL: Core 1 is ready (tid=0x802006800)
EAL: Core 3 is ready (tid=0x802007000)
EAL: Core 2 is ready (tid=0x802006c00)
EAL: PCI device 0000:01:00.0 on NUMA socket 0
EAL:   probe driver: 8086:10fb rte_ixgbe_pmd
EAL:   PCI memory mapped at 0x80074a000
EAL:   PCI memory mapped at 0x8007ca000
EAL: PCI device 0000:01:00.1 on NUMA socket 0
EAL:   probe driver: 8086:10fb rte_ixgbe_pmd
EAL:   PCI memory mapped at 0x8007ce000
EAL:   PCI memory mapped at 0x80084e000
EAL: PCI device 0000:02:00.0 on NUMA socket 0
EAL:   probe driver: 8086:10fb rte_ixgbe_pmd
EAL:   PCI memory mapped at 0x800852000
EAL:   PCI memory mapped at 0x8008d2000
EAL: PCI device 0000:02:00.1 on NUMA socket 0
EAL:   probe driver: 8086:10fb rte_ixgbe_pmd
EAL:   PCI memory mapped at 0x801b3f000
EAL:   PCI memory mapped at 0x8008d6000
hello from core 1
hello from core 2
hello from core 3
hello from core 0

Note

To run a DPDK process as a non-root user, adjust the permissions on the /dev/contigmem and /dev/uio device nodes as described in section Running DPDK Applications Without Root Privileges

Note

For an explanation of the command-line parameters that can be passed to an DPDK application, see section Running a Sample Application.

Compiling the DPDK Target from Source

System Requirements

The DPDK and its applications require the GNU make system (gmake) to build on FreeBSD. Optionally, gcc may also be used in place of clang to build the DPDK, in which case it too must be installed prior to compiling the DPDK. The installation of these tools is covered in this section.

Compiling the DPDK requires the FreeBSD kernel sources, which should be included during the installation of FreeBSD on the development platform. The DPDK also requires the use of FreeBSD ports to compile and function.

To use the FreeBSD ports system, it is required to update and extract the FreeBSD ports tree by issuing the following commands:

portsnap fetch
portsnap extract

If the environment requires proxies for external communication, these can be set using:

setenv http_proxy <my_proxy_host>:<port>
setenv ftp_proxy <my_proxy_host>:<port>

The FreeBSD ports below need to be installed prior to building the DPDK. In general these can be installed using the following set of commands:

cd /usr/ports/<port_location>

make config-recursive

make install

make clean

Each port location can be found using:

whereis <port_name>

The ports required and their locations are as follows:

  • dialog4ports: /usr/ports/ports-mgmt/dialog4ports
  • GNU make(gmake): /usr/ports/devel/gmake
  • coreutils: /usr/ports/sysutils/coreutils

For compiling and using the DPDK with gcc, the compiler must be installed from the ports collection:

  • gcc: version 4.9 is recommended /usr/ports/lang/gcc49. Ensure that CPU_OPTS is selected (default is OFF).

When running the make config-recursive command, a dialog may be presented to the user. For the installation of the DPDK, the default options were used.

Note

To avoid multiple dialogs being presented to the user during make install, it is advisable before running the make install command to re-run the make config-recursive command until no more dialogs are seen.

Install the DPDK and Browse Sources

First, uncompress the archive and move to the DPDK source directory:

unzip DPDK-<version>.zip
cd DPDK-<version>

The DPDK is composed of several directories:

  • lib: Source code of DPDK libraries
  • app: Source code of DPDK applications (automatic tests)
  • examples: Source code of DPDK applications
  • config, buildtools, mk: Framework-related makefiles, scripts and configuration

Installation of the DPDK Target Environments

The format of a DPDK target is:

ARCH-MACHINE-EXECENV-TOOLCHAIN

Where:

  • ARCH is: x86_64
  • MACHINE is: native
  • EXECENV is: bsdapp
  • TOOLCHAIN is: gcc | clang

The configuration files for the DPDK targets can be found in the DPDK/config directory in the form of:

defconfig_ARCH-MACHINE-EXECENV-TOOLCHAIN

Note

Configuration files are provided with the RTE_MACHINE optimization level set. Within the configuration files, the RTE_MACHINE configuration value is set to native, which means that the compiled software is tuned for the platform on which it is built. For more information on this setting, and its possible values, see the DPDK Programmers Guide.

To make the target, use gmake install T=<target>.

For example to compile for FreeBSD use:

gmake install T=x86_64-native-bsdapp-clang

Note

If the compiler binary to be used does not correspond to that given in the TOOLCHAIN part of the target, the compiler command may need to be explicitly specified. For example, if compiling for gcc, where the gcc binary is called gcc4.9, the command would need to be gmake install T=<target> CC=gcc4.9.

Browsing the Installed DPDK Environment Target

Once a target is created, it contains all the libraries and header files for the DPDK environment that are required to build customer applications. In addition, the test and testpmd applications are built under the build/app directory, which may be used for testing. A kmod directory is also present that contains the kernel modules to install.

Loading the DPDK contigmem Module

To run a DPDK application, physically contiguous memory is required. In the absence of non-transparent superpages, the included sources for the contigmem kernel module provides the ability to present contiguous blocks of memory for the DPDK to use. The contigmem module must be loaded into the running kernel before any DPDK is run. The module is found in the kmod sub-directory of the DPDK target directory.

The amount of physically contiguous memory along with the number of physically contiguous blocks to be reserved by the module can be set at runtime prior to module loading using:

kenv hw.contigmem.num_buffers=n
kenv hw.contigmem.buffer_size=m

The kernel environment variables can also be specified during boot by placing the following in /boot/loader.conf:

hw.contigmem.num_buffers=n hw.contigmem.buffer_size=m

The variables can be inspected using the following command:

sysctl -a hw.contigmem

Where n is the number of blocks and m is the size in bytes of each area of contiguous memory. A default of two buffers of size 1073741824 bytes (1 Gigabyte) each is set during module load if they are not specified in the environment.

The module can then be loaded using kldload (assuming that the current directory is the DPDK target directory):

kldload ./kmod/contigmem.ko

It is advisable to include the loading of the contigmem module during the boot process to avoid issues with potential memory fragmentation during later system up time. This can be achieved by copying the module to the /boot/kernel/ directory and placing the following into /boot/loader.conf:

contigmem_load="YES"

Note

The contigmem_load directive should be placed after any definitions of hw.contigmem.num_buffers and hw.contigmem.buffer_size if the default values are not to be used.

An error such as:

kldload: can't load ./x86_64-native-bsdapp-gcc/kmod/contigmem.ko:
         Exec format error

is generally attributed to not having enough contiguous memory available and can be verified via dmesg or /var/log/messages:

kernel: contigmalloc failed for buffer <n>

To avoid this error, reduce the number of buffers or the buffer size.

Loading the DPDK nic_uio Module

After loading the contigmem module, the nic_uio module must also be loaded into the running kernel prior to running any DPDK application. This module must be loaded using the kldload command as shown below (assuming that the current directory is the DPDK target directory).

kldload ./kmod/nic_uio.ko

Note

If the ports to be used are currently bound to a existing kernel driver then the hw.nic_uio.bdfs sysctl value will need to be set before loading the module. Setting this value is described in the next section below.

Currently loaded modules can be seen by using the kldstat command and a module can be removed from the running kernel by using kldunload <module_name>.

To load the module during boot, copy the nic_uio module to /boot/kernel and place the following into /boot/loader.conf:

nic_uio_load="YES"

Note

nic_uio_load="YES" must appear after the contigmem_load directive, if it exists.

By default, the nic_uio module will take ownership of network ports if they are recognized DPDK devices and are not owned by another module. However, since the FreeBSD kernel includes support, either built-in, or via a separate driver module, for most network card devices, it is likely that the ports to be used are already bound to a driver other than nic_uio. The following sub-section describe how to query and modify the device ownership of the ports to be used by DPDK applications.

Binding Network Ports to the nic_uio Module

Device ownership can be viewed using the pciconf -l command. The example below shows four Intel® 82599 network ports under if_ixgbe module ownership.

pciconf -l
ix0@pci0:1:0:0: class=0x020000 card=0x00038086 chip=0x10fb8086 rev=0x01 hdr=0x00
ix1@pci0:1:0:1: class=0x020000 card=0x00038086 chip=0x10fb8086 rev=0x01 hdr=0x00
ix2@pci0:2:0:0: class=0x020000 card=0x00038086 chip=0x10fb8086 rev=0x01 hdr=0x00
ix3@pci0:2:0:1: class=0x020000 card=0x00038086 chip=0x10fb8086 rev=0x01 hdr=0x00

The first column constitutes three components:

  1. Device name: ixN
  2. Unit name: pci0
  3. Selector (Bus:Device:Function): 1:0:0

Where no driver is associated with a device, the device name will be none.

By default, the FreeBSD kernel will include built-in drivers for the most common devices; a kernel rebuild would normally be required to either remove the drivers or configure them as loadable modules.

To avoid building a custom kernel, the nic_uio module can detach a network port from its current device driver. This is achieved by setting the hw.nic_uio.bdfs kernel environment variable prior to loading nic_uio, as follows:

hw.nic_uio.bdfs="b:d:f,b:d:f,..."

Where a comma separated list of selectors is set, the list must not contain any whitespace.

For example to re-bind ix2@pci0:2:0:0 and ix3@pci0:2:0:1 to the nic_uio module upon loading, use the following command:

kenv hw.nic_uio.bdfs="2:0:0,2:0:1"

The variable can also be specified during boot by placing the following into /boot/loader.conf, before the previously-described nic_uio_load line - as shown:

hw.nic_uio.bdfs="2:0:0,2:0:1"
nic_uio_load="YES"
Binding Network Ports Back to their Original Kernel Driver

If the original driver for a network port has been compiled into the kernel, it is necessary to reboot FreeBSD to restore the original device binding. Before doing so, update or remove the hw.nic_uio.bdfs in /boot/loader.conf.

If rebinding to a driver that is a loadable module, the network port binding can be reset without rebooting. To do so, unload both the target kernel module and the nic_uio module, modify or clear the hw.nic_uio.bdfs kernel environment (kenv) value, and reload the two drivers - first the original kernel driver, and then the nic_uio driver. Note: the latter does not need to be reloaded unless there are ports that are still to be bound to it.

Example commands to perform these steps are shown below:

kldunload nic_uio
kldunload <original_driver>

# To clear the value completely:
kenv -u hw.nic_uio.bdfs

# To update the list of ports to bind:
kenv hw.nic_uio.bdfs="b:d:f,b:d:f,..."

kldload <original_driver>

kldload nic_uio  # optional

Compiling and Running Sample Applications

The chapter describes how to compile and run applications in a DPDK environment. It also provides a pointer to where sample applications are stored.

Compiling a Sample Application

Once a DPDK target environment directory has been created (such as x86_64-native-bsdapp-clang), it contains all libraries and header files required to build an application.

When compiling an application in the FreeBSD environment on the DPDK, the following variables must be exported:

  • RTE_SDK - Points to the DPDK installation directory.
  • RTE_TARGET - Points to the DPDK target environment directory. For FreeBSD, this is the x86_64-native-bsdapp-clang or x86_64-native-bsdapp-gcc directory.

The following is an example of creating the helloworld application, which runs in the DPDK FreeBSD environment. While the example demonstrates compiling using gcc version 4.9, compiling with clang will be similar, except that the CC= parameter can probably be omitted. The helloworld example may be found in the ${RTE_SDK}/examples directory.

The directory contains the main.c file. This file, when combined with the libraries in the DPDK target environment, calls the various functions to initialize the DPDK environment, then launches an entry point (dispatch application) for each core to be utilized. By default, the binary is generated in the build directory.

setenv RTE_SDK /home/user/DPDK
cd $(RTE_SDK)
cd examples/helloworld/
setenv RTE_SDK $HOME/DPDK
setenv RTE_TARGET x86_64-native-bsdapp-gcc

gmake CC=gcc49
  CC main.o
  LD helloworld
  INSTALL-APP helloworld
  INSTALL-MAP helloworld.map

ls build/app
  helloworld helloworld.map

Note

In the above example, helloworld was in the directory structure of the DPDK. However, it could have been located outside the directory structure to keep the DPDK structure intact. In the following case, the helloworld application is copied to a new directory as a new starting point.

setenv RTE_SDK /home/user/DPDK
cp -r $(RTE_SDK)/examples/helloworld my_rte_app
cd my_rte_app/
setenv RTE_TARGET x86_64-native-bsdapp-gcc

gmake CC=gcc49
  CC main.o
  LD helloworld
  INSTALL-APP helloworld
  INSTALL-MAP helloworld.map

Running a Sample Application

  1. The contigmem and nic_uio modules must be set up prior to running an application.
  2. Any ports to be used by the application must be already bound to the nic_uio module, as described in section Binding Network Ports to the nic_uio Module, prior to running the application. The application is linked with the DPDK target environment’s Environment Abstraction Layer (EAL) library, which provides some options that are generic to every DPDK application.

The following is the list of options that can be given to the EAL:

./rte-app -l CORELIST [-n NUM] [-b <domain:bus:devid.func>] \
          [-r NUM] [-v] [--proc-type <primary|secondary|auto>]

Note

EAL has a common interface between all operating systems and is based on the Linux notation for PCI devices. For example, a FreeBSD device selector of pci0:2:0:1 is referred to as 02:00.1 in EAL.

The EAL options for FreeBSD are as follows:

  • -c COREMASK or -l CORELIST: A hexadecimal bit mask of the cores to run on. Note that core numbering can change between platforms and should be determined beforehand. The corelist is a list of cores to use instead of a core mask.
  • -n NUM: Number of memory channels per processor socket.
  • -b <domain:bus:devid.func>: Blacklisting of ports; prevent EAL from using specified PCI device (multiple -b options are allowed).
  • --use-device: Use the specified Ethernet device(s) only. Use comma-separate [domain:]bus:devid.func values. Cannot be used with -b option.
  • -r NUM: Number of memory ranks.
  • -v: Display version information on startup.
  • --proc-type: The type of process instance.

Other options, specific to Linux and are not supported under FreeBSD are as follows:

  • socket-mem: Memory to allocate from hugepages on specific sockets.
  • --huge-dir: The directory where hugetlbfs is mounted.
  • mbuf-pool-ops-name: Pool ops name for mbuf to use.
  • --file-prefix: The prefix text used for hugepage filenames.
  • -m MB: Memory to allocate from hugepages, regardless of processor socket. It is recommended that --socket-mem be used instead of this option.

The -c or -l option is mandatory; the others are optional.

Copy the DPDK application binary to your target, then run the application as follows (assuming the platform has four memory channels, and that cores 0-3 are present and are to be used for running the application):

./helloworld -l 0-3 -n 4

Note

The --proc-type and --file-prefix EAL options are used for running multiple DPDK processes. See the “Multi-process Sample Application” chapter in the DPDK Sample Applications User Guide and the DPDK Programmers Guide for more details.

Running DPDK Applications Without Root Privileges

Although applications using the DPDK use network ports and other hardware resources directly, with a number of small permission adjustments, it is possible to run these applications as a user other than “root”. To do so, the ownership, or permissions, on the following file system objects should be adjusted to ensure that the user account being used to run the DPDK application has access to them:

  • The userspace-io device files in /dev, for example, /dev/uio0, /dev/uio1, and so on
  • The userspace contiguous memory device: /dev/contigmem

Note

Please refer to the DPDK Release Notes for supported applications.

Sample Applications User Guides

Introduction to the DPDK Sample Applications

The DPDK Sample Applications are small standalone applications which demonstrate various features of DPDK. They can be considered as a cookbook of DPDK features. Users interested in getting started with DPDK can take the applications, try out the features, and then extend them to fit their needs.

The DPDK Sample Applications

Table Table 1.1 shows a list of some of the main sample applications that are available in the examples directory of DPDK:

Some of the DPDK Sample applications
Bonding Netmap Compatibility
Command Line Packet Ordering
Distributor Performance Thread
Ethtool Precision Time Protocol (PTP) Client
Exception Path Quality of Service (QoS) Metering
Hello World QoS Scheduler
Internet Protocol (IP) Fragmentation Quota and Watermark
IP Pipeline RX/TX Callbacks
IP Reassembly Server node EFD
IPsec Security Gateway Basic Forwarding/Skeleton App
IPv4 multicast Tunnel End Point (TEP) termination
Kernel NIC Interface Timer
Network Layer 2 Forwarding + variants Vhost
Network Layer 3 Forwarding + variants Vhost Xen
Link Status Interrupt VMDQ Forwarding
Load Balancer VMDQ and DCB Forwarding
Multi-process VM Power Management

These examples range from simple to reasonably complex but most are designed to demonstrate one particular feature of DPDK. Some of the more interesting examples are highlighted below.

  • Hello World: As with most introductions to a programming framework a good place to start is with the Hello World application. The Hello World example sets up the DPDK Environment Abstraction Layer (EAL), and prints a simple “Hello World” message to each of the DPDK enabled cores. This application doesn’t do any packet forwarding but it is a good way to test if the DPDK environment is compiled and set up properly.
  • Basic Forwarding/Skeleton Application: The Basic Forwarding/Skeleton contains the minimum amount of code required to enable basic packet forwarding with DPDK. This allows you to test if your network interfaces are working with DPDK.
  • Network Layer 2 forwarding: The Network Layer 2 forwarding, or l2fwd application does forwarding based on Ethernet MAC addresses like a simple switch.
  • Network Layer 3 forwarding: The Network Layer3 forwarding, or l3fwd application does forwarding based on Internet Protocol, IPv4 or IPv6 like a simple router.
  • Packet Distributor: The Packet Distributor demonstrates how to distribute packets arriving on an Rx port to different cores for processing and transmission.
  • Multi-Process Application: The multi-process application shows how two DPDK processes can work together using queues and memory pools to share information.
  • RX/TX callbacks Application: The RX/TX callbacks sample application is a packet forwarding application that demonstrates the use of user defined callbacks on received and transmitted packets. The application calculates the latency of a packet between RX (packet arrival) and TX (packet transmission) by adding callbacks to the RX and TX packet processing functions.
  • IPSec Security Gateway: The IPSec Security Gateway application is minimal example of something closer to a real world example. This is also a good example of an application using the DPDK Cryptodev framework.
  • Precision Time Protocol (PTP) client: The PTP client is another minimal implementation of a real world application. In this case the application is a PTP client that communicates with a PTP master clock to synchronize time on a Network Interface Card (NIC) using the IEEE1588 protocol.
  • Quality of Service (QoS) Scheduler: The QoS Scheduler application demonstrates the use of DPDK to provide QoS scheduling.

There are many more examples shown in the following chapters. Each of the documented sample applications show how to compile, configure and run the application as well as explaining the main functionality of the code.

Compiling the Sample Applications

This section explains how to compile the DPDK sample applications.

To compile all the sample applications

Set the path to DPDK source code if its not set:

export RTE_SDK=/path/to/rte_sdk

Go to DPDK source:

cd $RTE_SDK

Build DPDK:

make defconfig
make

Build the sample applications:

export RTE_TARGET=build
make -C examples

For other possible RTE_TARGET values and additional information on compiling see Compiling DPDK on Linux or Compiling DPDK on FreeBSD. Applications are output to: $RTE_SDK/examples/app-dir/build or $RTE_SDK/examples/app-dir/$RTE_TARGET.

In the example above the compiled application is written to the build subdirectory. To have the applications written to a different location, the O=/path/to/build/directory option may be specified in the make command.

make O=/tmp

To build the applications for debugging use the DEBUG option. This option adds some extra flags, disables compiler optimizations and sets verbose output.

make DEBUG=1

To compile a single application

Set the path to DPDK source code:

export RTE_SDK=/path/to/rte_sdk

Go to DPDK source:

cd $RTE_SDK

Build DPDK:

make defconfig
make

Go to the sample application directory. Unless otherwise specified the sample applications are located in $RTE_SDK/examples/.

Build the application:

export RTE_TARGET=build
make

Command Line Sample Application

This chapter describes the Command Line sample application that is part of the Data Plane Development Kit (DPDK).

Overview

The Command Line sample application is a simple application that demonstrates the use of the command line interface in the DPDK. This application is a readline-like interface that can be used to debug a DPDK application, in a Linux* application environment.

Note

The rte_cmdline library should not be used in production code since it is not validated to the same standard as other DPDK libraries. See also the “rte_cmdline library should not be used in production code due to limited testing” item in the “Known Issues” section of the Release Notes.

The Command Line sample application supports some of the features of the GNU readline library such as, completion, cut/paste and some other special bindings that make configuration and debug faster and easier.

The application shows how the rte_cmdline application can be extended to handle a list of objects. There are three simple commands:

  • add obj_name IP: Add a new object with an IP/IPv6 address associated to it.
  • del obj_name: Delete the specified object.
  • show obj_name: Show the IP associated with the specified object.

Note

To terminate the application, use Ctrl-d.

Compiling the Application

To compile the sample application see Compiling the Sample Applications

The application is located in the cmd_line sub-directory.

Running the Application

To run the application in linuxapp environment, issue the following command:

$ ./build/cmdline -l 0-3 -n 4

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The following sections provide some explanation of the code.

EAL Initialization and cmdline Start

The first task is the initialization of the Environment Abstraction Layer (EAL). This is achieved as follows:

int main(int argc, char **argv)
{
    ret = rte_eal_init(argc, argv);
    if (ret < 0)
        rte_panic("Cannot init EAL\n");

Then, a new command line object is created and started to interact with the user through the console:

cl = cmdline_stdin_new(main_ctx, "example> ");
cmdline_interact(cl);
cmdline_stdin_exit(cl);

The cmd line_interact() function returns when the user types Ctrl-d and in this case, the application exits.

Defining a cmdline Context

A cmdline context is a list of commands that are listed in a NULL-terminated table, for example:

cmdline_parse_ctx_t main_ctx[] = {
    (cmdline_parse_inst_t *) &cmd_obj_del_show,
    (cmdline_parse_inst_t *) &cmd_obj_add,
    (cmdline_parse_inst_t *) &cmd_help,
     NULL,
};

Each command (of type cmdline_parse_inst_t) is defined statically. It contains a pointer to a callback function that is executed when the command is parsed, an opaque pointer, a help string and a list of tokens in a NULL-terminated table.

The rte_cmdline application provides a list of pre-defined token types:

  • String Token: Match a static string, a list of static strings or any string.
  • Number Token: Match a number that can be signed or unsigned, from 8-bit to 32-bit.
  • IP Address Token: Match an IPv4 or IPv6 address or network.
  • Ethernet* Address Token: Match a MAC address.

In this example, a new token type obj_list is defined and implemented in the parse_obj_list.c and parse_obj_list.h files.

For example, the cmd_obj_del_show command is defined as shown below:

struct cmd_obj_add_result {
    cmdline_fixed_string_t action;
    cmdline_fixed_string_t name;
    struct object *obj;
};

static void cmd_obj_del_show_parsed(void *parsed_result, struct cmdline *cl, attribute ((unused)) void *data)
{
   /* ... */
}

cmdline_parse_token_string_t cmd_obj_action = TOKEN_STRING_INITIALIZER(struct cmd_obj_del_show_result, action, "show#del");

parse_token_obj_list_t cmd_obj_obj = TOKEN_OBJ_LIST_INITIALIZER(struct cmd_obj_del_show_result, obj, &global_obj_list);

cmdline_parse_inst_t cmd_obj_del_show = {
    .f = cmd_obj_del_show_parsed, /* function to call */
    .data = NULL,  /* 2nd arg of func */
    .help_str = "Show/del an object",
    .tokens = { /* token list, NULL terminated */
        (void *)&cmd_obj_action,
        (void *)&cmd_obj_obj,
         NULL,
    },
};

This command is composed of two tokens:

  • The first token is a string token that can be show or del.
  • The second token is an object that was previously added using the add command in the global_obj_list variable.

Once the command is parsed, the rte_cmdline application fills a cmd_obj_del_show_result structure. A pointer to this structure is given as an argument to the callback function and can be used in the body of this function.

Ethtool Sample Application

The Ethtool sample application shows an implementation of an ethtool-like API and provides a console environment that allows its use to query and change Ethernet card parameters. The sample is based upon a simple L2 frame reflector.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the ethtool sub-directory.

Running the Application

The application requires an available core for each port, plus one. The only available options are the standard ones for the EAL:

./ethtool-app/ethtool-app/${RTE_TARGET}/ethtool [EAL options]

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Using the application

The application is console-driven using the cmdline DPDK interface:

EthApp>

From this interface the available commands and descriptions of what they do as as follows:

  • drvinfo: Print driver info
  • eeprom: Dump EEPROM to file
  • link: Print port link states
  • macaddr: Gets/sets MAC address
  • mtu: Set NIC MTU
  • open: Open port
  • pause: Get/set port pause state
  • portstats: Print port statistics
  • regs: Dump port register(s) to file
  • ringparam: Get/set ring parameters
  • rxmode: Toggle port Rx mode
  • stop: Stop port
  • validate: Check that given MAC address is valid unicast address
  • vlan: Add/remove VLAN id
  • quit: Exit program

Explanation

The sample program has two parts: A background packet reflector that runs on a slave core, and a foreground Ethtool Shell that runs on the master core. These are described below.

Packet Reflector

The background packet reflector is intended to demonstrate basic packet processing on NIC ports controlled by the Ethtool shim. Each incoming MAC frame is rewritten so that it is returned to the sender, using the port in question’s own MAC address as the source address, and is then sent out on the same port.

Ethtool Shell

The foreground part of the Ethtool sample is a console-based interface that accepts commands as described in using the application. Individual call-back functions handle the detail associated with each command, which make use of the functions defined in the Ethtool interface to the DPDK functions.

Ethtool interface

The Ethtool interface is built as a separate library, and implements the following functions:

  • rte_ethtool_get_drvinfo()
  • rte_ethtool_get_regs_len()
  • rte_ethtool_get_regs()
  • rte_ethtool_get_link()
  • rte_ethtool_get_eeprom_len()
  • rte_ethtool_get_eeprom()
  • rte_ethtool_set_eeprom()
  • rte_ethtool_get_pauseparam()
  • rte_ethtool_set_pauseparam()
  • rte_ethtool_net_open()
  • rte_ethtool_net_stop()
  • rte_ethtool_net_get_mac_addr()
  • rte_ethtool_net_set_mac_addr()
  • rte_ethtool_net_validate_addr()
  • rte_ethtool_net_change_mtu()
  • rte_ethtool_net_get_stats64()
  • rte_ethtool_net_vlan_rx_add_vid()
  • rte_ethtool_net_vlan_rx_kill_vid()
  • rte_ethtool_net_set_rx_mode()
  • rte_ethtool_get_ringparam()
  • rte_ethtool_set_ringparam()

Exception Path Sample Application

The Exception Path sample application is a simple example that demonstrates the use of the DPDK to set up an exception path for packets to go through the Linux* kernel. This is done by using virtual TAP network interfaces. These can be read from and written to by the DPDK application and appear to the kernel as a standard network interface.

Overview

The application creates two threads for each NIC port being used. One thread reads from the port and writes the data unmodified to a thread-specific TAP interface. The second thread reads from a TAP interface and writes the data unmodified to the NIC port.

The packet flow through the exception path application is as shown in the following figure.

_images/exception_path_example.svg

Packet Flow

To make throughput measurements, kernel bridges must be setup to forward data between the bridges appropriately.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the exception_path sub-directory.

Running the Application

The application requires a number of command line options:

.build/exception_path [EAL options] -- -p PORTMASK -i IN_CORES -o OUT_CORES

where:

  • -p PORTMASK: A hex bitmask of ports to use
  • -i IN_CORES: A hex bitmask of cores which read from NIC
  • -o OUT_CORES: A hex bitmask of cores which write to NIC

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

The number of bits set in each bitmask must be the same. The coremask -c or the corelist -l parameter of the EAL options should include IN_CORES and OUT_CORES. The same bit must not be set in IN_CORES and OUT_CORES. The affinities between ports and cores are set beginning with the least significant bit of each mask, that is, the port represented by the lowest bit in PORTMASK is read from by the core represented by the lowest bit in IN_CORES, and written to by the core represented by the lowest bit in OUT_CORES.

For example to run the application with two ports and four cores:

./build/exception_path -l 0-3 -n 4 -- -p 3 -i 3 -o c
Getting Statistics

While the application is running, statistics on packets sent and received can be displayed by sending the SIGUSR1 signal to the application from another terminal:

killall -USR1 exception_path

The statistics can be reset by sending a SIGUSR2 signal in a similar way.

Explanation

The following sections provide some explanation of the code.

Initialization

Setup of the mbuf pool, driver and queues is similar to the setup done in the L2 Forwarding Sample Application (in Real and Virtualized Environments). In addition, the TAP interfaces must also be created. A TAP interface is created for each lcore that is being used. The code for creating the TAP interface is as follows:

/*
 *   Create a tap network interface, or use existing one with same name.
 *   If name[0]='\0' then a name is automatically assigned and returned in name.
 */

static int tap_create(char *name)
{
    struct ifreq ifr;
    int fd, ret;

    fd = open("/dev/net/tun", O_RDWR);
    if (fd < 0)
        return fd;

    memset(&ifr, 0, sizeof(ifr));

    /* TAP device without packet information */

    ifr.ifr_flags = IFF_TAP | IFF_NO_PI;
    if (name && *name)
        rte_snprinf(ifr.ifr_name, IFNAMSIZ, name);

    ret = ioctl(fd, TUNSETIFF, (void *) &ifr);

    if (ret < 0) {
        close(fd);
        return ret;

    }

    if (name)
        snprintf(name, IFNAMSIZ, ifr.ifr_name);

    return fd;
}

The other step in the initialization process that is unique to this sample application is the association of each port with two cores:

  • One core to read from the port and write to a TAP interface
  • A second core to read from a TAP interface and write to the port

This is done using an array called port_ids[], which is indexed by the lcore IDs. The population of this array is shown below:

 tx_port = 0;
 rx_port = 0;

 RTE_LCORE_FOREACH(i) {
     if (input_cores_mask & (1ULL << i)) {
         /* Skip ports that are not enabled */
         while ((ports_mask & (1 << rx_port)) == 0) {
             rx_port++;
             if (rx_port > (sizeof(ports_mask) * 8))
                 goto fail; /* not enough ports */
         }
         port_ids[i] = rx_port++;
     } else if (output_cores_mask & (1ULL << i)) {
         /* Skip ports that are not enabled */
         while ((ports_mask & (1 << tx_port)) == 0) {
             tx_port++;
             if (tx_port > (sizeof(ports_mask) * 8))
                goto fail; /* not enough ports */
         }
         port_ids[i] = tx_port++;
     }
}
Packet Forwarding

After the initialization steps are complete, the main_loop() function is run on each lcore. This function first checks the lcore_id against the user provided input_cores_mask and output_cores_mask to see if this core is reading from or writing to a TAP interface.

For the case that reads from a NIC port, the packet reception is the same as in the L2 Forwarding sample application (see Receive, Process and Transmit Packets). The packet transmission is done by calling write() with the file descriptor of the appropriate TAP interface and then explicitly freeing the mbuf back to the pool.

/* Loop forever reading from NIC and writing to tap */

for (;;) {
    struct rte_mbuf *pkts_burst[PKT_BURST_SZ];
    unsigned i;

    const unsigned nb_rx = rte_eth_rx_burst(port_ids[lcore_id], 0, pkts_burst, PKT_BURST_SZ);

    lcore_stats[lcore_id].rx += nb_rx;

    for (i = 0; likely(i < nb_rx); i++) {
        struct rte_mbuf *m = pkts_burst[i];
        int ret = write(tap_fd, rte_pktmbuf_mtod(m, void*),

        rte_pktmbuf_data_len(m));
        rte_pktmbuf_free(m);
        if (unlikely(ret<0))
            lcore_stats[lcore_id].dropped++;
        else
            lcore_stats[lcore_id].tx++;
    }
}

For the other case that reads from a TAP interface and writes to a NIC port, packets are retrieved by doing a read() from the file descriptor of the appropriate TAP interface. This fills in the data into the mbuf, then other fields are set manually. The packet can then be transmitted as normal.

/* Loop forever reading from tap and writing to NIC */

for (;;) {
    int ret;
    struct rte_mbuf *m = rte_pktmbuf_alloc(pktmbuf_pool);

    if (m == NULL)
        continue;

    ret = read(tap_fd, m->pkt.data, MAX_PACKET_SZ); lcore_stats[lcore_id].rx++;
    if (unlikely(ret < 0)) {
        FATAL_ERROR("Reading from %s interface failed", tap_name);
    }

    m->pkt.nb_segs = 1;
    m->pkt.next = NULL;
    m->pkt.data_len = (uint16_t)ret;

    ret = rte_eth_tx_burst(port_ids[lcore_id], 0, &m, 1);
    if (unlikely(ret < 1)) {
        rte_pktmuf_free(m);
        lcore_stats[lcore_id].dropped++;
    }
    else {
        lcore_stats[lcore_id].tx++;
    }
}

To set up loops for measuring throughput, TAP interfaces can be connected using bridging. The steps to do this are described in the section that follows.

Managing TAP Interfaces and Bridges

The Exception Path sample application creates TAP interfaces with names of the format tap_dpdk_nn, where nn is the lcore ID. These TAP interfaces need to be configured for use:

ifconfig tap_dpdk_00 up

To set up a bridge between two interfaces so that packets sent to one interface can be read from another, use the brctl tool:

brctl addbr "br0"
brctl addif br0 tap_dpdk_00
brctl addif br0 tap_dpdk_03
ifconfig br0 up

The TAP interfaces created by this application exist only when the application is running, so the steps above need to be repeated each time the application is run. To avoid this, persistent TAP interfaces can be created using openvpn:

openvpn --mktun --dev tap_dpdk_00

If this method is used, then the steps above have to be done only once and the same TAP interfaces can be reused each time the application is run. To remove bridges and persistent TAP interfaces, the following commands are used:

ifconfig br0 down
brctl delbr br0
openvpn --rmtun --dev tap_dpdk_00

Hello World Sample Application

The Hello World sample application is an example of the simplest DPDK application that can be written. The application simply prints an “helloworld” message on every enabled lcore.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the helloworld sub-directory.

Running the Application

To run the example in a linuxapp environment:

$ ./build/helloworld -l 0-3 -n 4

Refer to DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The following sections provide some explanation of code.

EAL Initialization

The first task is to initialize the Environment Abstraction Layer (EAL). This is done in the main() function using the following code:

int

main(int argc, char **argv)

{
    ret = rte_eal_init(argc, argv);
    if (ret < 0)
        rte_panic("Cannot init EAL\n");

This call finishes the initialization process that was started before main() is called (in case of a Linuxapp environment). The argc and argv arguments are provided to the rte_eal_init() function. The value returned is the number of parsed arguments.

Starting Application Unit Lcores

Once the EAL is initialized, the application is ready to launch a function on an lcore. In this example, lcore_hello() is called on every available lcore. The following is the definition of the function:

static int
lcore_hello( attribute ((unused)) void *arg)
{
    unsigned lcore_id;

    lcore_id = rte_lcore_id();
    printf("hello from core %u\n", lcore_id);
    return 0;
}

The code that launches the function on each lcore is as follows:

/* call lcore_hello() on every slave lcore */

RTE_LCORE_FOREACH_SLAVE(lcore_id) {
   rte_eal_remote_launch(lcore_hello, NULL, lcore_id);
}

/* call it on master lcore too */

lcore_hello(NULL);

The following code is equivalent and simpler:

rte_eal_mp_remote_launch(lcore_hello, NULL, CALL_MASTER);

Refer to the DPDK API Reference for detailed information on the rte_eal_mp_remote_launch() function.

Basic Forwarding Sample Application

The Basic Forwarding sample application is a simple skeleton example of a forwarding application.

It is intended as a demonstration of the basic components of a DPDK forwarding application. For more detailed implementations see the L2 and L3 forwarding sample applications.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the skeleton sub-directory.

Running the Application

To run the example in a linuxapp environment:

./build/basicfwd -l 1 -n 4

Refer to DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The following sections provide an explanation of the main components of the code.

All DPDK library functions used in the sample code are prefixed with rte_ and are explained in detail in the DPDK API Documentation.

The Main Function

The main() function performs the initialization and calls the execution threads for each lcore.

The first task is to initialize the Environment Abstraction Layer (EAL). The argc and argv arguments are provided to the rte_eal_init() function. The value returned is the number of parsed arguments:

int ret = rte_eal_init(argc, argv);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Error with EAL initialization\n");

The main() also allocates a mempool to hold the mbufs (Message Buffers) used by the application:

mbuf_pool = rte_mempool_create("MBUF_POOL",
                               NUM_MBUFS * nb_ports,
                               MBUF_SIZE,
                               MBUF_CACHE_SIZE,
                               sizeof(struct rte_pktmbuf_pool_private),
                               rte_pktmbuf_pool_init, NULL,
                               rte_pktmbuf_init,      NULL,
                               rte_socket_id(),
                               0);

Mbufs are the packet buffer structure used by DPDK. They are explained in detail in the “Mbuf Library” section of the DPDK Programmer’s Guide.

The main() function also initializes all the ports using the user defined port_init() function which is explained in the next section:

for (portid = 0; portid < nb_ports; portid++) {
    if (port_init(portid, mbuf_pool) != 0) {
        rte_exit(EXIT_FAILURE,
                 "Cannot init port %" PRIu8 "\n", portid);
    }
}

Once the initialization is complete, the application is ready to launch a function on an lcore. In this example lcore_main() is called on a single lcore.

lcore_main();

The lcore_main() function is explained below.

The Port Initialization Function

The main functional part of the port initialization used in the Basic Forwarding application is shown below:

static inline int
port_init(uint16_t port, struct rte_mempool *mbuf_pool)
{
    struct rte_eth_conf port_conf = port_conf_default;
    const uint16_t rx_rings = 1, tx_rings = 1;
    struct ether_addr addr;
    int retval;
    uint16_t q;

    if (port >= rte_eth_dev_count())
        return -1;

    /* Configure the Ethernet device. */
    retval = rte_eth_dev_configure(port, rx_rings, tx_rings, &port_conf);
    if (retval != 0)
        return retval;

    /* Allocate and set up 1 RX queue per Ethernet port. */
    for (q = 0; q < rx_rings; q++) {
        retval = rte_eth_rx_queue_setup(port, q, RX_RING_SIZE,
                rte_eth_dev_socket_id(port), NULL, mbuf_pool);
        if (retval < 0)
            return retval;
    }

    /* Allocate and set up 1 TX queue per Ethernet port. */
    for (q = 0; q < tx_rings; q++) {
        retval = rte_eth_tx_queue_setup(port, q, TX_RING_SIZE,
                rte_eth_dev_socket_id(port), NULL);
        if (retval < 0)
            return retval;
    }

    /* Start the Ethernet port. */
    retval = rte_eth_dev_start(port);
    if (retval < 0)
        return retval;

    /* Enable RX in promiscuous mode for the Ethernet device. */
    rte_eth_promiscuous_enable(port);

    return 0;
}

The Ethernet ports are configured with default settings using the rte_eth_dev_configure() function and the port_conf_default struct:

static const struct rte_eth_conf port_conf_default = {
    .rxmode = { .max_rx_pkt_len = ETHER_MAX_LEN }
};

For this example the ports are set up with 1 RX and 1 TX queue using the rte_eth_rx_queue_setup() and rte_eth_tx_queue_setup() functions.

The Ethernet port is then started:

retval  = rte_eth_dev_start(port);

Finally the RX port is set in promiscuous mode:

rte_eth_promiscuous_enable(port);
The Lcores Main

As we saw above the main() function calls an application function on the available lcores. For the Basic Forwarding application the lcore function looks like the following:

static __attribute__((noreturn)) void
lcore_main(void)
{
    const uint16_t nb_ports = rte_eth_dev_count();
    uint16_t port;

    /*
     * Check that the port is on the same NUMA node as the polling thread
     * for best performance.
     */
    for (port = 0; port < nb_ports; port++)
        if (rte_eth_dev_socket_id(port) > 0 &&
                rte_eth_dev_socket_id(port) !=
                        (int)rte_socket_id())
            printf("WARNING, port %u is on remote NUMA node to "
                    "polling thread.\n\tPerformance will "
                    "not be optimal.\n", port);

    printf("\nCore %u forwarding packets. [Ctrl+C to quit]\n",
            rte_lcore_id());

    /* Run until the application is quit or killed. */
    for (;;) {
        /*
         * Receive packets on a port and forward them on the paired
         * port. The mapping is 0 -> 1, 1 -> 0, 2 -> 3, 3 -> 2, etc.
         */
        for (port = 0; port < nb_ports; port++) {

            /* Get burst of RX packets, from first port of pair. */
            struct rte_mbuf *bufs[BURST_SIZE];
            const uint16_t nb_rx = rte_eth_rx_burst(port, 0,
                    bufs, BURST_SIZE);

            if (unlikely(nb_rx == 0))
                continue;

            /* Send burst of TX packets, to second port of pair. */
            const uint16_t nb_tx = rte_eth_tx_burst(port ^ 1, 0,
                    bufs, nb_rx);

            /* Free any unsent packets. */
            if (unlikely(nb_tx < nb_rx)) {
                uint16_t buf;
                for (buf = nb_tx; buf < nb_rx; buf++)
                    rte_pktmbuf_free(bufs[buf]);
            }
        }
    }
}

The main work of the application is done within the loop:

for (;;) {
    for (port = 0; port < nb_ports; port++) {

        /* Get burst of RX packets, from first port of pair. */
        struct rte_mbuf *bufs[BURST_SIZE];
        const uint16_t nb_rx = rte_eth_rx_burst(port, 0,
                bufs, BURST_SIZE);

        if (unlikely(nb_rx == 0))
            continue;

        /* Send burst of TX packets, to second port of pair. */
        const uint16_t nb_tx = rte_eth_tx_burst(port ^ 1, 0,
                bufs, nb_rx);

        /* Free any unsent packets. */
        if (unlikely(nb_tx < nb_rx)) {
            uint16_t buf;
            for (buf = nb_tx; buf < nb_rx; buf++)
                rte_pktmbuf_free(bufs[buf]);
        }
    }
}

Packets are received in bursts on the RX ports and transmitted in bursts on the TX ports. The ports are grouped in pairs with a simple mapping scheme using the an XOR on the port number:

0 -> 1
1 -> 0

2 -> 3
3 -> 2

etc.

The rte_eth_tx_burst() function frees the memory buffers of packets that are transmitted. If packets fail to transmit, (nb_tx < nb_rx), then they must be freed explicitly using rte_pktmbuf_free().

The forwarding loop can be interrupted and the application closed using Ctrl-C.

RX/TX Callbacks Sample Application

The RX/TX Callbacks sample application is a packet forwarding application that demonstrates the use of user defined callbacks on received and transmitted packets. The application performs a simple latency check, using callbacks, to determine the time packets spend within the application.

In the sample application a user defined callback is applied to all received packets to add a timestamp. A separate callback is applied to all packets prior to transmission to calculate the elapsed time, in CPU cycles.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the rxtx_callbacks sub-directory.

The callbacks feature requires that the CONFIG_RTE_ETHDEV_RXTX_CALLBACKS setting is on in the config/common_ config file that applies to the target. This is generally on by default:

CONFIG_RTE_ETHDEV_RXTX_CALLBACKS=y

Running the Application

To run the example in a linuxapp environment:

./build/rxtx_callbacks -l 1 -n 4

Refer to DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The rxtx_callbacks application is mainly a simple forwarding application based on the Basic Forwarding Sample Application. See that section of the documentation for more details of the forwarding part of the application.

The sections below explain the additional RX/TX callback code.

The Main Function

The main() function performs the application initialization and calls the execution threads for each lcore. This function is effectively identical to the main() function explained in Basic Forwarding Sample Application.

The lcore_main() function is also identical.

The main difference is in the user defined port_init() function where the callbacks are added. This is explained in the next section:

The Port Initialization Function

The main functional part of the port initialization is shown below with comments:

static inline int
port_init(uint16_t port, struct rte_mempool *mbuf_pool)
{
    struct rte_eth_conf port_conf = port_conf_default;
    const uint16_t rx_rings = 1, tx_rings = 1;
    struct ether_addr addr;
    int retval;
    uint16_t q;

    if (port >= rte_eth_dev_count())
        return -1;

    /* Configure the Ethernet device. */
    retval = rte_eth_dev_configure(port, rx_rings, tx_rings, &port_conf);
    if (retval != 0)
        return retval;

    /* Allocate and set up 1 RX queue per Ethernet port. */
    for (q = 0; q < rx_rings; q++) {
        retval = rte_eth_rx_queue_setup(port, q, RX_RING_SIZE,
                rte_eth_dev_socket_id(port), NULL, mbuf_pool);
        if (retval < 0)
            return retval;
    }

    /* Allocate and set up 1 TX queue per Ethernet port. */
    for (q = 0; q < tx_rings; q++) {
        retval = rte_eth_tx_queue_setup(port, q, TX_RING_SIZE,
                rte_eth_dev_socket_id(port), NULL);
        if (retval < 0)
            return retval;
    }

    /* Start the Ethernet port. */
    retval = rte_eth_dev_start(port);
    if (retval < 0)
        return retval;

    /* Enable RX in promiscuous mode for the Ethernet device. */
    rte_eth_promiscuous_enable(port);


    /* Add the callbacks for RX and TX.*/
    rte_eth_add_rx_callback(port, 0, add_timestamps, NULL);
    rte_eth_add_tx_callback(port, 0, calc_latency, NULL);

    return 0;
}

The RX and TX callbacks are added to the ports/queues as function pointers:

rte_eth_add_rx_callback(port, 0, add_timestamps, NULL);
rte_eth_add_tx_callback(port, 0, calc_latency,   NULL);

More than one callback can be added and additional information can be passed to callback function pointers as a void*. In the examples above NULL is used.

The add_timestamps() and calc_latency() functions are explained below.

The add_timestamps() Callback

The add_timestamps() callback is added to the RX port and is applied to all packets received:

static uint16_t
add_timestamps(uint16_t port __rte_unused, uint16_t qidx __rte_unused,
        struct rte_mbuf **pkts, uint16_t nb_pkts, void *_ __rte_unused)
{
    unsigned i;
    uint64_t now = rte_rdtsc();

    for (i = 0; i < nb_pkts; i++)
        pkts[i]->udata64 = now;

    return nb_pkts;
}

The DPDK function rte_rdtsc() is used to add a cycle count timestamp to each packet (see the cycles section of the DPDK API Documentation for details).

The calc_latency() Callback

The calc_latency() callback is added to the TX port and is applied to all packets prior to transmission:

static uint16_t
calc_latency(uint16_t port __rte_unused, uint16_t qidx __rte_unused,
        struct rte_mbuf **pkts, uint16_t nb_pkts, void *_ __rte_unused)
{
    uint64_t cycles = 0;
    uint64_t now = rte_rdtsc();
    unsigned i;

    for (i = 0; i < nb_pkts; i++)
        cycles += now - pkts[i]->udata64;

    latency_numbers.total_cycles += cycles;
    latency_numbers.total_pkts   += nb_pkts;

    if (latency_numbers.total_pkts > (100 * 1000 * 1000ULL)) {
        printf("Latency = %"PRIu64" cycles\n",
                latency_numbers.total_cycles / latency_numbers.total_pkts);

        latency_numbers.total_cycles = latency_numbers.total_pkts = 0;
    }

    return nb_pkts;
}

The calc_latency() function accumulates the total number of packets and the total number of cycles used. Once more than 100 million packets have been transmitted the average cycle count per packet is printed out and the counters are reset.

Flow Classify Sample Application

The Flow Classify sample application is based on the simple skeleton example of a forwarding application.

It is intended as a demonstration of the basic components of a DPDK forwarding application which uses the Flow Classify library API’s.

Please refer to the Flow Classification Library for more information.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the flow_classify sub-directory.

Running the Application

To run the example in a linuxapp environment:

cd ~/dpdk/examples/flow_classify
./build/flow_classify -c 4 -n 4 -- --rule_ipv4="../ipv4_rules_file.txt"

Please refer to the DPDK Getting Started Guide, section Compiling and Running Sample Applications for general information on running applications and the Environment Abstraction Layer (EAL) options.

Sample ipv4_rules_file.txt

#file format:
#src_ip/masklen dst_ip/masklen src_port : mask dst_port : mask proto/mask priority
#
2.2.2.3/24 2.2.2.7/24 32 : 0xffff 33 : 0xffff 17/0xff 0
9.9.9.3/24 9.9.9.7/24 32 : 0xffff 33 : 0xffff 17/0xff 1
9.9.9.3/24 9.9.9.7/24 32 : 0xffff 33 : 0xffff 6/0xff 2
9.9.8.3/24 9.9.8.7/24 32 : 0xffff 33 : 0xffff 6/0xff 3
6.7.8.9/24 2.3.4.5/24 32 : 0x0000 33 : 0x0000 132/0xff 4

Explanation

The following sections provide an explanation of the main components of the code.

All DPDK library functions used in the sample code are prefixed with rte_ and are explained in detail in the DPDK API Documentation.

ACL field definitions for the IPv4 5 tuple rule

The following field definitions are used when creating the ACL table during initialisation of the Flow Classify application..

 enum {
     PROTO_FIELD_IPV4,
     SRC_FIELD_IPV4,
     DST_FIELD_IPV4,
     SRCP_FIELD_IPV4,
     DSTP_FIELD_IPV4,
     NUM_FIELDS_IPV4
};

enum {
    PROTO_INPUT_IPV4,
    SRC_INPUT_IPV4,
    DST_INPUT_IPV4,
    SRCP_DESTP_INPUT_IPV4
};

static struct rte_acl_field_def ipv4_defs[NUM_FIELDS_IPV4] = {
    /* first input field - always one byte long. */
    {
        .type = RTE_ACL_FIELD_TYPE_BITMASK,
        .size = sizeof(uint8_t),
        .field_index = PROTO_FIELD_IPV4,
        .input_index = PROTO_INPUT_IPV4,
        .offset = sizeof(struct ether_hdr) +
            offsetof(struct ipv4_hdr, next_proto_id),
    },
    /* next input field (IPv4 source address) - 4 consecutive bytes. */
    {
        /* rte_flow uses a bit mask for IPv4 addresses */
        .type = RTE_ACL_FIELD_TYPE_BITMASK,
        .size = sizeof(uint32_t),
        .field_index = SRC_FIELD_IPV4,
        .input_index = SRC_INPUT_IPV4,
        .offset = sizeof(struct ether_hdr) +
            offsetof(struct ipv4_hdr, src_addr),
    },
    /* next input field (IPv4 destination address) - 4 consecutive bytes. */
    {
        /* rte_flow uses a bit mask for IPv4 addresses */
        .type = RTE_ACL_FIELD_TYPE_BITMASK,
        .size = sizeof(uint32_t),
        .field_index = DST_FIELD_IPV4,
        .input_index = DST_INPUT_IPV4,
        .offset = sizeof(struct ether_hdr) +
            offsetof(struct ipv4_hdr, dst_addr),
    },
    /*
     * Next 2 fields (src & dst ports) form 4 consecutive bytes.
     * They share the same input index.
     */
    {
        /* rte_flow uses a bit mask for protocol ports */
        .type = RTE_ACL_FIELD_TYPE_BITMASK,
        .size = sizeof(uint16_t),
        .field_index = SRCP_FIELD_IPV4,
        .input_index = SRCP_DESTP_INPUT_IPV4,
        .offset = sizeof(struct ether_hdr) +
            sizeof(struct ipv4_hdr) +
            offsetof(struct tcp_hdr, src_port),
    },
    {
         /* rte_flow uses a bit mask for protocol ports */
         .type = RTE_ACL_FIELD_TYPE_BITMASK,
         .size = sizeof(uint16_t),
         .field_index = DSTP_FIELD_IPV4,
         .input_index = SRCP_DESTP_INPUT_IPV4,
         .offset = sizeof(struct ether_hdr) +
             sizeof(struct ipv4_hdr) +
             offsetof(struct tcp_hdr, dst_port),
    },
};
The Main Function

The main() function performs the initialization and calls the execution threads for each lcore.

The first task is to initialize the Environment Abstraction Layer (EAL). The argc and argv arguments are provided to the rte_eal_init() function. The value returned is the number of parsed arguments:

int ret = rte_eal_init(argc, argv);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Error with EAL initialization\n");

It then parses the flow_classify application arguments

ret = parse_args(argc, argv);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Invalid flow_classify parameters\n");

The main() function also allocates a mempool to hold the mbufs (Message Buffers) used by the application:

mbuf_pool = rte_mempool_create("MBUF_POOL",
                               NUM_MBUFS * nb_ports,
                               MBUF_SIZE,
                               MBUF_CACHE_SIZE,
                               sizeof(struct rte_pktmbuf_pool_private),
                               rte_pktmbuf_pool_init, NULL,
                               rte_pktmbuf_init, NULL,
                               rte_socket_id(),
                               0);

mbufs are the packet buffer structure used by DPDK. They are explained in detail in the “Mbuf Library” section of the DPDK Programmer’s Guide.

The main() function also initializes all the ports using the user defined port_init() function which is explained in the next section:

for (portid = 0; portid < nb_ports; portid++) {
    if (port_init(portid, mbuf_pool) != 0) {
        rte_exit(EXIT_FAILURE,
                 "Cannot init port %" PRIu8 "\n", portid);
    }
}

The main() function creates the flow classifier object and adds an ACL table to the flow classifier.

struct flow_classifier {
    struct rte_flow_classifier *cls;
    uint32_t table_id[RTE_FLOW_CLASSIFY_TABLE_MAX];
};

struct flow_classifier_acl {
    struct flow_classifier cls;
} __rte_cache_aligned;

/* Memory allocation */
size = RTE_CACHE_LINE_ROUNDUP(sizeof(struct flow_classifier_acl));
cls_app = rte_zmalloc(NULL, size, RTE_CACHE_LINE_SIZE);
if (cls_app == NULL)
    rte_exit(EXIT_FAILURE, "Cannot allocate classifier memory\n");

cls_params.name = "flow_classifier";
cls_params.socket_id = socket_id;
cls_params.type = RTE_FLOW_CLASSIFY_TABLE_TYPE_ACL;

cls_app->cls = rte_flow_classifier_create(&cls_params);
if (cls_app->cls == NULL) {
    rte_free(cls_app);
    rte_exit(EXIT_FAILURE, "Cannot create classifier\n");
}

/* initialise ACL table params */
table_acl_params.name = "table_acl_ipv4_5tuple";
table_acl_params.n_rule_fields = RTE_DIM(ipv4_defs);
table_acl_params.n_rules = FLOW_CLASSIFY_MAX_RULE_NUM;
memcpy(table_acl_params.field_format, ipv4_defs, sizeof(ipv4_defs));

/* initialise table create params */
cls_table_params.ops = &rte_table_acl_ops,
cls_table_params.arg_create = &table_acl_params,
cls_table_params.table_metadata_size = 0;

ret = rte_flow_classify_table_create(cls_app->cls, &cls_table_params,
              &cls->table_id[0]);
if (ret) {
    rte_flow_classifier_free(cls_app->cls);
    rte_free(cls);
    rte_exit(EXIT_FAILURE, "Failed to create classifier table\n");
}

It then reads the ipv4_rules_file.txt file and initialises the parameters for the rte_flow_classify_table_entry_add API. This API adds a rule to the ACL table.

if (add_rules(parm_config.rule_ipv4_name)) {
    rte_flow_classifier_free(cls_app->cls);
    rte_free(cls_app);
    rte_exit(EXIT_FAILURE, "Failed to add rules\n");
}

Once the initialization is complete, the application is ready to launch a function on an lcore. In this example lcore_main() is called on a single lcore.

lcore_main(cls_app);

The lcore_main() function is explained below.

The Port Initialization Function

The main functional part of the port initialization used in the Basic Forwarding application is shown below:

static inline int
port_init(uint8_t port, struct rte_mempool *mbuf_pool)
{
    struct rte_eth_conf port_conf = port_conf_default;
    const uint16_t rx_rings = 1, tx_rings = 1;
    struct ether_addr addr;
    int retval;
    uint16_t q;

    if (port >= rte_eth_dev_count())
        return -1;

    /* Configure the Ethernet device. */
    retval = rte_eth_dev_configure(port, rx_rings, tx_rings, &port_conf);
    if (retval != 0)
        return retval;

    /* Allocate and set up 1 RX queue per Ethernet port. */
    for (q = 0; q < rx_rings; q++) {
        retval = rte_eth_rx_queue_setup(port, q, RX_RING_SIZE,
                rte_eth_dev_socket_id(port), NULL, mbuf_pool);
        if (retval < 0)
            return retval;
    }

    /* Allocate and set up 1 TX queue per Ethernet port. */
    for (q = 0; q < tx_rings; q++) {
        retval = rte_eth_tx_queue_setup(port, q, TX_RING_SIZE,
                rte_eth_dev_socket_id(port), NULL);
        if (retval < 0)
            return retval;
    }

    /* Start the Ethernet port. */
    retval = rte_eth_dev_start(port);
    if (retval < 0)
        return retval;

    /* Display the port MAC address. */
    rte_eth_macaddr_get(port, &addr);
    printf("Port %u MAC: %02" PRIx8 " %02" PRIx8 " %02" PRIx8
           " %02" PRIx8 " %02" PRIx8 " %02" PRIx8 "\n",
           port,
           addr.addr_bytes[0], addr.addr_bytes[1],
           addr.addr_bytes[2], addr.addr_bytes[3],
           addr.addr_bytes[4], addr.addr_bytes[5]);

    /* Enable RX in promiscuous mode for the Ethernet device. */
    rte_eth_promiscuous_enable(port);

    return 0;
}

The Ethernet ports are configured with default settings using the rte_eth_dev_configure() function and the port_conf_default struct.

static const struct rte_eth_conf port_conf_default = {
    .rxmode = { .max_rx_pkt_len = ETHER_MAX_LEN }
};

For this example the ports are set up with 1 RX and 1 TX queue using the rte_eth_rx_queue_setup() and rte_eth_tx_queue_setup() functions.

The Ethernet port is then started:

retval  = rte_eth_dev_start(port);

Finally the RX port is set in promiscuous mode:

rte_eth_promiscuous_enable(port);
The Add Rules function

The add_rules function reads the ipv4_rules_file.txt file and calls the add_classify_rule function which calls the rte_flow_classify_table_entry_add API.

static int
add_rules(const char *rule_path)
{
    FILE *fh;
    char buff[LINE_MAX];
    unsigned int i = 0;
    unsigned int total_num = 0;
    struct rte_eth_ntuple_filter ntuple_filter;

    fh = fopen(rule_path, "rb");
    if (fh == NULL)
        rte_exit(EXIT_FAILURE, "%s: Open %s failed\n", __func__,
                 rule_path);

    fseek(fh, 0, SEEK_SET);

    i = 0;
    while (fgets(buff, LINE_MAX, fh) != NULL) {
        i++;

        if (is_bypass_line(buff))
            continue;

        if (total_num >= FLOW_CLASSIFY_MAX_RULE_NUM - 1) {
            printf("\nINFO: classify rule capacity %d reached\n",
                   total_num);
            break;
        }

        if (parse_ipv4_5tuple_rule(buff, &ntuple_filter) != 0)
            rte_exit(EXIT_FAILURE,
                     "%s Line %u: parse rules error\n",
                     rule_path, i);

        if (add_classify_rule(&ntuple_filter) != 0)
            rte_exit(EXIT_FAILURE, "add rule error\n");

        total_num++;
    }

    fclose(fh);
    return 0;
}
The Lcore Main function

As we saw above the main() function calls an application function on the available lcores. The lcore_main function calls the rte_flow_classifier_query API. For the Basic Forwarding application the lcore_main function looks like the following:

/* flow classify data */
static int num_classify_rules;
static struct rte_flow_classify_rule *rules[MAX_NUM_CLASSIFY];
static struct rte_flow_classify_ipv4_5tuple_stats ntuple_stats;
static struct rte_flow_classify_stats classify_stats = {
        .stats = (void *)&ntuple_stats
};

static __attribute__((noreturn)) void
lcore_main(cls_app)
{
    const uint8_t nb_ports = rte_eth_dev_count();
    uint8_t port;

    /*
     * Check that the port is on the same NUMA node as the polling thread
     * for best performance.
     */
    for (port = 0; port < nb_ports; port++)
        if (rte_eth_dev_socket_id(port) > 0 &&
            rte_eth_dev_socket_id(port) != (int)rte_socket_id()) {
            printf("\n\n");
            printf("WARNING: port %u is on remote NUMA node\n",
                   port);
            printf("to polling thread.\n");
            printf("Performance will not be optimal.\n");

            printf("\nCore %u forwarding packets. \n",
                   rte_lcore_id());
            printf("[Ctrl+C to quit]\n
        }

    /* Run until the application is quit or killed. */
    for (;;) {
        /*
         * Receive packets on a port and forward them on the paired
         * port. The mapping is 0 -> 1, 1 -> 0, 2 -> 3, 3 -> 2, etc.
         */
        for (port = 0; port < nb_ports; port++) {

            /* Get burst of RX packets, from first port of pair. */
            struct rte_mbuf *bufs[BURST_SIZE];
            const uint16_t nb_rx = rte_eth_rx_burst(port, 0,
                    bufs, BURST_SIZE);

            if (unlikely(nb_rx == 0))
                continue;

            for (i = 0; i < MAX_NUM_CLASSIFY; i++) {
                if (rules[i]) {
                    ret = rte_flow_classifier_query(
                        cls_app->cls,
                        cls_app->table_id[0],
                        bufs, nb_rx, rules[i],
                        &classify_stats);
                    if (ret)
                        printf(
                            "rule [%d] query failed ret [%d]\n\n",
                            i, ret);
                    else {
                        printf(
                            "rule[%d] count=%"PRIu64"\n",
                            i, ntuple_stats.counter1);

                        printf("proto = %d\n",
                            ntuple_stats.ipv4_5tuple.proto);
                    }
                 }
             }

            /* Send burst of TX packets, to second port of pair. */
            const uint16_t nb_tx = rte_eth_tx_burst(port ^ 1, 0,
                    bufs, nb_rx);

            /* Free any unsent packets. */
            if (unlikely(nb_tx < nb_rx)) {
                uint16_t buf;
                for (buf = nb_tx; buf < nb_rx; buf++)
                    rte_pktmbuf_free(bufs[buf]);
            }
        }
    }
}

The main work of the application is done within the loop:

for (;;) {
    for (port = 0; port < nb_ports; port++) {

        /* Get burst of RX packets, from first port of pair. */
        struct rte_mbuf *bufs[BURST_SIZE];
        const uint16_t nb_rx = rte_eth_rx_burst(port, 0,
                bufs, BURST_SIZE);

        if (unlikely(nb_rx == 0))
            continue;

        /* Send burst of TX packets, to second port of pair. */
        const uint16_t nb_tx = rte_eth_tx_burst(port ^ 1, 0,
                bufs, nb_rx);

        /* Free any unsent packets. */
        if (unlikely(nb_tx < nb_rx)) {
            uint16_t buf;
            for (buf = nb_tx; buf < nb_rx; buf++)
                rte_pktmbuf_free(bufs[buf]);
        }
    }
}

Packets are received in bursts on the RX ports and transmitted in bursts on the TX ports. The ports are grouped in pairs with a simple mapping scheme using the an XOR on the port number:

0 -> 1
1 -> 0

2 -> 3
3 -> 2

etc.

The rte_eth_tx_burst() function frees the memory buffers of packets that are transmitted. If packets fail to transmit, (nb_tx < nb_rx), then they must be freed explicitly using rte_pktmbuf_free().

The forwarding loop can be interrupted and the application closed using Ctrl-C.

Basic RTE Flow Filtering Sample Application

The Basic RTE flow filtering sample application is a simple example of a creating a RTE flow rule.

It is intended as a demonstration of the basic components RTE flow rules.

Compiling the Application

To compile the application export the path to the DPDK source tree and go to the example directory:

export RTE_SDK=/path/to/rte_sdk

cd ${RTE_SDK}/examples/flow_filtering

Set the target, for example:

export RTE_TARGET=x86_64-native-linuxapp-gcc

See the DPDK Getting Started Guide for possible RTE_TARGET values.

Build the application as follows:

make

Running the Application

To run the example in a linuxapp environment:

./build/flow -l 1 -n 1

Refer to DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The example is build from 2 main files, main.c which holds the example logic and flow_blocks.c that holds the implementation for building the flow rule.

The following sections provide an explanation of the main components of the code.

All DPDK library functions used in the sample code are prefixed with rte_ and are explained in detail in the DPDK API Documentation.

The Main Function

The main() function located in main.c file performs the initialization and runs the main loop function.

The first task is to initialize the Environment Abstraction Layer (EAL). The argc and argv arguments are provided to the rte_eal_init() function. The value returned is the number of parsed arguments:

int ret = rte_eal_init(argc, argv);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Error with EAL initialization\n");

The main() also allocates a mempool to hold the mbufs (Message Buffers) used by the application:

mbuf_pool = rte_pktmbuf_pool_create("mbuf_pool", 4096, 128, 0,
                                         RTE_MBUF_DEFAULT_BUF_SIZE,
                                         rte_socket_id());

Mbufs are the packet buffer structure used by DPDK. They are explained in detail in the “Mbuf Library” section of the DPDK Programmer’s Guide.

The main() function also initializes all the ports using the user defined init_port() function which is explained in the next section:

init_port();

Once the initialization is complete, we set the flow rule using the following code:

/* create flow for send packet with */
flow = generate_ipv4_flow(port_id, selected_queue,
                             SRC_IP, EMPTY_MASK,
                             DEST_IP, FULL_MASK, &error);
if (!flow) {
       printf("Flow can't be created %d message: %s\n",
                    error.type,
                    error.message ? error.message : "(no stated reason)");
       rte_exit(EXIT_FAILURE, "error in creating flow");
}

In the last part the application is ready to launch the main_loop() function. Which is explained below.

main_loop();
The Port Initialization Function

The main functional part of the port initialization used in the flow filtering application is shown below:

init_port(void)
{
        int ret;
        uint16_t i;
        struct rte_eth_conf port_conf = {
                .rxmode = {
                        .split_hdr_size = 0,
                        /**< Header Split disabled */
                        .header_split   = 0,
                        /**< IP checksum offload disabled */
                        .hw_ip_checksum = 0,
                        /**< VLAN filtering disabled */
                        .hw_vlan_filter = 0,
                        /**< Jumbo Frame Support disabled */
                        .jumbo_frame    = 0,
                        /**< CRC stripped by hardware */
                        .hw_strip_crc   = 1,
                },
        };

        printf(":: initializing port: %d\n", port_id);
        ret = rte_eth_dev_configure(port_id,
                                nr_queues, nr_queues, &port_conf);
        if (ret < 0) {
                rte_exit(EXIT_FAILURE,
                        ":: cannot configure device: err=%d, port=%u\n",
                        ret, port_id);
        }

        /* only set Rx queues: something we care only so far */
        for (i = 0; i < nr_queues; i++) {
                ret = rte_eth_rx_queue_setup(port_id, i, 512,
                                     rte_eth_dev_socket_id(port_id),
                                     NULL,
                                     mbuf_pool);
                if (ret < 0) {
                        rte_exit(EXIT_FAILURE,
                           ":: Rx queue setup failed: err=%d, port=%u\n",
                           ret, port_id);
                }
        }


        rte_eth_promiscuous_enable(port_id);

        ret = rte_eth_dev_start(port_id);
        if (ret < 0) {
                rte_exit(EXIT_FAILURE,
                        "rte_eth_dev_start:err=%d, port=%u\n",
                        ret, port_id);
        }

        assert_link_status();

        printf(":: initializing port: %d done\n", port_id);
}

The Ethernet port is configured with default settings using the rte_eth_dev_configure() function and the port_conf_default struct:

struct rte_eth_conf port_conf = {
             .rxmode = {
                     .split_hdr_size = 0,
                     /**< Header Split disabled */
                     .header_split   = 0,
                     /**< IP checksum offload disabled */
                     .hw_ip_checksum = 0,
                     /**< VLAN filtering disabled */
                     .hw_vlan_filter = 0,
                     /**< Jumbo Frame Support disabled */
                     .jumbo_frame    = 0,
                     /**< CRC stripped by hardware */
                     .hw_strip_crc   = 1,
             },
};

ret = rte_eth_dev_configure(port_id, nr_queues, nr_queues, &port_conf);
if (ret < 0) {
     rte_exit(EXIT_FAILURE,
              ":: cannot configure device: err=%d, port=%u\n",
              ret, port_id);
}

For this example we are configuring number of rx queues that are connected to a single port.

 for (i = 0; i < nr_queues; i++) {
        ret = rte_eth_rx_queue_setup(port_id, i, 512,
                                     rte_eth_dev_socket_id(port_id),
                                     NULL,
                                     mbuf_pool);
        if (ret < 0) {
                rte_exit(EXIT_FAILURE,
                        ":: Rx queue setup failed: err=%d, port=%u\n",
                        ret, port_id);
        }
}

In the next step we create and apply the flow rule. which is to send packets with destination ip equals to 192.168.1.1 to queue number 1. The detail explanation of the generate_ipv4_flow() appears later in this document:

flow = generate_ipv4_flow(port_id, selected_queue,
                          SRC_IP, EMPTY_MASK,
                          DEST_IP, FULL_MASK, &error);

We are setting the RX port to promiscuous mode:

rte_eth_promiscuous_enable(port_id);

The last step is to start the port.

ret = rte_eth_dev_start(port_id);
if (ret < 0)  {
     rte_exit(EXIT_FAILURE, "rte_eth_dev_start:err%d, port=%u\n",
                     ret, port_id);
}
The main_loop function

As we saw above the main() function calls an application function to handle the main loop. For the flow filtering application the main_loop function looks like the following:

static void
main_loop(void)
{
        struct rte_mbuf *mbufs[32];
        struct ether_hdr *eth_hdr;
        uint16_t nb_rx;
        uint16_t i;
        uint16_t j;

        while (!force_quit) {
                for (i = 0; i < nr_queues; i++) {
                        nb_rx = rte_eth_rx_burst(port_id,
                                                i, mbufs, 32);
                        if (nb_rx) {
                                for (j = 0; j < nb_rx; j++) {
                                        struct rte_mbuf *m = mbufs[j];

                                        eth_hdr = rte_pktmbuf_mtod(m,
                                                     struct ether_hdr *);
                                        print_ether_addr("src=",
                                                     &eth_hdr->s_addr);
                                        print_ether_addr(" - dst=",
                                                     &eth_hdr->d_addr);
                                        printf(" - queue=0x%x",
                                                        (unsigned int)i);
                                        printf("\n");
                                        rte_pktmbuf_free(m);
                                }
                        }
                }
        }
        /* closing and releasing resources */
        rte_flow_flush(port_id, &error);
        rte_eth_dev_stop(port_id);
        rte_eth_dev_close(port_id);
}

The main work of the application is reading the packets from all queues and printing for each packet the destination queue:

while (!force_quit) {
    for (i = 0; i < nr_queues; i++) {
               nb_rx = rte_eth_rx_burst(port_id, i, mbufs, 32);
            if (nb_rx) {
                    for (j = 0; j < nb_rx; j++) {
                         struct rte_mbuf *m = mbufs[j];
                         eth_hdr = rte_pktmbuf_mtod(m, struct ether_hdr *);
                         print_ether_addr("src=", &eth_hdr->s_addr);
                         print_ether_addr(" - dst=", &eth_hdr->d_addr);
                         printf(" - queue=0x%x", (unsigned int)i);
                         printf("\n");
                         rte_pktmbuf_free(m);
                    }
            }
       }
}

The forwarding loop can be interrupted and the application closed using Ctrl-C. Which results in closing the port and the device using rte_eth_dev_stop and rte_eth_dev_close

The generate_ipv4_flow function

The generate_ipv4_rule function is responsible for creating the flow rule. This function is located in the flow_blocks.c file.

static struct rte_flow *
generate_ipv4_flow(uint8_t port_id, uint16_t rx_q,
                uint32_t src_ip, uint32_t src_mask,
                uint32_t dest_ip, uint32_t dest_mask,
                struct rte_flow_error *error)
{
        struct rte_flow_attr attr;
        struct rte_flow_item pattern[MAX_PATTERN_NUM];
        struct rte_flow_action action[MAX_PATTERN_NUM];
        struct rte_flow *flow = NULL;
        struct rte_flow_action_queue queue = { .index = rx_q };
        struct rte_flow_item_eth eth_spec;
        struct rte_flow_item_eth eth_mask;
        struct rte_flow_item_vlan vlan_spec;
        struct rte_flow_item_vlan vlan_mask;
        struct rte_flow_item_ipv4 ip_spec;
        struct rte_flow_item_ipv4 ip_mask;

        memset(pattern, 0, sizeof(pattern));
        memset(action, 0, sizeof(action));

        /*
         * set the rule attribute.
         * in this case only ingress packets will be checked.
         */
        memset(&attr, 0, sizeof(struct rte_flow_attr));
        attr.ingress = 1;

        /*
         * create the action sequence.
         * one action only,  move packet to queue
         */

        action[0].type = RTE_FLOW_ACTION_TYPE_QUEUE;
        action[0].conf = &queue;
        action[1].type = RTE_FLOW_ACTION_TYPE_END;

        /*
         * set the first level of the pattern (eth).
         * since in this example we just want to get the
         * ipv4 we set this level to allow all.
         */
        memset(&eth_spec, 0, sizeof(struct rte_flow_item_eth));
        memset(&eth_mask, 0, sizeof(struct rte_flow_item_eth));
        eth_spec.type = 0;
        eth_mask.type = 0;
        pattern[0].type = RTE_FLOW_ITEM_TYPE_ETH;
        pattern[0].spec = &eth_spec;
        pattern[0].mask = &eth_mask;

        /*
         * setting the second level of the pattern (vlan).
         * since in this example we just want to get the
         * ipv4 we also set this level to allow all.
         */
        memset(&vlan_spec, 0, sizeof(struct rte_flow_item_vlan));
        memset(&vlan_mask, 0, sizeof(struct rte_flow_item_vlan));
        pattern[1].type = RTE_FLOW_ITEM_TYPE_VLAN;
        pattern[1].spec = &vlan_spec;
        pattern[1].mask = &vlan_mask;

        /*
         * setting the third level of the pattern (ip).
         * in this example this is the level we care about
         * so we set it according to the parameters.
         */
        memset(&ip_spec, 0, sizeof(struct rte_flow_item_ipv4));
        memset(&ip_mask, 0, sizeof(struct rte_flow_item_ipv4));
        ip_spec.hdr.dst_addr = htonl(dest_ip);
        ip_mask.hdr.dst_addr = dest_mask;
        ip_spec.hdr.src_addr = htonl(src_ip);
        ip_mask.hdr.src_addr = src_mask;
        pattern[2].type = RTE_FLOW_ITEM_TYPE_IPV4;
        pattern[2].spec = &ip_spec;
        pattern[2].mask = &ip_mask;

        /* the final level must be always type end */
        pattern[3].type = RTE_FLOW_ITEM_TYPE_END;

        int res = rte_flow_validate(port_id, &attr, pattern, action, error);
        if(!res)
            flow = rte_flow_create(port_id, &attr, pattern, action, error);

        return flow;
}

The first part of the function is declaring the structures that will be used.

struct rte_flow_attr attr;
struct rte_flow_item pattern[MAX_PATTERN_NUM];
struct rte_flow_action action[MAX_PATTERN_NUM];
struct rte_flow *flow;
struct rte_flow_error error;
struct rte_flow_action_queue queue = { .index = rx_q };
struct rte_flow_item_eth eth_spec;
struct rte_flow_item_eth eth_mask;
struct rte_flow_item_vlan vlan_spec;
struct rte_flow_item_vlan vlan_mask;
struct rte_flow_item_ipv4 ip_spec;
struct rte_flow_item_ipv4 ip_mask;

The following part create the flow attributes, in our case ingress.

memset(&attr, 0, sizeof(struct rte_flow_attr));
attr.ingress = 1;

The third part defines the action to be taken when a packet matches the rule. In this case send the packet to queue.

action[0].type = RTE_FLOW_ACTION_TYPE_QUEUE;
action[0].conf = &queue;
action[1].type = RTE_FLOW_ACTION_TYPE_END;

The forth part is responsible for creating the pattern and is build from number of step. In each step we build one level of the pattern starting with the lowest one.

Setting the first level of the pattern ETH:

memset(&eth_spec, 0, sizeof(struct rte_flow_item_eth));
memset(&eth_mask, 0, sizeof(struct rte_flow_item_eth));
eth_spec.type = 0;
eth_mask.type = 0;
pattern[0].type = RTE_FLOW_ITEM_TYPE_ETH;
pattern[0].spec = &eth_spec;
pattern[0].mask = &eth_mask;

Setting the second level of the pattern VLAN:

memset(&vlan_spec, 0, sizeof(struct rte_flow_item_vlan));
memset(&vlan_mask, 0, sizeof(struct rte_flow_item_vlan));
pattern[1].type = RTE_FLOW_ITEM_TYPE_VLAN;
pattern[1].spec = &vlan_spec;
pattern[1].mask = &vlan_mask;

Setting the third level ip:

memset(&ip_spec, 0, sizeof(struct rte_flow_item_ipv4));
memset(&ip_mask, 0, sizeof(struct rte_flow_item_ipv4));
ip_spec.hdr.dst_addr = htonl(dest_ip);
ip_mask.hdr.dst_addr = dest_mask;
ip_spec.hdr.src_addr = htonl(src_ip);
ip_mask.hdr.src_addr = src_mask;
pattern[2].type = RTE_FLOW_ITEM_TYPE_IPV4;
pattern[2].spec = &ip_spec;
pattern[2].mask = &ip_mask;

Closing the pattern part.

pattern[3].type = RTE_FLOW_ITEM_TYPE_END;

The last part of the function is to validate the rule and create it.

int res = rte_flow_validate(port_id, &attr, pattern, action, &error);
if (!res)
     flow = rte_flow_create(port_id, &attr, pattern, action, &error);

IP Fragmentation Sample Application

The IPv4 Fragmentation application is a simple example of packet processing using the Data Plane Development Kit (DPDK). The application does L3 forwarding with IPv4 and IPv6 packet fragmentation.

Overview

The application demonstrates the use of zero-copy buffers for packet fragmentation. The initialization and run-time paths are very similar to those of the L2 Forwarding Sample Application (in Real and Virtualized Environments). This guide highlights the differences between the two applications.

There are three key differences from the L2 Forwarding sample application:

  • The first difference is that the IP Fragmentation sample application makes use of indirect buffers.
  • The second difference is that the forwarding decision is taken based on information read from the input packet’s IP header.
  • The third difference is that the application differentiates between IP and non-IP traffic by means of offload flags.

The Longest Prefix Match (LPM for IPv4, LPM6 for IPv6) table is used to store/lookup an outgoing port number, associated with that IP address. Any unmatched packets are forwarded to the originating port.

By default, input frame sizes up to 9.5 KB are supported. Before forwarding, the input IP packet is fragmented to fit into the “standard” Ethernet* v2 MTU (1500 bytes).

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the ip_fragmentation sub-directory.

Running the Application

The LPM object is created and loaded with the pre-configured entries read from global l3fwd_ipv4_route_array and l3fwd_ipv6_route_array tables. For each input packet, the packet forwarding decision (that is, the identification of the output interface for the packet) is taken as a result of LPM lookup. If the IP packet size is greater than default output MTU, then the input packet is fragmented and several fragments are sent via the output interface.

Application usage:

./build/ip_fragmentation [EAL options] -- -p PORTMASK [-q NQ]

where:

  • -p PORTMASK is a hexadecimal bitmask of ports to configure
  • -q NQ is the number of queue (=ports) per lcore (the default is 1)

To run the example in linuxapp environment with 2 lcores (2,4) over 2 ports(0,2) with 1 RX queue per lcore:

./build/ip_fragmentation -l 2,4 -n 3 -- -p 5
EAL: coremask set to 14
EAL: Detected lcore 0 on socket 0
EAL: Detected lcore 1 on socket 1
EAL: Detected lcore 2 on socket 0
EAL: Detected lcore 3 on socket 1
EAL: Detected lcore 4 on socket 0
...

Initializing port 0 on lcore 2... Address:00:1B:21:76:FA:2C, rxq=0 txq=2,0 txq=4,1
done: Link Up - speed 10000 Mbps - full-duplex
Skipping disabled port 1
Initializing port 2 on lcore 4... Address:00:1B:21:5C:FF:54, rxq=0 txq=2,0 txq=4,1
done: Link Up - speed 10000 Mbps - full-duplex
Skipping disabled port 3IP_FRAG: Socket 0: adding route 100.10.0.0/16 (port 0)
IP_FRAG: Socket 0: adding route 100.20.0.0/16 (port 1)
...
IP_FRAG: Socket 0: adding route 0101:0101:0101:0101:0101:0101:0101:0101/48 (port 0)
IP_FRAG: Socket 0: adding route 0201:0101:0101:0101:0101:0101:0101:0101/48 (port 1)
...
IP_FRAG: entering main loop on lcore 4
IP_FRAG: -- lcoreid=4 portid=2
IP_FRAG: entering main loop on lcore 2
IP_FRAG: -- lcoreid=2 portid=0

To run the example in linuxapp environment with 1 lcore (4) over 2 ports(0,2) with 2 RX queues per lcore:

./build/ip_fragmentation -l 4 -n 3 -- -p 5 -q 2

To test the application, flows should be set up in the flow generator that match the values in the l3fwd_ipv4_route_array and/or l3fwd_ipv6_route_array table.

The default l3fwd_ipv4_route_array table is:

struct l3fwd_ipv4_route l3fwd_ipv4_route_array[] = {
    {IPv4(100, 10, 0, 0), 16, 0},
    {IPv4(100, 20, 0, 0), 16, 1},
    {IPv4(100, 30, 0, 0), 16, 2},
    {IPv4(100, 40, 0, 0), 16, 3},
    {IPv4(100, 50, 0, 0), 16, 4},
    {IPv4(100, 60, 0, 0), 16, 5},
    {IPv4(100, 70, 0, 0), 16, 6},
    {IPv4(100, 80, 0, 0), 16, 7},
};

The default l3fwd_ipv6_route_array table is:

struct l3fwd_ipv6_route l3fwd_ipv6_route_array[] = {
    {{1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 0},
    {{2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 1},
    {{3, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 2},
    {{4, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 3},
    {{5, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 4},
    {{6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 5},
    {{7, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 6},
    {{8, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 7},
};

For example, for the input IPv4 packet with destination address: 100.10.1.1 and packet length 9198 bytes, seven IPv4 packets will be sent out from port #0 to the destination address 100.10.1.1: six of those packets will have length 1500 bytes and one packet will have length 318 bytes. IP Fragmentation sample application provides basic NUMA support in that all the memory structures are allocated on all sockets that have active lcores on them.

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

IPv4 Multicast Sample Application

The IPv4 Multicast application is a simple example of packet processing using the Data Plane Development Kit (DPDK). The application performs L3 multicasting.

Overview

The application demonstrates the use of zero-copy buffers for packet forwarding. The initialization and run-time paths are very similar to those of the L2 Forwarding Sample Application (in Real and Virtualized Environments). This guide highlights the differences between the two applications. There are two key differences from the L2 Forwarding sample application:

  • The IPv4 Multicast sample application makes use of indirect buffers.
  • The forwarding decision is taken based on information read from the input packet’s IPv4 header.

The lookup method is the Four-byte Key (FBK) hash-based method. The lookup table is composed of pairs of destination IPv4 address (the FBK) and a port mask associated with that IPv4 address.

Note

The max port mask supported in the given hash table is 0xf, so only first four ports can be supported. If using non-consecutive ports, use the destination IPv4 address accordingly.

For convenience and simplicity, this sample application does not take IANA-assigned multicast addresses into account, but instead equates the last four bytes of the multicast group (that is, the last four bytes of the destination IP address) with the mask of ports to multicast packets to. Also, the application does not consider the Ethernet addresses; it looks only at the IPv4 destination address for any given packet.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the ipv4_multicast sub-directory.

Running the Application

The application has a number of command line options:

./build/ipv4_multicast [EAL options] -- -p PORTMASK [-q NQ]

where,

  • -p PORTMASK: Hexadecimal bitmask of ports to configure
  • -q NQ: determines the number of queues per lcore

Note

Unlike the basic L2/L3 Forwarding sample applications, NUMA support is not provided in the IPv4 Multicast sample application.

Typically, to run the IPv4 Multicast sample application, issue the following command (as root):

./build/ipv4_multicast -l 0-3 -n 3 -- -p 0x3 -q 1

In this command:

  • The -l option enables cores 0, 1, 2 and 3
  • The -n option specifies 3 memory channels
  • The -p option enables ports 0 and 1
  • The -q option assigns 1 queue to each lcore

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The following sections provide some explanation of the code. As mentioned in the overview section, the initialization and run-time paths are very similar to those of the L2 Forwarding Sample Application (in Real and Virtualized Environments). The following sections describe aspects that are specific to the IPv4 Multicast sample application.

Memory Pool Initialization

The IPv4 Multicast sample application uses three memory pools. Two of the pools are for indirect buffers used for packet duplication purposes. Memory pools for indirect buffers are initialized differently from the memory pool for direct buffers:

packet_pool = rte_pktmbuf_pool_create("packet_pool", NB_PKT_MBUF, 32,
                    0, PKT_MBUF_DATA_SIZE, rte_socket_id());
header_pool = rte_pktmbuf_pool_create("header_pool", NB_HDR_MBUF, 32,
                    0, HDR_MBUF_DATA_SIZE, rte_socket_id());
clone_pool = rte_pktmbuf_pool_create("clone_pool", NB_CLONE_MBUF, 32,
                    0, 0, rte_socket_id());

The reason for this is because indirect buffers are not supposed to hold any packet data and therefore can be initialized with lower amount of reserved memory for each buffer.

Hash Initialization

The hash object is created and loaded with the pre-configured entries read from a global array:

static int

init_mcast_hash(void)
{
    uint32_t i;
    mcast_hash_params.socket_id = rte_socket_id();

    mcast_hash = rte_fbk_hash_create(&mcast_hash_params);
    if (mcast_hash == NULL){
        return -1;
    }

    for (i = 0; i < N_MCAST_GROUPS; i ++){
        if (rte_fbk_hash_add_key(mcast_hash, mcast_group_table[i].ip, mcast_group_table[i].port_mask) < 0) {
                    return -1;
        }
    }
    return 0;
}
Forwarding

All forwarding is done inside the mcast_forward() function. Firstly, the Ethernet* header is removed from the packet and the IPv4 address is extracted from the IPv4 header:

/* Remove the Ethernet header from the input packet */

iphdr = (struct ipv4_hdr *)rte_pktmbuf_adj(m, sizeof(struct ether_hdr));
RTE_ASSERT(iphdr != NULL);
dest_addr = rte_be_to_cpu_32(iphdr->dst_addr);

Then, the packet is checked to see if it has a multicast destination address and if the routing table has any ports assigned to the destination address:

if (!IS_IPV4_MCAST(dest_addr) ||
   (hash = rte_fbk_hash_lookup(mcast_hash, dest_addr)) <= 0 ||
   (port_mask = hash & enabled_port_mask) == 0) {
       rte_pktmbuf_free(m);
       return;
}

Then, the number of ports in the destination portmask is calculated with the help of the bitcnt() function:

/* Get number of bits set. */

static inline uint32_t bitcnt(uint32_t v)
{
    uint32_t n;

    for (n = 0; v != 0; v &= v - 1, n++)
       ;
    return n;
}

This is done to determine which forwarding algorithm to use. This is explained in more detail in the next section.

Thereafter, a destination Ethernet address is constructed:

/* construct destination Ethernet address */

dst_eth_addr = ETHER_ADDR_FOR_IPV4_MCAST(dest_addr);

Since Ethernet addresses are also part of the multicast process, each outgoing packet carries the same destination Ethernet address. The destination Ethernet address is constructed from the lower 23 bits of the multicast group OR-ed with the Ethernet address 01:00:5e:00:00:00, as per RFC 1112:

#define ETHER_ADDR_FOR_IPV4_MCAST(x) \
    (rte_cpu_to_be_64(0x01005e000000ULL | ((x) & 0x7fffff)) >> 16)

Then, packets are dispatched to the destination ports according to the portmask associated with a multicast group:

for (port = 0; use_clone != port_mask; port_mask >>= 1, port++) {
    /* Prepare output packet and send it out. */

    if ((port_mask & 1) != 0) {
        if (likely ((mc = mcast_out_pkt(m, use_clone)) != NULL))
            mcast_send_pkt(mc, &dst_eth_addr.as_addr, qconf, port);
        else if (use_clone == 0)
             rte_pktmbuf_free(m);
   }
}

The actual packet transmission is done in the mcast_send_pkt() function:

static inline void mcast_send_pkt(struct rte_mbuf *pkt, struct ether_addr *dest_addr, struct lcore_queue_conf *qconf, uint16_t port)
{
    struct ether_hdr *ethdr;
    uint16_t len;

    /* Construct Ethernet header. */

    ethdr = (struct ether_hdr *)rte_pktmbuf_prepend(pkt, (uint16_t) sizeof(*ethdr));

    RTE_ASSERT(ethdr != NULL);

    ether_addr_copy(dest_addr, &ethdr->d_addr);
    ether_addr_copy(&ports_eth_addr[port], &ethdr->s_addr);
    ethdr->ether_type = rte_be_to_cpu_16(ETHER_TYPE_IPv4);

    /* Put new packet into the output queue */

    len = qconf->tx_mbufs[port].len;
    qconf->tx_mbufs[port].m_table[len] = pkt;
    qconf->tx_mbufs[port].len = ++len;

    /* Transmit packets */

    if (unlikely(MAX_PKT_BURST == len))
        send_burst(qconf, port);
}
Buffer Cloning

This is the most important part of the application since it demonstrates the use of zero- copy buffer cloning. There are two approaches for creating the outgoing packet and although both are based on the data zero-copy idea, there are some differences in the detail.

The first approach creates a clone of the input packet, for example, walk though all segments of the input packet and for each of segment, create a new buffer and attach that new buffer to the segment (refer to rte_pktmbuf_clone() in the rte_mbuf library for more details). A new buffer is then allocated for the packet header and is prepended to the cloned buffer.

The second approach does not make a clone, it just increments the reference counter for all input packet segment, allocates a new buffer for the packet header and prepends it to the input packet.

Basically, the first approach reuses only the input packet’s data, but creates its own copy of packet’s metadata. The second approach reuses both input packet’s data and metadata.

The advantage of first approach is that each outgoing packet has its own copy of the metadata, so we can safely modify the data pointer of the input packet. That allows us to skip creation if the output packet is for the last destination port and instead modify input packet’s header in place. For example, for N destination ports, we need to invoke mcast_out_pkt() (N-1) times.

The advantage of the second approach is that there is less work to be done for each outgoing packet, that is, the “clone” operation is skipped completely. However, there is a price to pay. The input packet’s metadata must remain intact, so for N destination ports, we need to invoke mcast_out_pkt() (N) times.

Therefore, for a small number of outgoing ports (and segments in the input packet), first approach is faster. As the number of outgoing ports (and/or input segments) grows, the second approach becomes more preferable.

Depending on the number of segments or the number of ports in the outgoing portmask, either the first (with cloning) or the second (without cloning) approach is taken:

use_clone = (port_num <= MCAST_CLONE_PORTS && m->pkt.nb_segs <= MCAST_CLONE_SEGS);

It is the mcast_out_pkt() function that performs the packet duplication (either with or without actually cloning the buffers):

static inline struct rte_mbuf *mcast_out_pkt(struct rte_mbuf *pkt, int use_clone)
{
    struct rte_mbuf *hdr;

    /* Create new mbuf for the header. */

    if (unlikely ((hdr = rte_pktmbuf_alloc(header_pool)) == NULL))
        return NULL;

    /* If requested, then make a new clone packet. */

    if (use_clone != 0 && unlikely ((pkt = rte_pktmbuf_clone(pkt, clone_pool)) == NULL)) {
        rte_pktmbuf_free(hdr);
        return NULL;
    }

    /* prepend new header */

    hdr->pkt.next = pkt;

    /* update header's fields */

    hdr->pkt.pkt_len = (uint16_t)(hdr->pkt.data_len + pkt->pkt.pkt_len);
    hdr->pkt.nb_segs = pkt->pkt.nb_segs + 1;

    /* copy metadata from source packet */

    hdr->pkt.in_port = pkt->pkt.in_port;
    hdr->pkt.vlan_macip = pkt->pkt.vlan_macip;
    hdr->pkt.hash = pkt->pkt.hash;
    hdr->ol_flags = pkt->ol_flags;
    rte_mbuf_sanity_check(hdr, RTE_MBUF_PKT, 1);

    return hdr;
}

IP Reassembly Sample Application

The L3 Forwarding application is a simple example of packet processing using the DPDK. The application performs L3 forwarding with reassembly for fragmented IPv4 and IPv6 packets.

Overview

The application demonstrates the use of the DPDK libraries to implement packet forwarding with reassembly for IPv4 and IPv6 fragmented packets. The initialization and run- time paths are very similar to those of the L2 Forwarding Sample Application (in Real and Virtualized Environments). The main difference from the L2 Forwarding sample application is that it reassembles fragmented IPv4 and IPv6 packets before forwarding. The maximum allowed size of reassembled packet is 9.5 KB.

There are two key differences from the L2 Forwarding sample application:

  • The first difference is that the forwarding decision is taken based on information read from the input packet’s IP header.
  • The second difference is that the application differentiates between IP and non-IP traffic by means of offload flags.

The Longest Prefix Match (LPM for IPv4, LPM6 for IPv6) table is used to store/lookup an outgoing port number, associated with that IPv4 address. Any unmatched packets are forwarded to the originating port.Compiling the Application

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the ip_reassembly sub-directory.

Running the Application

The application has a number of command line options:

./build/ip_reassembly [EAL options] -- -p PORTMASK [-q NQ] [--maxflows=FLOWS>] [--flowttl=TTL[(s|ms)]]

where:

  • -p PORTMASK: Hexadecimal bitmask of ports to configure
  • -q NQ: Number of RX queues per lcore
  • –maxflows=FLOWS: determines maximum number of active fragmented flows (1-65535). Default value: 4096.
  • –flowttl=TTL[(s|ms)]: determines maximum Time To Live for fragmented packet. If all fragments of the packet wouldn’t appear within given time-out, then they are considered as invalid and will be dropped. Valid range is 1ms - 3600s. Default value: 1s.

To run the example in linuxapp environment with 2 lcores (2,4) over 2 ports(0,2) with 1 RX queue per lcore:

./build/ip_reassembly -l 2,4 -n 3 -- -p 5
EAL: coremask set to 14
EAL: Detected lcore 0 on socket 0
EAL: Detected lcore 1 on socket 1
EAL: Detected lcore 2 on socket 0
EAL: Detected lcore 3 on socket 1
EAL: Detected lcore 4 on socket 0
...

Initializing port 0 on lcore 2... Address:00:1B:21:76:FA:2C, rxq=0 txq=2,0 txq=4,1
done: Link Up - speed 10000 Mbps - full-duplex
Skipping disabled port 1
Initializing port 2 on lcore 4... Address:00:1B:21:5C:FF:54, rxq=0 txq=2,0 txq=4,1
done: Link Up - speed 10000 Mbps - full-duplex
Skipping disabled port 3IP_FRAG: Socket 0: adding route 100.10.0.0/16 (port 0)
IP_RSMBL: Socket 0: adding route 100.20.0.0/16 (port 1)
...

IP_RSMBL: Socket 0: adding route 0101:0101:0101:0101:0101:0101:0101:0101/48 (port 0)
IP_RSMBL: Socket 0: adding route 0201:0101:0101:0101:0101:0101:0101:0101/48 (port 1)
...

IP_RSMBL: entering main loop on lcore 4
IP_RSMBL: -- lcoreid=4 portid=2
IP_RSMBL: entering main loop on lcore 2
IP_RSMBL: -- lcoreid=2 portid=0

To run the example in linuxapp environment with 1 lcore (4) over 2 ports(0,2) with 2 RX queues per lcore:

./build/ip_reassembly -l 4 -n 3 -- -p 5 -q 2

To test the application, flows should be set up in the flow generator that match the values in the l3fwd_ipv4_route_array and/or l3fwd_ipv6_route_array table.

Please note that in order to test this application, the traffic generator should be generating valid fragmented IP packets. For IPv6, the only supported case is when no other extension headers other than fragment extension header are present in the packet.

The default l3fwd_ipv4_route_array table is:

struct l3fwd_ipv4_route l3fwd_ipv4_route_array[] = {
    {IPv4(100, 10, 0, 0), 16, 0},
    {IPv4(100, 20, 0, 0), 16, 1},
    {IPv4(100, 30, 0, 0), 16, 2},
    {IPv4(100, 40, 0, 0), 16, 3},
    {IPv4(100, 50, 0, 0), 16, 4},
    {IPv4(100, 60, 0, 0), 16, 5},
    {IPv4(100, 70, 0, 0), 16, 6},
    {IPv4(100, 80, 0, 0), 16, 7},
};

The default l3fwd_ipv6_route_array table is:

struct l3fwd_ipv6_route l3fwd_ipv6_route_array[] = {
    {{1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 0},
    {{2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 1},
    {{3, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 2},
    {{4, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 3},
    {{5, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 4},
    {{6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 5},
    {{7, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 6},
    {{8, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 48, 7},
};

For example, for the fragmented input IPv4 packet with destination address: 100.10.1.1, a reassembled IPv4 packet be sent out from port #0 to the destination address 100.10.1.1 once all the fragments are collected.

Explanation

The following sections provide some explanation of the sample application code. As mentioned in the overview section, the initialization and run-time paths are very similar to those of the L2 Forwarding Sample Application (in Real and Virtualized Environments). The following sections describe aspects that are specific to the IP reassemble sample application.

IPv4 Fragment Table Initialization

This application uses the rte_ip_frag library. Please refer to Programmer’s Guide for more detailed explanation of how to use this library. Fragment table maintains information about already received fragments of the packet. Each IP packet is uniquely identified by triple <Source IP address>, <Destination IP address>, <ID>. To avoid lock contention, each RX queue has its own Fragment Table, e.g. the application can’t handle the situation when different fragments of the same packet arrive through different RX queues. Each table entry can hold information about packet consisting of up to RTE_LIBRTE_IP_FRAG_MAX_FRAGS fragments.

frag_cycles = (rte_get_tsc_hz() + MS_PER_S - 1) / MS_PER_S * max_flow_ttl;

if ((qconf->frag_tbl[queue] = rte_ip_frag_tbl_create(max_flow_num, IPV4_FRAG_TBL_BUCKET_ENTRIES, max_flow_num, frag_cycles, socket)) == NULL)
{
    RTE_LOG(ERR, IP_RSMBL, "ip_frag_tbl_create(%u) on " "lcore: %u for queue: %u failed\n",  max_flow_num, lcore, queue);
    return -1;
}
Mempools Initialization

The reassembly application demands a lot of mbuf’s to be allocated. At any given time up to (2 * max_flow_num * RTE_LIBRTE_IP_FRAG_MAX_FRAGS * <maximum number of mbufs per packet>) can be stored inside Fragment Table waiting for remaining fragments. To keep mempool size under reasonable limits and to avoid situation when one RX queue can starve other queues, each RX queue uses its own mempool.

nb_mbuf = RTE_MAX(max_flow_num, 2UL * MAX_PKT_BURST) * RTE_LIBRTE_IP_FRAG_MAX_FRAGS;
nb_mbuf *= (port_conf.rxmode.max_rx_pkt_len + BUF_SIZE - 1) / BUF_SIZE;
nb_mbuf *= 2; /* ipv4 and ipv6 */
nb_mbuf += RTE_TEST_RX_DESC_DEFAULT + RTE_TEST_TX_DESC_DEFAULT;
nb_mbuf = RTE_MAX(nb_mbuf, (uint32_t)NB_MBUF);

snprintf(buf, sizeof(buf), "mbuf_pool_%u_%u", lcore, queue);

if ((rxq->pool = rte_mempool_create(buf, nb_mbuf, MBUF_SIZE, 0, sizeof(struct rte_pktmbuf_pool_private), rte_pktmbuf_pool_init, NULL,
    rte_pktmbuf_init, NULL, socket, MEMPOOL_F_SP_PUT | MEMPOOL_F_SC_GET)) == NULL) {

        RTE_LOG(ERR, IP_RSMBL, "mempool_create(%s) failed", buf);
        return -1;
}
Packet Reassembly and Forwarding

For each input packet, the packet forwarding operation is done by the l3fwd_simple_forward() function. If the packet is an IPv4 or IPv6 fragment, then it calls rte_ipv4_reassemble_packet() for IPv4 packets, or rte_ipv6_reassemble_packet() for IPv6 packets. These functions either return a pointer to valid mbuf that contains reassembled packet, or NULL (if the packet can’t be reassembled for some reason). Then l3fwd_simple_forward() continues with the code for the packet forwarding decision (that is, the identification of the output interface for the packet) and actual transmit of the packet.

The rte_ipv4_reassemble_packet() or rte_ipv6_reassemble_packet() are responsible for:

  1. Searching the Fragment Table for entry with packet’s <IP Source Address, IP Destination Address, Packet ID>
  2. If the entry is found, then check if that entry already timed-out. If yes, then free all previously received fragments, and remove information about them from the entry.
  3. If no entry with such key is found, then try to create a new one by one of two ways:
    1. Use as empty entry
    2. Delete a timed-out entry, free mbufs associated with it mbufs and store a new entry with specified key in it.
  4. Update the entry with new fragment information and check if a packet can be reassembled (the packet’s entry contains all fragments).
    1. If yes, then, reassemble the packet, mark table’s entry as empty and return the reassembled mbuf to the caller.
    2. If no, then just return a NULL to the caller.

If at any stage of packet processing a reassembly function encounters an error (can’t insert new entry into the Fragment table, or invalid/timed-out fragment), then it will free all associated with the packet fragments, mark the table entry as invalid and return NULL to the caller.

Debug logging and Statistics Collection

The RTE_LIBRTE_IP_FRAG_TBL_STAT controls statistics collection for the IP Fragment Table. This macro is disabled by default. To make ip_reassembly print the statistics to the standard output, the user must send either an USR1, INT or TERM signal to the process. For all of these signals, the ip_reassembly process prints Fragment table statistics for each RX queue, plus the INT and TERM will cause process termination as usual.

Kernel NIC Interface Sample Application

The Kernel NIC Interface (KNI) is a DPDK control plane solution that allows userspace applications to exchange packets with the kernel networking stack. To accomplish this, DPDK userspace applications use an IOCTL call to request the creation of a KNI virtual device in the Linux* kernel. The IOCTL call provides interface information and the DPDK’s physical address space, which is re-mapped into the kernel address space by the KNI kernel loadable module that saves the information to a virtual device context. The DPDK creates FIFO queues for packet ingress and egress to the kernel module for each device allocated.

The KNI kernel loadable module is a standard net driver, which upon receiving the IOCTL call access the DPDK’s FIFO queue to receive/transmit packets from/to the DPDK userspace application. The FIFO queues contain pointers to data packets in the DPDK. This:

  • Provides a faster mechanism to interface with the kernel net stack and eliminates system calls
  • Facilitates the DPDK using standard Linux* userspace net tools (tcpdump, ftp, and so on)
  • Eliminate the copy_to_user and copy_from_user operations on packets.

The Kernel NIC Interface sample application is a simple example that demonstrates the use of the DPDK to create a path for packets to go through the Linux* kernel. This is done by creating one or more kernel net devices for each of the DPDK ports. The application allows the use of standard Linux tools (ethtool, ifconfig, tcpdump) with the DPDK ports and also the exchange of packets between the DPDK application and the Linux* kernel.

Overview

The Kernel NIC Interface sample application uses two threads in user space for each physical NIC port being used, and allocates one or more KNI device for each physical NIC port with kernel module’s support. For a physical NIC port, one thread reads from the port and writes to KNI devices, and another thread reads from KNI devices and writes the data unmodified to the physical NIC port. It is recommended to configure one KNI device for each physical NIC port. If configured with more than one KNI devices for a physical NIC port, it is just for performance testing, or it can work together with VMDq support in future.

The packet flow through the Kernel NIC Interface application is as shown in the following figure.

_images/kernel_nic.png

Kernel NIC Application Packet Flow

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the kni sub-directory.

Note

This application is intended as a linuxapp only.

Loading the Kernel Module

Loading the KNI kernel module without any parameter is the typical way a DPDK application gets packets into and out of the kernel net stack. This way, only one kernel thread is created for all KNI devices for packet receiving in kernel side:

#insmod rte_kni.ko

Pinning the kernel thread to a specific core can be done using a taskset command such as following:

#taskset -p 100000 `pgrep --fl kni_thread | awk '{print $1}'`

This command line tries to pin the specific kni_thread on the 20th lcore (lcore numbering starts at 0), which means it needs to check if that lcore is available on the board. This command must be sent after the application has been launched, as insmod does not start the kni thread.

For optimum performance, the lcore in the mask must be selected to be on the same socket as the lcores used in the KNI application.

To provide flexibility of performance, the kernel module of the KNI, located in the kmod sub-directory of the DPDK target directory, can be loaded with parameter of kthread_mode as follows:

  • #insmod rte_kni.ko kthread_mode=single

    This mode will create only one kernel thread for all KNI devices for packet receiving in kernel side. By default, it is in this single kernel thread mode. It can set core affinity for this kernel thread by using Linux command taskset.

  • #insmod rte_kni.ko kthread_mode =multiple

    This mode will create a kernel thread for each KNI device for packet receiving in kernel side. The core affinity of each kernel thread is set when creating the KNI device. The lcore ID for each kernel thread is provided in the command line of launching the application. Multiple kernel thread mode can provide scalable higher performance.

To measure the throughput in a loopback mode, the kernel module of the KNI, located in the kmod sub-directory of the DPDK target directory, can be loaded with parameters as follows:

  • #insmod rte_kni.ko lo_mode=lo_mode_fifo

    This loopback mode will involve ring enqueue/dequeue operations in kernel space.

  • #insmod rte_kni.ko lo_mode=lo_mode_fifo_skb

    This loopback mode will involve ring enqueue/dequeue operations and sk buffer copies in kernel space.

Running the Application

The application requires a number of command line options:

kni [EAL options] -- -P -p PORTMASK --config="(port,lcore_rx,lcore_tx[,lcore_kthread,...])[,port,lcore_rx,lcore_tx[,lcore_kthread,...]]"

Where:

  • -P: Set all ports to promiscuous mode so that packets are accepted regardless of the packet’s Ethernet MAC destination address. Without this option, only packets with the Ethernet MAC destination address set to the Ethernet address of the port are accepted.
  • -p PORTMASK: Hexadecimal bitmask of ports to configure.
  • –config=”(port,lcore_rx, lcore_tx[,lcore_kthread, ...]) [, port,lcore_rx, lcore_tx[,lcore_kthread, ...]]”: Determines which lcores of RX, TX, kernel thread are mapped to which ports.

Refer to DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

The -c coremask or -l corelist parameter of the EAL options should include the lcores indicated by the lcore_rx and lcore_tx, but does not need to include lcores indicated by lcore_kthread as they are used to pin the kernel thread on. The -p PORTMASK parameter should include the ports indicated by the port in –config, neither more nor less.

The lcore_kthread in –config can be configured none, one or more lcore IDs. In multiple kernel thread mode, if configured none, a KNI device will be allocated for each port, while no specific lcore affinity will be set for its kernel thread. If configured one or more lcore IDs, one or more KNI devices will be allocated for each port, while specific lcore affinity will be set for its kernel thread. In single kernel thread mode, if configured none, a KNI device will be allocated for each port. If configured one or more lcore IDs, one or more KNI devices will be allocated for each port while no lcore affinity will be set as there is only one kernel thread for all KNI devices.

For example, to run the application with two ports served by six lcores, one lcore of RX, one lcore of TX, and one lcore of kernel thread for each port:

./build/kni -l 4-7 -n 4 -- -P -p 0x3 -config="(0,4,6,8),(1,5,7,9)"

KNI Operations

Once the KNI application is started, one can use different Linux* commands to manage the net interfaces. If more than one KNI devices configured for a physical port, only the first KNI device will be paired to the physical device. Operations on other KNI devices will not affect the physical port handled in user space application.

Assigning an IP address:

#ifconfig vEth0_0 192.168.0.1

Displaying the NIC registers:

#ethtool -d vEth0_0

Dumping the network traffic:

#tcpdump -i vEth0_0

When the DPDK userspace application is closed, all the KNI devices are deleted from Linux*.

Explanation

The following sections provide some explanation of code.

Initialization

Setup of mbuf pool, driver and queues is similar to the setup done in the L2 Forwarding Sample Application (in Real and Virtualized Environments).. In addition, one or more kernel NIC interfaces are allocated for each of the configured ports according to the command line parameters.

The code for allocating the kernel NIC interfaces for a specific port is as follows:

 static int
 kni_alloc(uint16_t port_id)
 {
     uint8_t i;
     struct rte_kni *kni;
     struct rte_kni_conf conf;
     struct kni_port_params **params = kni_port_params_array;

     if (port_id >= RTE_MAX_ETHPORTS || !params[port_id])
         return -1;

     params[port_id]->nb_kni = params[port_id]->nb_lcore_k ? params[port_id]->nb_lcore_k : 1;

     for (i = 0; i < params[port_id]->nb_kni; i++) {

         /* Clear conf at first */

         memset(&conf, 0, sizeof(conf));
         if (params[port_id]->nb_lcore_k) {
             snprintf(conf.name, RTE_KNI_NAMESIZE, "vEth%u_%u", port_id, i);
             conf.core_id = params[port_id]->lcore_k[i];
             conf.force_bind = 1;
         } else
             snprintf(conf.name, RTE_KNI_NAMESIZE, "vEth%u", port_id);
             conf.group_id = (uint16_t)port_id;
             conf.mbuf_size = MAX_PACKET_SZ;

             /*
              *   The first KNI device associated to a port
              *   is the master, for multiple kernel thread
              *   environment.
              */

             if (i == 0) {
                 struct rte_kni_ops ops;
                 struct rte_eth_dev_info dev_info;

                 memset(&dev_info, 0, sizeof(dev_info)); rte_eth_dev_info_get(port_id, &dev_info);

                 conf.addr = dev_info.pci_dev->addr;
                 conf.id = dev_info.pci_dev->id;

                 memset(&ops, 0, sizeof(ops));

                 ops.port_id = port_id;
                 ops.change_mtu = kni_change_mtu;
                 ops.config_network_if = kni_config_network_interface;

                 kni = rte_kni_alloc(pktmbuf_pool, &conf, &ops);
             } else
                 kni = rte_kni_alloc(pktmbuf_pool, &conf, NULL);

             if (!kni)
                 rte_exit(EXIT_FAILURE, "Fail to create kni for "
                         "port: %d\n", port_id);

             params[port_id]->kni[i] = kni;
         }
     return 0;
}

The other step in the initialization process that is unique to this sample application is the association of each port with lcores for RX, TX and kernel threads.

  • One lcore to read from the port and write to the associated one or more KNI devices
  • Another lcore to read from one or more KNI devices and write to the port
  • Other lcores for pinning the kernel threads on one by one

This is done by using the`kni_port_params_array[]` array, which is indexed by the port ID. The code is as follows:

static int
parse_config(const char *arg)
{
    const char *p, *p0 = arg;
    char s[256], *end;
    unsigned size;
    enum fieldnames {
        FLD_PORT = 0,
        FLD_LCORE_RX,
        FLD_LCORE_TX,
        _NUM_FLD = KNI_MAX_KTHREAD + 3,
    };
    int i, j, nb_token;
    char *str_fld[_NUM_FLD];
    unsigned long int_fld[_NUM_FLD];
    uint16_t port_id, nb_kni_port_params = 0;

    memset(&kni_port_params_array, 0, sizeof(kni_port_params_array));

    while (((p = strchr(p0, '(')) != NULL) && nb_kni_port_params < RTE_MAX_ETHPORTS) {
        p++;
        if ((p0 = strchr(p, ')')) == NULL)
            goto fail;

        size = p0 - p;

        if (size >= sizeof(s)) {
            printf("Invalid config parameters\n");
            goto fail;
        }

        snprintf(s, sizeof(s), "%.*s", size, p);
        nb_token = rte_strsplit(s, sizeof(s), str_fld, _NUM_FLD, ',');

        if (nb_token <= FLD_LCORE_TX) {
            printf("Invalid config parameters\n");
            goto fail;
        }

        for (i = 0; i < nb_token; i++) {
            errno = 0;
            int_fld[i] = strtoul(str_fld[i], &end, 0);
            if (errno != 0 || end == str_fld[i]) {
                printf("Invalid config parameters\n");
                goto fail;
            }
        }

        i = 0;
        port_id = (uint8_t)int_fld[i++];

        if (port_id >= RTE_MAX_ETHPORTS) {
            printf("Port ID %u could not exceed the maximum %u\n", port_id, RTE_MAX_ETHPORTS);
            goto fail;
        }

        if (kni_port_params_array[port_id]) {
            printf("Port %u has been configured\n", port_id);
            goto fail;
        }

        kni_port_params_array[port_id] = (struct kni_port_params*)rte_zmalloc("KNI_port_params", sizeof(struct kni_port_params), RTE_CACHE_LINE_SIZE);
        kni_port_params_array[port_id]->port_id = port_id;
        kni_port_params_array[port_id]->lcore_rx = (uint8_t)int_fld[i++];
        kni_port_params_array[port_id]->lcore_tx = (uint8_t)int_fld[i++];

        if (kni_port_params_array[port_id]->lcore_rx >= RTE_MAX_LCORE || kni_port_params_array[port_id]->lcore_tx >= RTE_MAX_LCORE) {
            printf("lcore_rx %u or lcore_tx %u ID could not "
                    "exceed the maximum %u\n",
                    kni_port_params_array[port_id]->lcore_rx, kni_port_params_array[port_id]->lcore_tx, RTE_MAX_LCORE);
            goto fail;
       }

    for (j = 0; i < nb_token && j < KNI_MAX_KTHREAD; i++, j++)
        kni_port_params_array[port_id]->lcore_k[j] = (uint8_t)int_fld[i];
        kni_port_params_array[port_id]->nb_lcore_k = j;
    }

    print_config();

    return 0;

fail:

    for (i = 0; i < RTE_MAX_ETHPORTS; i++) {
        if (kni_port_params_array[i]) {
            rte_free(kni_port_params_array[i]);
            kni_port_params_array[i] = NULL;
        }
    }

    return -1;

}
Packet Forwarding

After the initialization steps are completed, the main_loop() function is run on each lcore. This function first checks the lcore_id against the user provided lcore_rx and lcore_tx to see if this lcore is reading from or writing to kernel NIC interfaces.

For the case that reads from a NIC port and writes to the kernel NIC interfaces, the packet reception is the same as in L2 Forwarding sample application (see Receive, Process and Transmit Packets). The packet transmission is done by sending mbufs into the kernel NIC interfaces by rte_kni_tx_burst(). The KNI library automatically frees the mbufs after the kernel successfully copied the mbufs.

/**
 *   Interface to burst rx and enqueue mbufs into rx_q
 */

static void
kni_ingress(struct kni_port_params *p)
{
    uint8_t i, nb_kni, port_id;
    unsigned nb_rx, num;
    struct rte_mbuf *pkts_burst[PKT_BURST_SZ];

    if (p == NULL)
        return;

    nb_kni = p->nb_kni;
    port_id = p->port_id;

    for (i = 0; i < nb_kni; i++) {
        /* Burst rx from eth */
        nb_rx = rte_eth_rx_burst(port_id, 0, pkts_burst, PKT_BURST_SZ);
        if (unlikely(nb_rx > PKT_BURST_SZ)) {
            RTE_LOG(ERR, APP, "Error receiving from eth\n");
            return;
        }

        /* Burst tx to kni */
        num = rte_kni_tx_burst(p->kni[i], pkts_burst, nb_rx);
        kni_stats[port_id].rx_packets += num;
        rte_kni_handle_request(p->kni[i]);

        if (unlikely(num < nb_rx)) {
            /* Free mbufs not tx to kni interface */
            kni_burst_free_mbufs(&pkts_burst[num], nb_rx - num);
            kni_stats[port_id].rx_dropped += nb_rx - num;
        }
    }
}

For the other case that reads from kernel NIC interfaces and writes to a physical NIC port, packets are retrieved by reading mbufs from kernel NIC interfaces by rte_kni_rx_burst(). The packet transmission is the same as in the L2 Forwarding sample application (see Receive, Process and Transmit Packets).

/**
 *   Interface to dequeue mbufs from tx_q and burst tx
 */

static void

kni_egress(struct kni_port_params *p)
{
    uint8_t i, nb_kni, port_id;
    unsigned nb_tx, num;
    struct rte_mbuf *pkts_burst[PKT_BURST_SZ];

    if (p == NULL)
        return;

    nb_kni = p->nb_kni;
    port_id = p->port_id;

    for (i = 0; i < nb_kni; i++) {
        /* Burst rx from kni */
        num = rte_kni_rx_burst(p->kni[i], pkts_burst, PKT_BURST_SZ);
        if (unlikely(num > PKT_BURST_SZ)) {
            RTE_LOG(ERR, APP, "Error receiving from KNI\n");
            return;
        }

        /* Burst tx to eth */

        nb_tx = rte_eth_tx_burst(port_id, 0, pkts_burst, (uint16_t)num);

        kni_stats[port_id].tx_packets += nb_tx;

        if (unlikely(nb_tx < num)) {
            /* Free mbufs not tx to NIC */
            kni_burst_free_mbufs(&pkts_burst[nb_tx], num - nb_tx);
            kni_stats[port_id].tx_dropped += num - nb_tx;
        }
    }
}
Callbacks for Kernel Requests

To execute specific PMD operations in user space requested by some Linux* commands, callbacks must be implemented and filled in the struct rte_kni_ops structure. Currently, setting a new MTU and configuring the network interface (up/ down) are supported.

static struct rte_kni_ops kni_ops = {
    .change_mtu = kni_change_mtu,
    .config_network_if = kni_config_network_interface,
};

/* Callback for request of changing MTU */

static int
kni_change_mtu(uint16_t port_id, unsigned new_mtu)
{
    int ret;
    struct rte_eth_conf conf;

    if (port_id >= rte_eth_dev_count()) {
        RTE_LOG(ERR, APP, "Invalid port id %d\n", port_id);
        return -EINVAL;
    }

    RTE_LOG(INFO, APP, "Change MTU of port %d to %u\n", port_id, new_mtu);

    /* Stop specific port */

    rte_eth_dev_stop(port_id);

    memcpy(&conf, &port_conf, sizeof(conf));

    /* Set new MTU */

    if (new_mtu > ETHER_MAX_LEN)
        conf.rxmode.jumbo_frame = 1;
    else
        conf.rxmode.jumbo_frame = 0;

    /* mtu + length of header + length of FCS = max pkt length */

    conf.rxmode.max_rx_pkt_len = new_mtu + KNI_ENET_HEADER_SIZE + KNI_ENET_FCS_SIZE;

    ret = rte_eth_dev_configure(port_id, 1, 1, &conf);
    if (ret < 0) {
        RTE_LOG(ERR, APP, "Fail to reconfigure port %d\n", port_id);
        return ret;
    }

    /* Restart specific port */

    ret = rte_eth_dev_start(port_id);
    if (ret < 0) {
         RTE_LOG(ERR, APP, "Fail to restart port %d\n", port_id);
        return ret;
    }

    return 0;
}

/* Callback for request of configuring network interface up/down */

static int
kni_config_network_interface(uint16_t port_id, uint8_t if_up)
{
    int ret = 0;

    if (port_id >= rte_eth_dev_count() || port_id >= RTE_MAX_ETHPORTS) {
        RTE_LOG(ERR, APP, "Invalid port id %d\n", port_id);
        return -EINVAL;
    }

    RTE_LOG(INFO, APP, "Configure network interface of %d %s\n",

    port_id, if_up ? "up" : "down");

    if (if_up != 0) {
        /* Configure network interface up */
        rte_eth_dev_stop(port_id);
        ret = rte_eth_dev_start(port_id);
    } else /* Configure network interface down */
        rte_eth_dev_stop(port_id);

    if (ret < 0)
        RTE_LOG(ERR, APP, "Failed to start port %d\n", port_id);
    return ret;
}

Keep Alive Sample Application

The Keep Alive application is a simple example of a heartbeat/watchdog for packet processing cores. It demonstrates how to detect ‘failed’ DPDK cores and notify a fault management entity of this failure. Its purpose is to ensure the failure of the core does not result in a fault that is not detectable by a management entity.

Overview

The application demonstrates how to protect against ‘silent outages’ on packet processing cores. A Keep Alive Monitor Agent Core (master) monitors the state of packet processing cores (worker cores) by dispatching pings at a regular time interval (default is 5ms) and monitoring the state of the cores. Cores states are: Alive, MIA, Dead or Buried. MIA indicates a missed ping, and Dead indicates two missed pings within the specified time interval. When a core is Dead, a callback function is invoked to restart the packet processing core; A real life application might use this callback function to notify a higher level fault management entity of the core failure in order to take the appropriate corrective action.

Note: Only the worker cores are monitored. A local (on the host) mechanism or agent to supervise the Keep Alive Monitor Agent Core DPDK core is required to detect its failure.

Note: This application is based on the L2 Forwarding Sample Application (in Real and Virtualized Environments). As such, the initialization and run-time paths are very similar to those of the L2 forwarding application.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the l2fwd_keep_alive sub-directory.

Running the Application

The application has a number of command line options:

./build/l2fwd-keepalive [EAL options] \
        -- -p PORTMASK [-q NQ] [-K PERIOD] [-T PERIOD]

where,

  • p PORTMASK: A hexadecimal bitmask of the ports to configure
  • q NQ: A number of queues (=ports) per lcore (default is 1)
  • K PERIOD: Heartbeat check period in ms(5ms default; 86400 max)
  • T PERIOD: statistics will be refreshed each PERIOD seconds (0 to disable, 10 default, 86400 maximum).

To run the application in linuxapp environment with 4 lcores, 16 ports 8 RX queues per lcore and a ping interval of 10ms, issue the command:

./build/l2fwd-keepalive -l 0-3 -n 4 -- -q 8 -p ffff -K 10

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The following sections provide some explanation of the The Keep-Alive/’Liveliness’ conceptual scheme. As mentioned in the overview section, the initialization and run-time paths are very similar to those of the L2 Forwarding Sample Application (in Real and Virtualized Environments).

The Keep-Alive/’Liveliness’ conceptual scheme:

  • A Keep- Alive Agent Runs every N Milliseconds.
  • DPDK Cores respond to the keep-alive agent.
  • If keep-alive agent detects time-outs, it notifies the fault management entity through a callback function.

The following sections provide some explanation of the code aspects that are specific to the Keep Alive sample application.

The keepalive functionality is initialized with a struct rte_keepalive and the callback function to invoke in the case of a timeout.

rte_global_keepalive_info = rte_keepalive_create(&dead_core, NULL);
if (rte_global_keepalive_info == NULL)
    rte_exit(EXIT_FAILURE, "keepalive_create() failed");

The function that issues the pings keepalive_dispatch_pings() is configured to run every check_period milliseconds.

if (rte_timer_reset(&hb_timer,
        (check_period * rte_get_timer_hz()) / 1000,
        PERIODICAL,
        rte_lcore_id(),
        &rte_keepalive_dispatch_pings,
        rte_global_keepalive_info
        ) != 0 )
    rte_exit(EXIT_FAILURE, "Keepalive setup failure.\n");

The rest of the initialization and run-time path follows the same paths as the the L2 forwarding application. The only addition to the main processing loop is the mark alive functionality and the example random failures.

rte_keepalive_mark_alive(&rte_global_keepalive_info);
cur_tsc = rte_rdtsc();

/* Die randomly within 7 secs for demo purposes.. */
if (cur_tsc - tsc_initial > tsc_lifetime)
break;

The rte_keepalive_mark_alive function simply sets the core state to alive.

static inline void
rte_keepalive_mark_alive(struct rte_keepalive *keepcfg)
{
    keepcfg->state_flags[rte_lcore_id()] = ALIVE;
}

L2 Forwarding with Crypto Sample Application

The L2 Forwarding with Crypto (l2fwd-crypto) sample application is a simple example of packet processing using the Data Plane Development Kit (DPDK), in conjunction with the Cryptodev library.

Overview

The L2 Forwarding with Crypto sample application performs a crypto operation (cipher/hash) specified by the user from command line (or using the default values), with a crypto device capable of doing that operation, for each packet that is received on a RX_PORT and performs L2 forwarding. The destination port is the adjacent port from the enabled portmask, that is, if the first four ports are enabled (portmask 0xf), ports 0 and 1 forward into each other, and ports 2 and 3 forward into each other. Also, if MAC addresses updating is enabled, the MAC addresses are affected as follows:

  • The source MAC address is replaced by the TX_PORT MAC address
  • The destination MAC address is replaced by 02:00:00:00:00:TX_PORT_ID

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the l2fwd-crypt sub-directory.

Running the Application

The application requires a number of command line options:

./build/l2fwd-crypto [EAL options] -- [-p PORTMASK] [-q NQ] [-s] [-T PERIOD] /
[--cdev_type HW/SW/ANY] [--chain HASH_CIPHER/CIPHER_HASH/CIPHER_ONLY/HASH_ONLY/AEAD] /
[--cipher_algo ALGO] [--cipher_op ENCRYPT/DECRYPT] [--cipher_key KEY] /
[--cipher_key_random_size SIZE] [--cipher_iv IV] [--cipher_iv_random_size SIZE] /
[--auth_algo ALGO] [--auth_op GENERATE/VERIFY] [--auth_key KEY] /
[--auth_key_random_size SIZE] [--auth_iv IV] [--auth_iv_random_size SIZE] /
[--aead_algo ALGO] [--aead_op ENCRYPT/DECRYPT] [--aead_key KEY] /
[--aead_key_random_size SIZE] [--aead_iv] [--aead_iv_random_size SIZE] /
[--aad AAD] [--aad_random_size SIZE] /
[--digest size SIZE] [--sessionless] [--cryptodev_mask MASK] /
[--mac-updating] [--no-mac-updating]

where,

  • p PORTMASK: A hexadecimal bitmask of the ports to configure (default is all the ports)

  • q NQ: A number of queues (=ports) per lcore (default is 1)

  • s: manage all ports from single core

  • T PERIOD: statistics will be refreshed each PERIOD seconds

    (0 to disable, 10 default, 86400 maximum)

  • cdev_type: select preferred crypto device type: HW, SW or anything (ANY)

    (default is ANY)

  • chain: select the operation chaining to perform: Cipher->Hash (CIPHER_HASH),

    Hash->Cipher (HASH_CIPHER), Cipher (CIPHER_ONLY), Hash (HASH_ONLY)

    or AEAD (AEAD)

    (default is Cipher->Hash)

  • cipher_algo: select the ciphering algorithm (default is aes-cbc)

  • cipher_op: select the ciphering operation to perform: ENCRYPT or DECRYPT

    (default is ENCRYPT)

  • cipher_key: set the ciphering key to be used. Bytes has to be separated with ”:”

  • cipher_key_random_size: set the size of the ciphering key,

    which will be generated randomly.

    Note that if –cipher_key is used, this will be ignored.

  • cipher_iv: set the cipher IV to be used. Bytes has to be separated with ”:”

  • cipher_iv_random_size: set the size of the cipher IV, which will be generated randomly.

    Note that if –cipher_iv is used, this will be ignored.

  • auth_algo: select the authentication algorithm (default is sha1-hmac)

  • auth_op: select the authentication operation to perform: GENERATE or VERIFY

    (default is GENERATE)

  • auth_key: set the authentication key to be used. Bytes has to be separated with ”:”

  • auth_key_random_size: set the size of the authentication key,

    which will be generated randomly.

    Note that if –auth_key is used, this will be ignored.

  • auth_iv: set the auth IV to be used. Bytes has to be separated with ”:”

  • auth_iv_random_size: set the size of the auth IV, which will be generated randomly.

    Note that if –auth_iv is used, this will be ignored.

  • aead_algo: select the AEAD algorithm (default is aes-gcm)

  • aead_op: select the AEAD operation to perform: ENCRYPT or DECRYPT

    (default is ENCRYPT)

  • aead_key: set the AEAD key to be used. Bytes has to be separated with ”:”

  • aead_key_random_size: set the size of the AEAD key,

    which will be generated randomly.

    Note that if –aead_key is used, this will be ignored.

  • aead_iv: set the AEAD IV to be used. Bytes has to be separated with ”:”

  • aead_iv_random_size: set the size of the AEAD IV, which will be generated randomly.

    Note that if –aead_iv is used, this will be ignored.

  • aad: set the AAD to be used. Bytes has to be separated with ”:”

  • aad_random_size: set the size of the AAD, which will be generated randomly.

    Note that if –aad is used, this will be ignored.

  • digest_size: set the size of the digest to be generated/verified.

  • sessionless: no crypto session will be created.

  • cryptodev_mask: A hexadecimal bitmask of the cryptodevs to be used by the application.

    (default is all cryptodevs).

  • [no-]mac-updating: Enable or disable MAC addresses updating (enabled by default).

The application requires that crypto devices capable of performing the specified crypto operation are available on application initialization. This means that HW crypto device/s must be bound to a DPDK driver or a SW crypto device/s (virtual crypto PMD) must be created (using –vdev).

To run the application in linuxapp environment with 2 lcores, 2 ports and 2 crypto devices, issue the command:

$ ./build/l2fwd-crypto -l 0-1 -n 4 --vdev "crypto_aesni_mb0" \
--vdev "crypto_aesni_mb1" -- -p 0x3 --chain CIPHER_HASH \
--cipher_op ENCRYPT --cipher_algo aes-cbc \
--cipher_key 00:01:02:03:04:05:06:07:08:09:0a:0b:0c:0d:0e:0f \
--auth_op GENERATE --auth_algo aes-xcbc-mac \
--auth_key 10:11:12:13:14:15:16:17:18:19:1a:1b:1c:1d:1e:1f

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The L2 forward with Crypto application demonstrates the performance of a crypto operation on a packet received on a RX PORT before forwarding it to a TX PORT.

The following figure illustrates a sample flow of a packet in the application, from reception until transmission.

_images/l2_fwd_encrypt_flow.svg

Encryption flow Through the L2 Forwarding with Crypto Application

The following sections provide some explanation of the application.

Crypto operation specification

All the packets received in all the ports get transformed by the crypto device/s (ciphering and/or authentication). The crypto operation to be performed on the packet is parsed from the command line (go to “Running the Application section for all the options).

If no parameter is passed, the default crypto operation is:

  • Encryption with AES-CBC with 128 bit key.
  • Authentication with SHA1-HMAC (generation).
  • Keys, IV and AAD are generated randomly.

There are two methods to pass keys, IV and ADD from the command line:

  • Passing the full key, separated bytes by ”:”:

    --cipher_key 00:11:22:33:44
    
  • Passing the size, so key is generated randomly:

    --cipher_key_random_size 16
    
Note:
If full key is passed (first method) and the size is passed as well (second method), the latter will be ignored.

Size of these keys are checked (regardless the method), before starting the app, to make sure that it is supported by the crypto devices.

Crypto device initialization

Once the encryption operation is defined, crypto devices are initialized. The crypto devices must be either bound to a DPDK driver (if they are physical devices) or created using the EAL option –vdev (if they are virtual devices), when running the application.

The initialize_cryptodevs() function performs the device initialization. It iterates through the list of the available crypto devices and check which ones are capable of performing the operation. Each device has a set of capabilities associated with it, which are stored in the device info structure, so the function checks if the operation is within the structure of each device.

The following code checks if the device supports the specified cipher algorithm (similar for the authentication algorithm):

/* Check if device supports cipher algo */
i = 0;
opt_cipher_algo = options->cipher_xform.cipher.algo;
cap = &dev_info.capabilities[i];
while (cap->op != RTE_CRYPTO_OP_TYPE_UNDEFINED) {
        cap_cipher_algo = cap->sym.cipher.algo;
        if (cap->sym.xform_type ==
                        RTE_CRYPTO_SYM_XFORM_CIPHER) {
                if (cap_cipher_algo == opt_cipher_algo) {
                        if (check_type(options, &dev_info) == 0)
                                break;
                }
        }
        cap = &dev_info.capabilities[++i];
}

If a capable crypto device is found, key sizes are checked to see if they are supported (cipher key and IV for the ciphering):

/*
 * Check if length of provided cipher key is supported
 * by the algorithm chosen.
 */
if (options->ckey_param) {
        if (check_supported_size(
                        options->cipher_xform.cipher.key.length,
                        cap->sym.cipher.key_size.min,
                        cap->sym.cipher.key_size.max,
                        cap->sym.cipher.key_size.increment)
                                != 0) {
                printf("Unsupported cipher key length\n");
                return -1;
        }
/*
 * Check if length of the cipher key to be randomly generated
 * is supported by the algorithm chosen.
 */
} else if (options->ckey_random_size != -1) {
        if (check_supported_size(options->ckey_random_size,
                        cap->sym.cipher.key_size.min,
                        cap->sym.cipher.key_size.max,
                        cap->sym.cipher.key_size.increment)
                                != 0) {
                printf("Unsupported cipher key length\n");
                return -1;
        }
        options->cipher_xform.cipher.key.length =
                                options->ckey_random_size;
/* No size provided, use minimum size. */
} else
        options->cipher_xform.cipher.key.length =
                        cap->sym.cipher.key_size.min;

After all the checks, the device is configured and it is added to the crypto device list.

Note:
The number of crypto devices that supports the specified crypto operation must be at least the number of ports to be used.
Session creation

The crypto operation has a crypto session associated to it, which contains information such as the transform chain to perform (e.g. ciphering then hashing), pointers to the keys, lengths... etc.

This session is created and is later attached to the crypto operation:

static struct rte_cryptodev_sym_session *
initialize_crypto_session(struct l2fwd_crypto_options *options,
                uint8_t cdev_id)
{
        struct rte_crypto_sym_xform *first_xform;
        struct rte_cryptodev_sym_session *session;
        uint8_t socket_id = rte_cryptodev_socket_id(cdev_id);
        struct rte_mempool *sess_mp = session_pool_socket[socket_id];


        if (options->xform_chain == L2FWD_CRYPTO_AEAD) {
                first_xform = &options->aead_xform;
        } else if (options->xform_chain == L2FWD_CRYPTO_CIPHER_HASH) {
                first_xform = &options->cipher_xform;
                first_xform->next = &options->auth_xform;
        } else if (options->xform_chain == L2FWD_CRYPTO_HASH_CIPHER) {
                first_xform = &options->auth_xform;
                first_xform->next = &options->cipher_xform;
        } else if (options->xform_chain == L2FWD_CRYPTO_CIPHER_ONLY) {
                first_xform = &options->cipher_xform;
        } else {
                first_xform = &options->auth_xform;
        }

        session = rte_cryptodev_sym_session_create(sess_mp);

        if (session == NULL)
                return NULL;

       if (rte_cryptodev_sym_session_init(cdev_id, session,
                             first_xform, sess_mp) < 0)
                return NULL;

       return session;
}

...

port_cparams[i].session = initialize_crypto_session(options,
                             port_cparams[i].dev_id);
Crypto operation creation

Given N packets received from a RX PORT, N crypto operations are allocated and filled:

if (nb_rx) {
/*
 * If we can't allocate a crypto_ops, then drop
 * the rest of the burst and dequeue and
 * process the packets to free offload structs
 */
if (rte_crypto_op_bulk_alloc(
                l2fwd_crypto_op_pool,
                RTE_CRYPTO_OP_TYPE_SYMMETRIC,
                ops_burst, nb_rx) !=
                                nb_rx) {
        for (j = 0; j < nb_rx; j++)
                rte_pktmbuf_free(pkts_burst[i]);

        nb_rx = 0;
}

After filling the crypto operation (including session attachment), the mbuf which will be transformed is attached to it:

op->sym->m_src = m;

Since no destination mbuf is set, the source mbuf will be overwritten after the operation is done (in-place).

Crypto operation enqueuing/dequeuing

Once the operation has been created, it has to be enqueued in one of the crypto devices. Before doing so, for performance reasons, the operation stays in a buffer. When the buffer has enough operations (MAX_PKT_BURST), they are enqueued in the device, which will perform the operation at that moment:

static int
l2fwd_crypto_enqueue(struct rte_crypto_op *op,
                struct l2fwd_crypto_params *cparams)
{
        unsigned lcore_id, len;
        struct lcore_queue_conf *qconf;

        lcore_id = rte_lcore_id();

        qconf = &lcore_queue_conf[lcore_id];
        len = qconf->op_buf[cparams->dev_id].len;
        qconf->op_buf[cparams->dev_id].buffer[len] = op;
        len++;

        /* enough ops to be sent */
        if (len == MAX_PKT_BURST) {
                l2fwd_crypto_send_burst(qconf, MAX_PKT_BURST, cparams);
                len = 0;
        }

        qconf->op_buf[cparams->dev_id].len = len;
        return 0;
}

...

static int
l2fwd_crypto_send_burst(struct lcore_queue_conf *qconf, unsigned n,
                struct l2fwd_crypto_params *cparams)
{
        struct rte_crypto_op **op_buffer;
        unsigned ret;

        op_buffer = (struct rte_crypto_op **)
                        qconf->op_buf[cparams->dev_id].buffer;

        ret = rte_cryptodev_enqueue_burst(cparams->dev_id,
                        cparams->qp_id, op_buffer, (uint16_t) n);

        crypto_statistics[cparams->dev_id].enqueued += ret;
        if (unlikely(ret < n)) {
                crypto_statistics[cparams->dev_id].errors += (n - ret);
                do {
                        rte_pktmbuf_free(op_buffer[ret]->sym->m_src);
                        rte_crypto_op_free(op_buffer[ret]);
                } while (++ret < n);
        }

        return 0;
}

After this, the operations are dequeued from the device, and the transformed mbuf is extracted from the operation. Then, the operation is freed and the mbuf is forwarded as it is done in the L2 forwarding application.

/* Dequeue packets from Crypto device */
do {
        nb_rx = rte_cryptodev_dequeue_burst(
                        cparams->dev_id, cparams->qp_id,
                        ops_burst, MAX_PKT_BURST);

        crypto_statistics[cparams->dev_id].dequeued +=
                        nb_rx;

        /* Forward crypto'd packets */
        for (j = 0; j < nb_rx; j++) {
                m = ops_burst[j]->sym->m_src;

                rte_crypto_op_free(ops_burst[j]);
                l2fwd_simple_forward(m, portid);
        }
} while (nb_rx == MAX_PKT_BURST);

L2 Forwarding Sample Application (in Real and Virtualized Environments) with core load statistics.

The L2 Forwarding sample application is a simple example of packet processing using the Data Plane Development Kit (DPDK) which also takes advantage of Single Root I/O Virtualization (SR-IOV) features in a virtualized environment.

Note

This application is a variation of L2 Forwarding sample application. It demonstrate possible scheme of job stats library usage therefore some parts of this document is identical with original L2 forwarding application.

Overview

The L2 Forwarding sample application, which can operate in real and virtualized environments, performs L2 forwarding for each packet that is received. The destination port is the adjacent port from the enabled portmask, that is, if the first four ports are enabled (portmask 0xf), ports 1 and 2 forward into each other, and ports 3 and 4 forward into each other. Also, the MAC addresses are affected as follows:

  • The source MAC address is replaced by the TX port MAC address
  • The destination MAC address is replaced by 02:00:00:00:00:TX_PORT_ID

This application can be used to benchmark performance using a traffic-generator, as shown in the Fig. 17.1.

The application can also be used in a virtualized environment as shown in Fig. 17.2.

The L2 Forwarding application can also be used as a starting point for developing a new application based on the DPDK.

_images/l2_fwd_benchmark_setup.svg

Performance Benchmark Setup (Basic Environment)

_images/l2_fwd_virtenv_benchmark_setup.png

Performance Benchmark Setup (Virtualized Environment)

Virtual Function Setup Instructions

This application can use the virtual function available in the system and therefore can be used in a virtual machine without passing through the whole Network Device into a guest machine in a virtualized scenario. The virtual functions can be enabled in the host machine or the hypervisor with the respective physical function driver.

For example, in a Linux* host machine, it is possible to enable a virtual function using the following command:

modprobe ixgbe max_vfs=2,2

This command enables two Virtual Functions on each of Physical Function of the NIC, with two physical ports in the PCI configuration space. It is important to note that enabled Virtual Function 0 and 2 would belong to Physical Function 0 and Virtual Function 1 and 3 would belong to Physical Function 1, in this case enabling a total of four Virtual Functions.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the l2fwd-jobstats sub-directory.

Running the Application

The application requires a number of command line options:

./build/l2fwd-jobstats [EAL options] -- -p PORTMASK [-q NQ] [-l]

where,

  • p PORTMASK: A hexadecimal bitmask of the ports to configure
  • q NQ: A number of queues (=ports) per lcore (default is 1)
  • l: Use locale thousands separator when formatting big numbers.

To run the application in linuxapp environment with 4 lcores, 16 ports, 8 RX queues per lcore and thousands separator printing, issue the command:

$ ./build/l2fwd-jobstats -l 0-3 -n 4 -- -q 8 -p ffff -l

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The following sections provide some explanation of the code.

Command Line Arguments

The L2 Forwarding sample application takes specific parameters, in addition to Environment Abstraction Layer (EAL) arguments (see Running the Application). The preferred way to parse parameters is to use the getopt() function, since it is part of a well-defined and portable library.

The parsing of arguments is done in the l2fwd_parse_args() function. The method of argument parsing is not described here. Refer to the glibc getopt(3) man page for details.

EAL arguments are parsed first, then application-specific arguments. This is done at the beginning of the main() function:

/* init EAL */

ret = rte_eal_init(argc, argv);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Invalid EAL arguments\n");

argc -= ret;
argv += ret;

/* parse application arguments (after the EAL ones) */

ret = l2fwd_parse_args(argc, argv);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Invalid L2FWD arguments\n");
Mbuf Pool Initialization

Once the arguments are parsed, the mbuf pool is created. The mbuf pool contains a set of mbuf objects that will be used by the driver and the application to store network packet data:

/* create the mbuf pool */
l2fwd_pktmbuf_pool = rte_pktmbuf_pool_create("mbuf_pool", NB_MBUF,
            MEMPOOL_CACHE_SIZE, 0, RTE_MBUF_DEFAULT_BUF_SIZE,
            rte_socket_id());

if (l2fwd_pktmbuf_pool == NULL)
    rte_exit(EXIT_FAILURE, "Cannot init mbuf pool\n");

The rte_mempool is a generic structure used to handle pools of objects. In this case, it is necessary to create a pool that will be used by the driver. The number of allocated pkt mbufs is NB_MBUF, with a data room size of RTE_MBUF_DEFAULT_BUF_SIZE each. A per-lcore cache of MEMPOOL_CACHE_SIZE mbufs is kept. The memory is allocated in rte_socket_id() socket, but it is possible to extend this code to allocate one mbuf pool per socket.

The rte_pktmbuf_pool_create() function uses the default mbuf pool and mbuf initializers, respectively rte_pktmbuf_pool_init() and rte_pktmbuf_init(). An advanced application may want to use the mempool API to create the mbuf pool with more control.

Driver Initialization

The main part of the code in the main() function relates to the initialization of the driver. To fully understand this code, it is recommended to study the chapters that related to the Poll Mode Driver in the DPDK Programmer’s Guide and the DPDK API Reference.

nb_ports = rte_eth_dev_count();

if (nb_ports == 0)
    rte_exit(EXIT_FAILURE, "No Ethernet ports - bye\n");

/* reset l2fwd_dst_ports */

for (portid = 0; portid < RTE_MAX_ETHPORTS; portid++)
    l2fwd_dst_ports[portid] = 0;

last_port = 0;

/*
 * Each logical core is assigned a dedicated TX queue on each port.
 */
for (portid = 0; portid < nb_ports; portid++) {
    /* skip ports that are not enabled */
    if ((l2fwd_enabled_port_mask & (1 << portid)) == 0)
       continue;

    if (nb_ports_in_mask % 2) {
        l2fwd_dst_ports[portid] = last_port;
        l2fwd_dst_ports[last_port] = portid;
    }
    else
       last_port = portid;

    nb_ports_in_mask++;

    rte_eth_dev_info_get((uint8_t) portid, &dev_info);
}

The next step is to configure the RX and TX queues. For each port, there is only one RX queue (only one lcore is able to poll a given port). The number of TX queues depends on the number of available lcores. The rte_eth_dev_configure() function is used to configure the number of queues for a port:

ret = rte_eth_dev_configure((uint8_t)portid, 1, 1, &port_conf);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Cannot configure device: "
        "err=%d, port=%u\n",
        ret, portid);

The global configuration is stored in a static structure:

static const struct rte_eth_conf port_conf = {
    .rxmode = {
        .split_hdr_size = 0,
        .header_split = 0,   /**< Header Split disabled */
        .hw_ip_checksum = 0, /**< IP checksum offload disabled */
        .hw_vlan_filter = 0, /**< VLAN filtering disabled */
        .jumbo_frame = 0,    /**< Jumbo Frame Support disabled */
        .hw_strip_crc= 0,    /**< CRC stripped by hardware */
    },

    .txmode = {
        .mq_mode = ETH_DCB_NONE
    },
};
RX Queue Initialization

The application uses one lcore to poll one or several ports, depending on the -q option, which specifies the number of queues per lcore.

For example, if the user specifies -q 4, the application is able to poll four ports with one lcore. If there are 16 ports on the target (and if the portmask argument is -p ffff ), the application will need four lcores to poll all the ports.

ret = rte_eth_rx_queue_setup(portid, 0, nb_rxd,
            rte_eth_dev_socket_id(portid),
            NULL,
            l2fwd_pktmbuf_pool);

if (ret < 0)
    rte_exit(EXIT_FAILURE, "rte_eth_rx_queue_setup:err=%d, port=%u\n",
            ret, (unsigned) portid);

The list of queues that must be polled for a given lcore is stored in a private structure called struct lcore_queue_conf.

struct lcore_queue_conf {
    unsigned n_rx_port;
    unsigned rx_port_list[MAX_RX_QUEUE_PER_LCORE];
    truct mbuf_table tx_mbufs[RTE_MAX_ETHPORTS];

    struct rte_timer rx_timers[MAX_RX_QUEUE_PER_LCORE];
    struct rte_jobstats port_fwd_jobs[MAX_RX_QUEUE_PER_LCORE];

    struct rte_timer flush_timer;
    struct rte_jobstats flush_job;
    struct rte_jobstats idle_job;
    struct rte_jobstats_context jobs_context;

    rte_atomic16_t stats_read_pending;
    rte_spinlock_t lock;
} __rte_cache_aligned;

Values of struct lcore_queue_conf:

  • n_rx_port and rx_port_list[] are used in the main packet processing loop (see Section Receive, Process and Transmit Packets later in this chapter).
  • rx_timers and flush_timer are used to ensure forced TX on low packet rate.
  • flush_job, idle_job and jobs_context are librte_jobstats objects used for managing l2fwd jobs.
  • stats_read_pending and lock are used during job stats read phase.
TX Queue Initialization

Each lcore should be able to transmit on any port. For every port, a single TX queue is initialized.

/* init one TX queue on each port */

fflush(stdout);
ret = rte_eth_tx_queue_setup(portid, 0, nb_txd,
        rte_eth_dev_socket_id(portid),
        NULL);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "rte_eth_tx_queue_setup:err=%d, port=%u\n",
            ret, (unsigned) portid);
Jobs statistics initialization

There are several statistics objects available:

  • Flush job statistics
rte_jobstats_init(&qconf->flush_job, "flush", drain_tsc, drain_tsc,
        drain_tsc, 0);

rte_timer_init(&qconf->flush_timer);
ret = rte_timer_reset(&qconf->flush_timer, drain_tsc, PERIODICAL,
            lcore_id, &l2fwd_flush_job, NULL);

if (ret < 0) {
    rte_exit(1, "Failed to reset flush job timer for lcore %u: %s",
                lcore_id, rte_strerror(-ret));
}
  • Statistics per RX port
rte_jobstats_init(job, name, 0, drain_tsc, 0, MAX_PKT_BURST);
rte_jobstats_set_update_period_function(job, l2fwd_job_update_cb);

rte_timer_init(&qconf->rx_timers[i]);
ret = rte_timer_reset(&qconf->rx_timers[i], 0, PERIODICAL, lcore_id,
        l2fwd_fwd_job, (void *)(uintptr_t)i);

if (ret < 0) {
    rte_exit(1, "Failed to reset lcore %u port %u job timer: %s",
                lcore_id, qconf->rx_port_list[i], rte_strerror(-ret));
}

Following parameters are passed to rte_jobstats_init():

  • 0 as minimal poll period
  • drain_tsc as maximum poll period
  • MAX_PKT_BURST as desired target value (RX burst size)
Main loop

The forwarding path is reworked comparing to original L2 Forwarding application. In the l2fwd_main_loop() function three loops are placed.

for (;;) {
    rte_spinlock_lock(&qconf->lock);

    do {
        rte_jobstats_context_start(&qconf->jobs_context);

        /* Do the Idle job:
         * - Read stats_read_pending flag
         * - check if some real job need to be executed
         */
        rte_jobstats_start(&qconf->jobs_context, &qconf->idle_job);

        do {
            uint8_t i;
            uint64_t now = rte_get_timer_cycles();

            need_manage = qconf->flush_timer.expire < now;
            /* Check if we was esked to give a stats. */
            stats_read_pending =
                    rte_atomic16_read(&qconf->stats_read_pending);
            need_manage |= stats_read_pending;

            for (i = 0; i < qconf->n_rx_port && !need_manage; i++)
                need_manage = qconf->rx_timers[i].expire < now;

        } while (!need_manage);
        rte_jobstats_finish(&qconf->idle_job, qconf->idle_job.target);

        rte_timer_manage();
        rte_jobstats_context_finish(&qconf->jobs_context);
    } while (likely(stats_read_pending == 0));

    rte_spinlock_unlock(&qconf->lock);
    rte_pause();
}

First infinite for loop is to minimize impact of stats reading. Lock is only locked/unlocked when asked.

Second inner while loop do the whole jobs management. When any job is ready, the use rte_timer_manage() is used to call the job handler. In this place functions l2fwd_fwd_job() and l2fwd_flush_job() are called when needed. Then rte_jobstats_context_finish() is called to mark loop end - no other jobs are ready to execute. By this time stats are ready to be read and if stats_read_pending is set, loop breaks allowing stats to be read.

Third do-while loop is the idle job (idle stats counter). Its only purpose is monitoring if any job is ready or stats job read is pending for this lcore. Statistics from this part of code is considered as the headroom available for additional processing.

Receive, Process and Transmit Packets

The main task of l2fwd_fwd_job() function is to read ingress packets from the RX queue of particular port and forward it. This is done using the following code:

total_nb_rx = rte_eth_rx_burst((uint8_t) portid, 0, pkts_burst,
        MAX_PKT_BURST);

for (j = 0; j < total_nb_rx; j++) {
    m = pkts_burst[j];
    rte_prefetch0(rte_pktmbuf_mtod(m, void *));
    l2fwd_simple_forward(m, portid);
}

Packets are read in a burst of size MAX_PKT_BURST. Then, each mbuf in the table is processed by the l2fwd_simple_forward() function. The processing is very simple: process the TX port from the RX port, then replace the source and destination MAC addresses.

The rte_eth_rx_burst() function writes the mbuf pointers in a local table and returns the number of available mbufs in the table.

After first read second try is issued.

if (total_nb_rx == MAX_PKT_BURST) {
    const uint16_t nb_rx = rte_eth_rx_burst((uint8_t) portid, 0, pkts_burst,
            MAX_PKT_BURST);

    total_nb_rx += nb_rx;
    for (j = 0; j < nb_rx; j++) {
        m = pkts_burst[j];
        rte_prefetch0(rte_pktmbuf_mtod(m, void *));
        l2fwd_simple_forward(m, portid);
    }
}

This second read is important to give job stats library a feedback how many packets was processed.

/* Adjust period time in which we are running here. */
if (rte_jobstats_finish(job, total_nb_rx) != 0) {
    rte_timer_reset(&qconf->rx_timers[port_idx], job->period, PERIODICAL,
            lcore_id, l2fwd_fwd_job, arg);
}

To maximize performance exactly MAX_PKT_BURST is expected (the target value) to be read for each l2fwd_fwd_job() call. If total_nb_rx is smaller than target value job->period will be increased. If it is greater the period will be decreased.

Note

In the following code, one line for getting the output port requires some explanation.

During the initialization process, a static array of destination ports (l2fwd_dst_ports[]) is filled such that for each source port, a destination port is assigned that is either the next or previous enabled port from the portmask. Naturally, the number of ports in the portmask must be even, otherwise, the application exits.

static void
l2fwd_simple_forward(struct rte_mbuf *m, unsigned portid)
{
    struct ether_hdr *eth;
    void *tmp;
    unsigned dst_port;

    dst_port = l2fwd_dst_ports[portid];

    eth = rte_pktmbuf_mtod(m, struct ether_hdr *);

    /* 02:00:00:00:00:xx */

    tmp = &eth->d_addr.addr_bytes[0];

    *((uint64_t *)tmp) = 0x000000000002 + ((uint64_t) dst_port << 40);

    /* src addr */

    ether_addr_copy(&l2fwd_ports_eth_addr[dst_port], &eth->s_addr);

    l2fwd_send_packet(m, (uint8_t) dst_port);
}

Then, the packet is sent using the l2fwd_send_packet (m, dst_port) function. For this test application, the processing is exactly the same for all packets arriving on the same RX port. Therefore, it would have been possible to call the l2fwd_send_burst() function directly from the main loop to send all the received packets on the same TX port, using the burst-oriented send function, which is more efficient.

However, in real-life applications (such as, L3 routing), packet N is not necessarily forwarded on the same port as packet N-1. The application is implemented to illustrate that, so the same approach can be reused in a more complex application.

The l2fwd_send_packet() function stores the packet in a per-lcore and per-txport table. If the table is full, the whole packets table is transmitted using the l2fwd_send_burst() function:

/* Send the packet on an output interface */

static int
l2fwd_send_packet(struct rte_mbuf *m, uint16_t port)
{
    unsigned lcore_id, len;
    struct lcore_queue_conf *qconf;

    lcore_id = rte_lcore_id();
    qconf = &lcore_queue_conf[lcore_id];
    len = qconf->tx_mbufs[port].len;
    qconf->tx_mbufs[port].m_table[len] = m;
    len++;

    /* enough pkts to be sent */

    if (unlikely(len == MAX_PKT_BURST)) {
        l2fwd_send_burst(qconf, MAX_PKT_BURST, port);
        len = 0;
    }

    qconf->tx_mbufs[port].len = len; return 0;
}

To ensure that no packets remain in the tables, the flush job exists. The l2fwd_flush_job() is called periodically to for each lcore draining TX queue of each port. This technique introduces some latency when there are not many packets to send, however it improves performance:

static void
l2fwd_flush_job(__rte_unused struct rte_timer *timer, __rte_unused void *arg)
{
    uint64_t now;
    unsigned lcore_id;
    struct lcore_queue_conf *qconf;
    struct mbuf_table *m_table;
    uint16_t portid;

    lcore_id = rte_lcore_id();
    qconf = &lcore_queue_conf[lcore_id];

    rte_jobstats_start(&qconf->jobs_context, &qconf->flush_job);

    now = rte_get_timer_cycles();
    lcore_id = rte_lcore_id();
    qconf = &lcore_queue_conf[lcore_id];
    for (portid = 0; portid < RTE_MAX_ETHPORTS; portid++) {
        m_table = &qconf->tx_mbufs[portid];
        if (m_table->len == 0 || m_table->next_flush_time <= now)
            continue;

        l2fwd_send_burst(qconf, portid);
    }


    /* Pass target to indicate that this job is happy of time interval
     * in which it was called. */
    rte_jobstats_finish(&qconf->flush_job, qconf->flush_job.target);
}

L2 Forwarding Sample Application (in Real and Virtualized Environments)

The L2 Forwarding sample application is a simple example of packet processing using the Data Plane Development Kit (DPDK) which also takes advantage of Single Root I/O Virtualization (SR-IOV) features in a virtualized environment.

Note

Please note that previously a separate L2 Forwarding in Virtualized Environments sample application was used, however, in later DPDK versions these sample applications have been merged.

Overview

The L2 Forwarding sample application, which can operate in real and virtualized environments, performs L2 forwarding for each packet that is received on an RX_PORT. The destination port is the adjacent port from the enabled portmask, that is, if the first four ports are enabled (portmask 0xf), ports 1 and 2 forward into each other, and ports 3 and 4 forward into each other. Also, if MAC addresses updating is enabled, the MAC addresses are affected as follows:

  • The source MAC address is replaced by the TX_PORT MAC address
  • The destination MAC address is replaced by 02:00:00:00:00:TX_PORT_ID

This application can be used to benchmark performance using a traffic-generator, as shown in the Fig. 18.1, or in a virtualized environment as shown in Fig. 18.2.

_images/l2_fwd_benchmark_setup.svg

Performance Benchmark Setup (Basic Environment)

_images/l2_fwd_virtenv_benchmark_setup.png

Performance Benchmark Setup (Virtualized Environment)

This application may be used for basic VM to VM communication as shown in Fig. 18.3, when MAC addresses updating is disabled.

_images/l2_fwd_vm2vm.svg

Virtual Machine to Virtual Machine communication.

The L2 Forwarding application can also be used as a starting point for developing a new application based on the DPDK.

Virtual Function Setup Instructions

This application can use the virtual function available in the system and therefore can be used in a virtual machine without passing through the whole Network Device into a guest machine in a virtualized scenario. The virtual functions can be enabled in the host machine or the hypervisor with the respective physical function driver.

For example, in a Linux* host machine, it is possible to enable a virtual function using the following command:

modprobe ixgbe max_vfs=2,2

This command enables two Virtual Functions on each of Physical Function of the NIC, with two physical ports in the PCI configuration space. It is important to note that enabled Virtual Function 0 and 2 would belong to Physical Function 0 and Virtual Function 1 and 3 would belong to Physical Function 1, in this case enabling a total of four Virtual Functions.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the l2fwd sub-directory.

Running the Application

The application requires a number of command line options:

./build/l2fwd [EAL options] -- -p PORTMASK [-q NQ] --[no-]mac-updating

where,

  • p PORTMASK: A hexadecimal bitmask of the ports to configure
  • q NQ: A number of queues (=ports) per lcore (default is 1)
  • –[no-]mac-updating: Enable or disable MAC addresses updating (enabled by default).

To run the application in linuxapp environment with 4 lcores, 16 ports and 8 RX queues per lcore and MAC address updating enabled, issue the command:

$ ./build/l2fwd -l 0-3 -n 4 -- -q 8 -p ffff

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The following sections provide some explanation of the code.

Command Line Arguments

The L2 Forwarding sample application takes specific parameters, in addition to Environment Abstraction Layer (EAL) arguments. The preferred way to parse parameters is to use the getopt() function, since it is part of a well-defined and portable library.

The parsing of arguments is done in the l2fwd_parse_args() function. The method of argument parsing is not described here. Refer to the glibc getopt(3) man page for details.

EAL arguments are parsed first, then application-specific arguments. This is done at the beginning of the main() function:

/* init EAL */

ret = rte_eal_init(argc, argv);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Invalid EAL arguments\n");

argc -= ret;
argv += ret;

/* parse application arguments (after the EAL ones) */

ret = l2fwd_parse_args(argc, argv);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Invalid L2FWD arguments\n");
Mbuf Pool Initialization

Once the arguments are parsed, the mbuf pool is created. The mbuf pool contains a set of mbuf objects that will be used by the driver and the application to store network packet data:

/* create the mbuf pool */

l2fwd_pktmbuf_pool = rte_pktmbuf_pool_create("mbuf_pool", NB_MBUF,
    MEMPOOL_CACHE_SIZE, 0, RTE_MBUF_DEFAULT_BUF_SIZE,
    rte_socket_id());

if (l2fwd_pktmbuf_pool == NULL)
    rte_panic("Cannot init mbuf pool\n");

The rte_mempool is a generic structure used to handle pools of objects. In this case, it is necessary to create a pool that will be used by the driver. The number of allocated pkt mbufs is NB_MBUF, with a data room size of RTE_MBUF_DEFAULT_BUF_SIZE each. A per-lcore cache of 32 mbufs is kept. The memory is allocated in NUMA socket 0, but it is possible to extend this code to allocate one mbuf pool per socket.

The rte_pktmbuf_pool_create() function uses the default mbuf pool and mbuf initializers, respectively rte_pktmbuf_pool_init() and rte_pktmbuf_init(). An advanced application may want to use the mempool API to create the mbuf pool with more control.

Driver Initialization

The main part of the code in the main() function relates to the initialization of the driver. To fully understand this code, it is recommended to study the chapters that related to the Poll Mode Driver in the DPDK Programmer’s Guide - Rel 1.4 EAR and the DPDK API Reference.

if (rte_pci_probe() < 0)
    rte_exit(EXIT_FAILURE, "Cannot probe PCI\n");

nb_ports = rte_eth_dev_count();

if (nb_ports == 0)
    rte_exit(EXIT_FAILURE, "No Ethernet ports - bye\n");

/* reset l2fwd_dst_ports */

for (portid = 0; portid < RTE_MAX_ETHPORTS; portid++)
    l2fwd_dst_ports[portid] = 0;

last_port = 0;

/*
 * Each logical core is assigned a dedicated TX queue on each port.
 */

for (portid = 0; portid < nb_ports; portid++) {
    /* skip ports that are not enabled */

    if ((l2fwd_enabled_port_mask & (1 << portid)) == 0)
       continue;

    if (nb_ports_in_mask % 2) {
        l2fwd_dst_ports[portid] = last_port;
        l2fwd_dst_ports[last_port] = portid;
    }
    else
       last_port = portid;

    nb_ports_in_mask++;

    rte_eth_dev_info_get((uint8_t) portid, &dev_info);
}

Observe that:

  • rte_igb_pmd_init_all() simultaneously registers the driver as a PCI driver and as an Ethernet* Poll Mode Driver.
  • rte_pci_probe() parses the devices on the PCI bus and initializes recognized devices.

The next step is to configure the RX and TX queues. For each port, there is only one RX queue (only one lcore is able to poll a given port). The number of TX queues depends on the number of available lcores. The rte_eth_dev_configure() function is used to configure the number of queues for a port:

ret = rte_eth_dev_configure((uint8_t)portid, 1, 1, &port_conf);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Cannot configure device: "
        "err=%d, port=%u\n",
        ret, portid);

The global configuration is stored in a static structure:

static const struct rte_eth_conf port_conf = {
    .rxmode = {
        .split_hdr_size = 0,
        .header_split = 0,   /**< Header Split disabled */
        .hw_ip_checksum = 0, /**< IP checksum offload disabled */
        .hw_vlan_filter = 0, /**< VLAN filtering disabled */
        .jumbo_frame = 0,    /**< Jumbo Frame Support disabled */
        .hw_strip_crc= 0,    /**< CRC stripped by hardware */
    },

    .txmode = {
        .mq_mode = ETH_DCB_NONE
    },
};
RX Queue Initialization

The application uses one lcore to poll one or several ports, depending on the -q option, which specifies the number of queues per lcore.

For example, if the user specifies -q 4, the application is able to poll four ports with one lcore. If there are 16 ports on the target (and if the portmask argument is -p ffff ), the application will need four lcores to poll all the ports.

ret = rte_eth_rx_queue_setup((uint8_t) portid, 0, nb_rxd, SOCKET0, &rx_conf, l2fwd_pktmbuf_pool);
if (ret < 0)

    rte_exit(EXIT_FAILURE, "rte_eth_rx_queue_setup: "
        "err=%d, port=%u\n",
        ret, portid);

The list of queues that must be polled for a given lcore is stored in a private structure called struct lcore_queue_conf.

struct lcore_queue_conf {
    unsigned n_rx_port;
    unsigned rx_port_list[MAX_RX_QUEUE_PER_LCORE];
    struct mbuf_table tx_mbufs[L2FWD_MAX_PORTS];
} rte_cache_aligned;

struct lcore_queue_conf lcore_queue_conf[RTE_MAX_LCORE];

The values n_rx_port and rx_port_list[] are used in the main packet processing loop (see Receive, Process and Transmit Packets).

The global configuration for the RX queues is stored in a static structure:

static const struct rte_eth_rxconf rx_conf = {
    .rx_thresh = {
        .pthresh = RX_PTHRESH,
        .hthresh = RX_HTHRESH,
        .wthresh = RX_WTHRESH,
    },
};
TX Queue Initialization

Each lcore should be able to transmit on any port. For every port, a single TX queue is initialized.

/* init one TX queue on each port */

fflush(stdout);

ret = rte_eth_tx_queue_setup((uint8_t) portid, 0, nb_txd, rte_eth_dev_socket_id(portid), &tx_conf);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "rte_eth_tx_queue_setup:err=%d, port=%u\n", ret, (unsigned) portid);

The global configuration for TX queues is stored in a static structure:

static const struct rte_eth_txconf tx_conf = {
    .tx_thresh = {
        .pthresh = TX_PTHRESH,
        .hthresh = TX_HTHRESH,
        .wthresh = TX_WTHRESH,
    },
    .tx_free_thresh = RTE_TEST_TX_DESC_DEFAULT + 1, /* disable feature */
};
Receive, Process and Transmit Packets

In the l2fwd_main_loop() function, the main task is to read ingress packets from the RX queues. This is done using the following code:

/*
 * Read packet from RX queues
 */

for (i = 0; i < qconf->n_rx_port; i++) {
    portid = qconf->rx_port_list[i];
    nb_rx = rte_eth_rx_burst((uint8_t) portid, 0,  pkts_burst, MAX_PKT_BURST);

    for (j = 0; j < nb_rx; j++) {
        m = pkts_burst[j];
        rte_prefetch0[rte_pktmbuf_mtod(m, void *)); l2fwd_simple_forward(m, portid);
    }
}

Packets are read in a burst of size MAX_PKT_BURST. The rte_eth_rx_burst() function writes the mbuf pointers in a local table and returns the number of available mbufs in the table.

Then, each mbuf in the table is processed by the l2fwd_simple_forward() function. The processing is very simple: process the TX port from the RX port, then replace the source and destination MAC addresses if MAC addresses updating is enabled.

Note

In the following code, one line for getting the output port requires some explanation.

During the initialization process, a static array of destination ports (l2fwd_dst_ports[]) is filled such that for each source port, a destination port is assigned that is either the next or previous enabled port from the portmask. Naturally, the number of ports in the portmask must be even, otherwise, the application exits.

static void
l2fwd_simple_forward(struct rte_mbuf *m, unsigned portid)
{
    struct ether_hdr *eth;
    void *tmp;
    unsigned dst_port;

    dst_port = l2fwd_dst_ports[portid];

    eth = rte_pktmbuf_mtod(m, struct ether_hdr *);

    /* 02:00:00:00:00:xx */

    tmp = &eth->d_addr.addr_bytes[0];

    *((uint64_t *)tmp) = 0x000000000002 + ((uint64_t) dst_port << 40);

    /* src addr */

    ether_addr_copy(&l2fwd_ports_eth_addr[dst_port], &eth->s_addr);

    l2fwd_send_packet(m, (uint8_t) dst_port);
}

Then, the packet is sent using the l2fwd_send_packet (m, dst_port) function. For this test application, the processing is exactly the same for all packets arriving on the same RX port. Therefore, it would have been possible to call the l2fwd_send_burst() function directly from the main loop to send all the received packets on the same TX port, using the burst-oriented send function, which is more efficient.

However, in real-life applications (such as, L3 routing), packet N is not necessarily forwarded on the same port as packet N-1. The application is implemented to illustrate that, so the same approach can be reused in a more complex application.

The l2fwd_send_packet() function stores the packet in a per-lcore and per-txport table. If the table is full, the whole packets table is transmitted using the l2fwd_send_burst() function:

/* Send the packet on an output interface */

static int
l2fwd_send_packet(struct rte_mbuf *m, uint16_t port)
{
    unsigned lcore_id, len;
    struct lcore_queue_conf *qconf;

    lcore_id = rte_lcore_id();
    qconf = &lcore_queue_conf[lcore_id];
    len = qconf->tx_mbufs[port].len;
    qconf->tx_mbufs[port].m_table[len] = m;
    len++;

    /* enough pkts to be sent */

    if (unlikely(len == MAX_PKT_BURST)) {
        l2fwd_send_burst(qconf, MAX_PKT_BURST, port);
        len = 0;
    }

    qconf->tx_mbufs[port].len = len; return 0;
}

To ensure that no packets remain in the tables, each lcore does a draining of TX queue in its main loop. This technique introduces some latency when there are not many packets to send, however it improves performance:

cur_tsc = rte_rdtsc();

/*
 *   TX burst queue drain
 */

diff_tsc = cur_tsc - prev_tsc;

if (unlikely(diff_tsc > drain_tsc)) {
    for (portid = 0; portid < RTE_MAX_ETHPORTS; portid++) {
        if (qconf->tx_mbufs[portid].len == 0)
            continue;

        l2fwd_send_burst(&lcore_queue_conf[lcore_id], qconf->tx_mbufs[portid].len, (uint8_t) portid);

        qconf->tx_mbufs[portid].len = 0;
    }

    /* if timer is enabled */

    if (timer_period > 0) {
        /* advance the timer */

        timer_tsc += diff_tsc;

        /* if timer has reached its timeout */

        if (unlikely(timer_tsc >= (uint64_t) timer_period)) {
            /* do this only on master core */

            if (lcore_id == rte_get_master_lcore()) {
                print_stats();

                /* reset the timer */
                timer_tsc = 0;
            }
        }
    }

    prev_tsc = cur_tsc;
}

L2 Forwarding Sample Application with Cache Allocation Technology (CAT)

Basic Forwarding sample application is a simple skeleton example of a forwarding application. It has been extended to make use of CAT via extended command line options and linking against the libpqos library.

It is intended as a demonstration of the basic components of a DPDK forwarding application and use of the libpqos library to program CAT. For more detailed implementations see the L2 and L3 forwarding sample applications.

CAT and Code Data Prioritization (CDP) features allow management of the CPU’s last level cache. CAT introduces classes of service (COS) that are essentially bitmasks. In current CAT implementations, a bit in a COS bitmask corresponds to one cache way in last level cache. A CPU core is always assigned to one of the CAT classes. By programming CPU core assignment and COS bitmasks, applications can be given exclusive, shared, or mixed access to the CPU’s last level cache. CDP extends CAT so that there are two bitmasks per COS, one for data and one for code. The number of classes and number of valid bits in a COS bitmask is CPU model specific and COS bitmasks need to be contiguous. Sample code calls this bitmask cbm or capacity bitmask. By default, after reset, all CPU cores are assigned to COS 0 and all classes are programmed to allow fill into all cache ways. CDP is off by default.

For more information about CAT please see:

White paper demonstrating example use case:

Compiling the Application

Note

Requires libpqos from Intel’s intel-cmt-cat software package hosted on GitHub repository. For installation notes, please see README file.

GIT:

  1. To compile the application export the path to PQoS lib and the DPDK source tree and go to the example directory:

    export PQOS_INSTALL_PATH=/path/to/libpqos
    

To compile the sample application see Compiling the Sample Applications.

The application is located in the l2fwd-cat sub-directory.

Running the Application

To run the example in a linuxapp environment and enable CAT on cpus 0-2:

./build/l2fwd-cat -l 1 -n 4 -- --l3ca="0x3@(0-2)"

or to enable CAT and CDP on cpus 1,3:

./build/l2fwd-cat -l 1 -n 4 -- --l3ca="(0x00C00,0x00300)@(1,3)"

If CDP is not supported it will fail with following error message:

PQOS: CDP requested but not supported.
PQOS: Requested CAT configuration is not valid!
PQOS: Shutting down PQoS library...
EAL: Error - exiting with code: 1
  Cause: PQOS: L3CA init failed!

The option to enable CAT is:

  • --l3ca='<common_cbm@cpus>[,<(code_cbm,data_cbm)@cpus>...]':

    where cbm stands for capacity bitmask and must be expressed in hexadecimal form.

    common_cbm is a single mask, for a CDP enabled system, a group of two masks (code_cbm and data_cbm) is used.

    ( and ) are necessary if it’s a group.

    cpus could be a single digit/range or a group and must be expressed in decimal form.

    ( and ) are necessary if it’s a group.

    e.g. --l3ca='0x00F00@(1,3),0x0FF00@(4-6),0xF0000@7'

    • cpus 1 and 3 share its 4 ways with cpus 4, 5 and 6;
    • cpus 4, 5 and 6 share half (4 out of 8 ways) of its L3 with cpus 1 and 3;
    • cpus 4, 5 and 6 have exclusive access to 4 out of 8 ways;
    • cpu 7 has exclusive access to all of its 4 ways;

    e.g. --l3ca='(0x00C00,0x00300)@(1,3)' for CDP enabled system

    • cpus 1 and 3 have access to 2 ways for code and 2 ways for data, code and data ways are not overlapping.

Refer to DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

To reset or list CAT configuration and control CDP please use pqos tool from Intel’s intel-cmt-cat software package.

To enabled or disable CDP:

sudo ./pqos -S cdp-on

sudo ./pqos -S cdp-off

to reset CAT configuration:

sudo ./pqos -R

to list CAT config:

sudo ./pqos -s

For more info about pqos tool please see its man page or intel-cmt-cat wiki.

Explanation

The following sections provide an explanation of the main components of the code.

All DPDK library functions used in the sample code are prefixed with rte_ and are explained in detail in the DPDK API Documentation.

The Main Function

The main() function performs the initialization and calls the execution threads for each lcore.

The first task is to initialize the Environment Abstraction Layer (EAL). The argc and argv arguments are provided to the rte_eal_init() function. The value returned is the number of parsed arguments:

int ret = rte_eal_init(argc, argv);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Error with EAL initialization\n");

The next task is to initialize the PQoS library and configure CAT. The argc and argv arguments are provided to the cat_init() function. The value returned is the number of parsed arguments:

int ret = cat_init(argc, argv);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "PQOS: L3CA init failed!\n");

cat_init() is a wrapper function which parses the command, validates the requested parameters and configures CAT accordingly.

Parsing of command line arguments is done in parse_args(...). libpqos is then initialized with the pqos_init(...) call. Next, libpqos is queried for system CPU information and L3CA capabilities via pqos_cap_get(...) and pqos_cap_get_type(..., PQOS_CAP_TYPE_L3CA, ...) calls. When all capability and topology information is collected, the requested CAT configuration is validated. A check is then performed (on per socket basis) for a sufficient number of un-associated COS. COS are selected and configured via the pqos_l3ca_set(...) call. Finally, COS are associated to relevant CPUs via pqos_l3ca_assoc_set(...) calls.

atexit(...) is used to register cat_exit(...) to be called on a clean exit. cat_exit(...) performs a simple CAT clean-up, by associating COS 0 to all involved CPUs via pqos_l3ca_assoc_set(...) calls.

L3 Forwarding Sample Application

The L3 Forwarding application is a simple example of packet processing using the DPDK. The application performs L3 forwarding.

Overview

The application demonstrates the use of the hash and LPM libraries in the DPDK to implement packet forwarding. The initialization and run-time paths are very similar to those of the L2 Forwarding Sample Application (in Real and Virtualized Environments). The main difference from the L2 Forwarding sample application is that the forwarding decision is made based on information read from the input packet.

The lookup method is either hash-based or LPM-based and is selected at run time. When the selected lookup method is hash-based, a hash object is used to emulate the flow classification stage. The hash object is used in correlation with a flow table to map each input packet to its flow at runtime.

The hash lookup key is represented by a DiffServ 5-tuple composed of the following fields read from the input packet: Source IP Address, Destination IP Address, Protocol, Source Port and Destination Port. The ID of the output interface for the input packet is read from the identified flow table entry. The set of flows used by the application is statically configured and loaded into the hash at initialization time. When the selected lookup method is LPM based, an LPM object is used to emulate the forwarding stage for IPv4 packets. The LPM object is used as the routing table to identify the next hop for each input packet at runtime.

The LPM lookup key is represented by the Destination IP Address field read from the input packet. The ID of the output interface for the input packet is the next hop returned by the LPM lookup. The set of LPM rules used by the application is statically configured and loaded into the LPM object at initialization time.

In the sample application, hash-based forwarding supports IPv4 and IPv6. LPM-based forwarding supports IPv4 only.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the l3fwd sub-directory.

Running the Application

The application has a number of command line options:

./l3fwd [EAL options] -- -p PORTMASK
                         [-P]
                         [-E]
                         [-L]
                         --config(port,queue,lcore)[,(port,queue,lcore)]
                         [--eth-dest=X,MM:MM:MM:MM:MM:MM]
                         [--enable-jumbo [--max-pkt-len PKTLEN]]
                         [--no-numa]
                         [--hash-entry-num]
                         [--ipv6]
                         [--parse-ptype]

Where,

  • -p PORTMASK: Hexadecimal bitmask of ports to configure
  • -P: Optional, sets all ports to promiscuous mode so that packets are accepted regardless of the packet’s Ethernet MAC destination address. Without this option, only packets with the Ethernet MAC destination address set to the Ethernet address of the port are accepted.
  • -E: Optional, enable exact match.
  • -L: Optional, enable longest prefix match.
  • --config (port,queue,lcore)[,(port,queue,lcore)]: Determines which queues from which ports are mapped to which cores.
  • --eth-dest=X,MM:MM:MM:MM:MM:MM: Optional, ethernet destination for port X.
  • --enable-jumbo: Optional, enables jumbo frames.
  • --max-pkt-len: Optional, under the premise of enabling jumbo, maximum packet length in decimal (64-9600).
  • --no-numa: Optional, disables numa awareness.
  • --hash-entry-num: Optional, specifies the hash entry number in hexadecimal to be setup.
  • --ipv6: Optional, set if running ipv6 packets.
  • --parse-ptype: Optional, set to use software to analyze packet type. Without this option, hardware will check the packet type.

For example, consider a dual processor socket platform with 8 physical cores, where cores 0-7 and 16-23 appear on socket 0, while cores 8-15 and 24-31 appear on socket 1.

To enable L3 forwarding between two ports, assuming that both ports are in the same socket, using two cores, cores 1 and 2, (which are in the same socket too), use the following command:

./build/l3fwd -l 1,2 -n 4 -- -p 0x3 --config="(0,0,1),(1,0,2)"

In this command:

  • The -l option enables cores 1, 2
  • The -p option enables ports 0 and 1
  • The –config option enables one queue on each port and maps each (port,queue) pair to a specific core. The following table shows the mapping in this example:
Port Queue lcore Description
0 0 1 Map queue 0 from port 0 to lcore 1.
1 0 2 Map queue 0 from port 1 to lcore 2.

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The following sections provide some explanation of the sample application code. As mentioned in the overview section, the initialization and run-time paths are very similar to those of the L2 Forwarding Sample Application (in Real and Virtualized Environments). The following sections describe aspects that are specific to the L3 Forwarding sample application.

Hash Initialization

The hash object is created and loaded with the pre-configured entries read from a global array, and then generate the expected 5-tuple as key to keep consistence with those of real flow for the convenience to execute hash performance test on 4M/8M/16M flows.

Note

The Hash initialization will setup both ipv4 and ipv6 hash table, and populate the either table depending on the value of variable ipv6. To support the hash performance test with up to 8M single direction flows/16M bi-direction flows, populate_ipv4_many_flow_into_table() function will populate the hash table with specified hash table entry number(default 4M).

Note

Value of global variable ipv6 can be specified with –ipv6 in the command line. Value of global variable hash_entry_number, which is used to specify the total hash entry number for all used ports in hash performance test, can be specified with –hash-entry-num VALUE in command line, being its default value 4.

#if (APP_LOOKUP_METHOD == APP_LOOKUP_EXACT_MATCH)

    static void
    setup_hash(int socketid)
    {
        // ...

        if (hash_entry_number != HASH_ENTRY_NUMBER_DEFAULT) {
            if (ipv6 == 0) {
                /* populate the ipv4 hash */
                populate_ipv4_many_flow_into_table(ipv4_l3fwd_lookup_struct[socketid], hash_entry_number);
            } else {
                /* populate the ipv6 hash */
                populate_ipv6_many_flow_into_table( ipv6_l3fwd_lookup_struct[socketid], hash_entry_number);
            }
        } else
            if (ipv6 == 0) {
                /* populate the ipv4 hash */
                populate_ipv4_few_flow_into_table(ipv4_l3fwd_lookup_struct[socketid]);
            } else {
                /* populate the ipv6 hash */
                populate_ipv6_few_flow_into_table(ipv6_l3fwd_lookup_struct[socketid]);
            }
        }
    }
#endif
LPM Initialization

The LPM object is created and loaded with the pre-configured entries read from a global array.

#if (APP_LOOKUP_METHOD == APP_LOOKUP_LPM)

static void
setup_lpm(int socketid)
{
    unsigned i;
    int ret;
    char s[64];

    /* create the LPM table */

    snprintf(s, sizeof(s), "IPV4_L3FWD_LPM_%d", socketid);

    ipv4_l3fwd_lookup_struct[socketid] = rte_lpm_create(s, socketid, IPV4_L3FWD_LPM_MAX_RULES, 0);

    if (ipv4_l3fwd_lookup_struct[socketid] == NULL)
        rte_exit(EXIT_FAILURE, "Unable to create the l3fwd LPM table"
            " on socket %d\n", socketid);

    /* populate the LPM table */

    for (i = 0; i < IPV4_L3FWD_NUM_ROUTES; i++) {
        /* skip unused ports */

        if ((1 << ipv4_l3fwd_route_array[i].if_out & enabled_port_mask) == 0)
            continue;

        ret = rte_lpm_add(ipv4_l3fwd_lookup_struct[socketid], ipv4_l3fwd_route_array[i].ip,
                                ipv4_l3fwd_route_array[i].depth, ipv4_l3fwd_route_array[i].if_out);

        if (ret < 0) {
            rte_exit(EXIT_FAILURE, "Unable to add entry %u to the "
                    "l3fwd LPM table on socket %d\n", i, socketid);
        }

        printf("LPM: Adding route 0x%08x / %d (%d)\n",
            (unsigned)ipv4_l3fwd_route_array[i].ip, ipv4_l3fwd_route_array[i].depth, ipv4_l3fwd_route_array[i].if_out);
    }
}
#endif
Packet Forwarding for Hash-based Lookups

For each input packet, the packet forwarding operation is done by the l3fwd_simple_forward() or simple_ipv4_fwd_4pkts() function for IPv4 packets or the simple_ipv6_fwd_4pkts() function for IPv6 packets. The l3fwd_simple_forward() function provides the basic functionality for both IPv4 and IPv6 packet forwarding for any number of burst packets received, and the packet forwarding decision (that is, the identification of the output interface for the packet) for hash-based lookups is done by the get_ipv4_dst_port() or get_ipv6_dst_port() function. The get_ipv4_dst_port() function is shown below:

static inline uint8_t
get_ipv4_dst_port(void *ipv4_hdr, uint16_t portid, lookup_struct_t *ipv4_l3fwd_lookup_struct)
{
    int ret = 0;
    union ipv4_5tuple_host key;

    ipv4_hdr = (uint8_t *)ipv4_hdr + offsetof(struct ipv4_hdr, time_to_live);

    m128i data = _mm_loadu_si128(( m128i*)(ipv4_hdr));

    /* Get 5 tuple: dst port, src port, dst IP address, src IP address and protocol */

    key.xmm = _mm_and_si128(data, mask0);

    /* Find destination port */

    ret = rte_hash_lookup(ipv4_l3fwd_lookup_struct, (const void *)&key);

    return (uint8_t)((ret < 0)? portid : ipv4_l3fwd_out_if[ret]);
}

The get_ipv6_dst_port() function is similar to the get_ipv4_dst_port() function.

The simple_ipv4_fwd_4pkts() and simple_ipv6_fwd_4pkts() function are optimized for continuous 4 valid ipv4 and ipv6 packets, they leverage the multiple buffer optimization to boost the performance of forwarding packets with the exact match on hash table. The key code snippet of simple_ipv4_fwd_4pkts() is shown below:

static inline void
simple_ipv4_fwd_4pkts(struct rte_mbuf* m[4], uint16_t portid, struct lcore_conf *qconf)
{
    // ...

    data[0] = _mm_loadu_si128(( m128i*)(rte_pktmbuf_mtod(m[0], unsigned char *) + sizeof(struct ether_hdr) + offsetof(struct ipv4_hdr, time_to_live)));
    data[1] = _mm_loadu_si128(( m128i*)(rte_pktmbuf_mtod(m[1], unsigned char *) + sizeof(struct ether_hdr) + offsetof(struct ipv4_hdr, time_to_live)));
    data[2] = _mm_loadu_si128(( m128i*)(rte_pktmbuf_mtod(m[2], unsigned char *) + sizeof(struct ether_hdr) + offsetof(struct ipv4_hdr, time_to_live)));
    data[3] = _mm_loadu_si128(( m128i*)(rte_pktmbuf_mtod(m[3], unsigned char *) + sizeof(struct ether_hdr) + offsetof(struct ipv4_hdr, time_to_live)));

    key[0].xmm = _mm_and_si128(data[0], mask0);
    key[1].xmm = _mm_and_si128(data[1], mask0);
    key[2].xmm = _mm_and_si128(data[2], mask0);
    key[3].xmm = _mm_and_si128(data[3], mask0);

    const void *key_array[4] = {&key[0], &key[1], &key[2],&key[3]};

    rte_hash_lookup_bulk(qconf->ipv4_lookup_struct, &key_array[0], 4, ret);

    dst_port[0] = (ret[0] < 0)? portid:ipv4_l3fwd_out_if[ret[0]];
    dst_port[1] = (ret[1] < 0)? portid:ipv4_l3fwd_out_if[ret[1]];
    dst_port[2] = (ret[2] < 0)? portid:ipv4_l3fwd_out_if[ret[2]];
    dst_port[3] = (ret[3] < 0)? portid:ipv4_l3fwd_out_if[ret[3]];

    // ...
}

The simple_ipv6_fwd_4pkts() function is similar to the simple_ipv4_fwd_4pkts() function.

Known issue: IP packets with extensions or IP packets which are not TCP/UDP cannot work well at this mode.

Packet Forwarding for LPM-based Lookups

For each input packet, the packet forwarding operation is done by the l3fwd_simple_forward() function, but the packet forwarding decision (that is, the identification of the output interface for the packet) for LPM-based lookups is done by the get_ipv4_dst_port() function below:

static inline uint16_t
get_ipv4_dst_port(struct ipv4_hdr *ipv4_hdr, uint16_t portid, lookup_struct_t *ipv4_l3fwd_lookup_struct)
{
    uint8_t next_hop;

    return ((rte_lpm_lookup(ipv4_l3fwd_lookup_struct, rte_be_to_cpu_32(ipv4_hdr->dst_addr), &next_hop) == 0)? next_hop : portid);
}

L3 Forwarding with Power Management Sample Application

Introduction

The L3 Forwarding with Power Management application is an example of power-aware packet processing using the DPDK. The application is based on existing L3 Forwarding sample application, with the power management algorithms to control the P-states and C-states of the Intel processor via a power management library.

Overview

The application demonstrates the use of the Power libraries in the DPDK to implement packet forwarding. The initialization and run-time paths are very similar to those of the L3 Forwarding Sample Application. The main difference from the L3 Forwarding sample application is that this application introduces power-aware optimization algorithms by leveraging the Power library to control P-state and C-state of processor based on packet load.

The DPDK includes poll-mode drivers to configure Intel NIC devices and their receive (Rx) and transmit (Tx) queues. The design principle of this PMD is to access the Rx and Tx descriptors directly without any interrupts to quickly receive, process and deliver packets in the user space.

In general, the DPDK executes an endless packet processing loop on dedicated IA cores that include the following steps:

  • Retrieve input packets through the PMD to poll Rx queue
  • Process each received packet or provide received packets to other processing cores through software queues
  • Send pending output packets to Tx queue through the PMD

In this way, the PMD achieves better performance than a traditional interrupt-mode driver, at the cost of keeping cores active and running at the highest frequency, hence consuming the maximum power all the time. However, during the period of processing light network traffic, which happens regularly in communication infrastructure systems due to well-known “tidal effect”, the PMD is still busy waiting for network packets, which wastes a lot of power.

Processor performance states (P-states) are the capability of an Intel processor to switch between different supported operating frequencies and voltages. If configured correctly, according to system workload, this feature provides power savings. CPUFreq is the infrastructure provided by the Linux* kernel to control the processor performance state capability. CPUFreq supports a user space governor that enables setting frequency via manipulating the virtual file device from a user space application. The Power library in the DPDK provides a set of APIs for manipulating a virtual file device to allow user space application to set the CPUFreq governor and set the frequency of specific cores.

This application includes a P-state power management algorithm to generate a frequency hint to be sent to CPUFreq. The algorithm uses the number of received and available Rx packets on recent polls to make a heuristic decision to scale frequency up/down. Specifically, some thresholds are checked to see whether a specific core running an DPDK polling thread needs to increase frequency a step up based on the near to full trend of polled Rx queues. Also, it decreases frequency a step if packet processed per loop is far less than the expected threshold or the thread’s sleeping time exceeds a threshold.

C-States are also known as sleep states. They allow software to put an Intel core into a low power idle state from which it is possible to exit via an event, such as an interrupt. However, there is a tradeoff between the power consumed in the idle state and the time required to wake up from the idle state (exit latency). Therefore, as you go into deeper C-states, the power consumed is lower but the exit latency is increased. Each C-state has a target residency. It is essential that when entering into a C-state, the core remains in this C-state for at least as long as the target residency in order to fully realize the benefits of entering the C-state. CPUIdle is the infrastructure provide by the Linux kernel to control the processor C-state capability. Unlike CPUFreq, CPUIdle does not provide a mechanism that allows the application to change C-state. It actually has its own heuristic algorithms in kernel space to select target C-state to enter by executing privileged instructions like HLT and MWAIT, based on the speculative sleep duration of the core. In this application, we introduce a heuristic algorithm that allows packet processing cores to sleep for a short period if there is no Rx packet received on recent polls. In this way, CPUIdle automatically forces the corresponding cores to enter deeper C-states instead of always running to the C0 state waiting for packets.

Note

To fully demonstrate the power saving capability of using C-states, it is recommended to enable deeper C3 and C6 states in the BIOS during system boot up.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the l3fwd-power sub-directory.

Running the Application

The application has a number of command line options:

./build/l3fwd_power [EAL options] -- -p PORTMASK [-P]  --config(port,queue,lcore)[,(port,queue,lcore)] [--enable-jumbo [--max-pkt-len PKTLEN]] [--no-numa]

where,

  • -p PORTMASK: Hexadecimal bitmask of ports to configure
  • -P: Sets all ports to promiscuous mode so that packets are accepted regardless of the packet’s Ethernet MAC destination address. Without this option, only packets with the Ethernet MAC destination address set to the Ethernet address of the port are accepted.
  • –config (port,queue,lcore)[,(port,queue,lcore)]: determines which queues from which ports are mapped to which cores.
  • –enable-jumbo: optional, enables jumbo frames
  • –max-pkt-len: optional, maximum packet length in decimal (64-9600)
  • –no-numa: optional, disables numa awareness

See L3 Forwarding Sample Application for details. The L3fwd-power example reuses the L3fwd command line options.

Explanation

The following sections provide some explanation of the sample application code. As mentioned in the overview section, the initialization and run-time paths are identical to those of the L3 forwarding application. The following sections describe aspects that are specific to the L3 Forwarding with Power Management sample application.

Power Library Initialization

The Power library is initialized in the main routine. It changes the P-state governor to userspace for specific cores that are under control. The Timer library is also initialized and several timers are created later on, responsible for checking if it needs to scale down frequency at run time by checking CPU utilization statistics.

Note

Only the power management related initialization is shown.

int main(int argc, char **argv)
{
    struct lcore_conf *qconf;
    int ret;
    unsigned nb_ports;
    uint16_t queueid, portid;
    unsigned lcore_id;
    uint64_t hz;
    uint32_t n_tx_queue, nb_lcores;
    uint8_t nb_rx_queue, queue, socketid;

    // ...

    /* init RTE timer library to be used to initialize per-core timers */

    rte_timer_subsystem_init();

    // ...


    /* per-core initialization */

    for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++) {
        if (rte_lcore_is_enabled(lcore_id) == 0)
            continue;

        /* init power management library for a specified core */

        ret = rte_power_init(lcore_id);
        if (ret)
            rte_exit(EXIT_FAILURE, "Power management library "
                "initialization failed on core%d\n", lcore_id);

        /* init timer structures for each enabled lcore */

        rte_timer_init(&power_timers[lcore_id]);

        hz = rte_get_hpet_hz();

        rte_timer_reset(&power_timers[lcore_id], hz/TIMER_NUMBER_PER_SECOND, SINGLE, lcore_id, power_timer_cb, NULL);

        // ...
    }

    // ...
}
Monitoring Loads of Rx Queues

In general, the polling nature of the DPDK prevents the OS power management subsystem from knowing if the network load is actually heavy or light. In this sample, sampling network load work is done by monitoring received and available descriptors on NIC Rx queues in recent polls. Based on the number of returned and available Rx descriptors, this example implements algorithms to generate frequency scaling hints and speculative sleep duration, and use them to control P-state and C-state of processors via the power management library. Frequency (P-state) control and sleep state (C-state) control work individually for each logical core, and the combination of them contributes to a power efficient packet processing solution when serving light network loads.

The rte_eth_rx_burst() function and the newly-added rte_eth_rx_queue_count() function are used in the endless packet processing loop to return the number of received and available Rx descriptors. And those numbers of specific queue are passed to P-state and C-state heuristic algorithms to generate hints based on recent network load trends.

Note

Only power control related code is shown.

static
attribute ((noreturn)) int main_loop( attribute ((unused)) void *dummy)
{
    // ...

    while (1) {
    // ...

    /**
     * Read packet from RX queues
     */

    lcore_scaleup_hint = FREQ_CURRENT;
    lcore_rx_idle_count = 0;

    for (i = 0; i < qconf->n_rx_queue; ++i)
    {
        rx_queue = &(qconf->rx_queue_list[i]);
        rx_queue->idle_hint = 0;
        portid = rx_queue->port_id;
        queueid = rx_queue->queue_id;

        nb_rx = rte_eth_rx_burst(portid, queueid, pkts_burst, MAX_PKT_BURST);
        stats[lcore_id].nb_rx_processed += nb_rx;

        if (unlikely(nb_rx == 0)) {
            /**
             * no packet received from rx queue, try to
             * sleep for a while forcing CPU enter deeper
             * C states.
             */

            rx_queue->zero_rx_packet_count++;

            if (rx_queue->zero_rx_packet_count <= MIN_ZERO_POLL_COUNT)
                continue;

            rx_queue->idle_hint = power_idle_heuristic(rx_queue->zero_rx_packet_count);
            lcore_rx_idle_count++;
        } else {
            rx_ring_length = rte_eth_rx_queue_count(portid, queueid);

            rx_queue->zero_rx_packet_count = 0;

            /**
             * do not scale up frequency immediately as
             * user to kernel space communication is costly
             * which might impact packet I/O for received
             * packets.
             */

            rx_queue->freq_up_hint = power_freq_scaleup_heuristic(lcore_id, rx_ring_length);
        }

        /* Prefetch and forward packets */

        // ...
    }

    if (likely(lcore_rx_idle_count != qconf->n_rx_queue)) {
        for (i = 1, lcore_scaleup_hint = qconf->rx_queue_list[0].freq_up_hint; i < qconf->n_rx_queue; ++i) {
            x_queue = &(qconf->rx_queue_list[i]);

            if (rx_queue->freq_up_hint > lcore_scaleup_hint)

                lcore_scaleup_hint = rx_queue->freq_up_hint;
        }

        if (lcore_scaleup_hint == FREQ_HIGHEST)

            rte_power_freq_max(lcore_id);

        else if (lcore_scaleup_hint == FREQ_HIGHER)
            rte_power_freq_up(lcore_id);
        } else {
            /**
             *  All Rx queues empty in recent consecutive polls,
             *  sleep in a conservative manner, meaning sleep as
             * less as possible.
             */

            for (i = 1, lcore_idle_hint = qconf->rx_queue_list[0].idle_hint; i < qconf->n_rx_queue; ++i) {
                rx_queue = &(qconf->rx_queue_list[i]);
                if (rx_queue->idle_hint < lcore_idle_hint)
                    lcore_idle_hint = rx_queue->idle_hint;
            }

            if ( lcore_idle_hint < SLEEP_GEAR1_THRESHOLD)
                /**
                 *   execute "pause" instruction to avoid context
                 *   switch for short sleep.
                 */
                rte_delay_us(lcore_idle_hint);
            else
                /* long sleep force ruining thread to suspend */
                usleep(lcore_idle_hint);

           stats[lcore_id].sleep_time += lcore_idle_hint;
        }
    }
}
P-State Heuristic Algorithm

The power_freq_scaleup_heuristic() function is responsible for generating a frequency hint for the specified logical core according to available descriptor number returned from rte_eth_rx_queue_count(). On every poll for new packets, the length of available descriptor on an Rx queue is evaluated, and the algorithm used for frequency hinting is as follows:

  • If the size of available descriptors exceeds 96, the maximum frequency is hinted.
  • If the size of available descriptors exceeds 64, a trend counter is incremented by 100.
  • If the length of the ring exceeds 32, the trend counter is incremented by 1.
  • When the trend counter reached 10000 the frequency hint is changed to the next higher frequency.

Note

The assumption is that the Rx queue size is 128 and the thresholds specified above must be adjusted accordingly based on actual hardware Rx queue size, which are configured via the rte_eth_rx_queue_setup() function.

In general, a thread needs to poll packets from multiple Rx queues. Most likely, different queue have different load, so they would return different frequency hints. The algorithm evaluates all the hints and then scales up frequency in an aggressive manner by scaling up to highest frequency as long as one Rx queue requires. In this way, we can minimize any negative performance impact.

On the other hand, frequency scaling down is controlled in the timer callback function. Specifically, if the sleep times of a logical core indicate that it is sleeping more than 25% of the sampling period, or if the average packet per iteration is less than expectation, the frequency is decreased by one step.

C-State Heuristic Algorithm

Whenever recent rte_eth_rx_burst() polls return 5 consecutive zero packets, an idle counter begins incrementing for each successive zero poll. At the same time, the function power_idle_heuristic() is called to generate speculative sleep duration in order to force logical to enter deeper sleeping C-state. There is no way to control C- state directly, and the CPUIdle subsystem in OS is intelligent enough to select C-state to enter based on actual sleep period time of giving logical core. The algorithm has the following sleeping behavior depending on the idle counter:

  • If idle count less than 100, the counter value is used as a microsecond sleep value through rte_delay_us() which execute pause instructions to avoid costly context switch but saving power at the same time.
  • If idle count is between 100 and 999, a fixed sleep interval of 100 μs is used. A 100 μs sleep interval allows the core to enter the C1 state while keeping a fast response time in case new traffic arrives.
  • If idle count is greater than 1000, a fixed sleep value of 1 ms is used until the next timer expiration is used. This allows the core to enter the C3/C6 states.

Note

The thresholds specified above need to be adjusted for different Intel processors and traffic profiles.

If a thread polls multiple Rx queues and different queue returns different sleep duration values, the algorithm controls the sleep time in a conservative manner by sleeping for the least possible time in order to avoid a potential performance impact.

L3 Forwarding with Access Control Sample Application

The L3 Forwarding with Access Control application is a simple example of packet processing using the DPDK. The application performs a security check on received packets. Packets that are in the Access Control List (ACL), which is loaded during initialization, are dropped. Others are forwarded to the correct port.

Overview

The application demonstrates the use of the ACL library in the DPDK to implement access control and packet L3 forwarding. The application loads two types of rules at initialization:

  • Route information rules, which are used for L3 forwarding
  • Access Control List (ACL) rules that blacklist (or block) packets with a specific characteristic

When packets are received from a port, the application extracts the necessary information from the TCP/IP header of the received packet and performs a lookup in the rule database to figure out whether the packets should be dropped (in the ACL range) or forwarded to desired ports. The initialization and run-time paths are similar to those of the L3 Forwarding Sample Application. However, there are significant differences in the two applications. For example, the original L3 forwarding application uses either LPM or an exact match algorithm to perform forwarding port lookup, while this application uses the ACL library to perform both ACL and route entry lookup. The following sections provide more detail.

Classification for both IPv4 and IPv6 packets is supported in this application. The application also assumes that all the packets it processes are TCP/UDP packets and always extracts source/destination port information from the packets.

Tuple Packet Syntax

The application implements packet classification for the IPv4/IPv6 5-tuple syntax specifically. The 5-tuple syntax consist of a source IP address, a destination IP address, a source port, a destination port and a protocol identifier. The fields in the 5-tuple syntax have the following formats:

  • Source IP address and destination IP address : Each is either a 32-bit field (for IPv4), or a set of 4 32-bit fields (for IPv6) represented by a value and a mask length. For example, an IPv4 range of 192.168.1.0 to 192.168.1.255 could be represented by a value = [192, 168, 1, 0] and a mask length = 24.
  • Source port and destination port : Each is a 16-bit field, represented by a lower start and a higher end. For example, a range of ports 0 to 8192 could be represented by lower = 0 and higher = 8192.
  • Protocol identifier : An 8-bit field, represented by a value and a mask, that covers a range of values. To verify that a value is in the range, use the following expression: “(VAL & mask) == value”

The trick in how to represent a range with a mask and value is as follows. A range can be enumerated in binary numbers with some bits that are never changed and some bits that are dynamically changed. Set those bits that dynamically changed in mask and value with 0. Set those bits that never changed in the mask with 1, in value with number expected. For example, a range of 6 to 7 is enumerated as 0b110 and 0b111. Bit 1-7 are bits never changed and bit 0 is the bit dynamically changed. Therefore, set bit 0 in mask and value with 0, set bits 1-7 in mask with 1, and bits 1-7 in value with number 0b11. So, mask is 0xfe, value is 0x6.

Note

The library assumes that each field in the rule is in LSB or Little Endian order when creating the database. It internally converts them to MSB or Big Endian order. When performing a lookup, the library assumes the input is in MSB or Big Endian order.

Access Rule Syntax

In this sample application, each rule is a combination of the following:

  • 5-tuple field: This field has a format described in Section.
  • priority field: A weight to measure the priority of the rules. The rule with the higher priority will ALWAYS be returned if the specific input has multiple matches in the rule database. Rules with lower priority will NEVER be returned in any cases.
  • userdata field: A user-defined field that could be any value. It can be the forwarding port number if the rule is a route table entry or it can be a pointer to a mapping address if the rule is used for address mapping in the NAT application. The key point is that it is a useful reserved field for user convenience.
ACL and Route Rules

The application needs to acquire ACL and route rules before it runs. Route rules are mandatory, while ACL rules are optional. To simplify the complexity of the priority field for each rule, all ACL and route entries are assumed to be in the same file. To read data from the specified file successfully, the application assumes the following:

  • Each rule occupies a single line.
  • Only the following four rule line types are valid in this application:
  • ACL rule line, which starts with a leading character ‘@’
  • Route rule line, which starts with a leading character ‘R’
  • Comment line, which starts with a leading character ‘#’
  • Empty line, which consists of a space, form-feed (‘f’), newline (‘n’), carriage return (‘r’), horizontal tab (‘t’), or vertical tab (‘v’).

Other lines types are considered invalid.

  • Rules are organized in descending order of priority, which means rules at the head of the file always have a higher priority than those further down in the file.
  • A typical IPv4 ACL rule line should have a format as shown below:
_images/ipv4_acl_rule.png

A typical IPv4 ACL rule

IPv4 addresses are specified in CIDR format as specified in RFC 4632. They consist of the dot notation for the address and a prefix length separated by ‘/’. For example, 192.168.0.34/32, where the address is 192.168.0.34 and the prefix length is 32.

Ports are specified as a range of 16-bit numbers in the format MIN:MAX, where MIN and MAX are the inclusive minimum and maximum values of the range. The range 0:65535 represents all possible ports in a range. When MIN and MAX are the same value, a single port is represented, for example, 20:20.

The protocol identifier is an 8-bit value and a mask separated by ‘/’. For example: 6/0xfe matches protocol values 6 and 7.

  • Route rules start with a leading character ‘R’ and have the same format as ACL rules except an extra field at the tail that indicates the forwarding port number.
Rules File Example
_images/example_rules.png

Rules example

Each rule is explained as follows:

  • Rule 1 (the first line) tells the application to drop those packets with source IP address = [1.2.3.*], destination IP address = [192.168.0.36], protocol = [6]/[7]
  • Rule 2 (the second line) is similar to Rule 1, except the source IP address is ignored. It tells the application to forward packets with destination IP address = [192.168.0.36], protocol = [6]/[7], destined to port 1.
  • Rule 3 (the third line) tells the application to forward all packets to port 0. This is something like a default route entry.

As described earlier, the application assume rules are listed in descending order of priority, therefore Rule 1 has the highest priority, then Rule 2, and finally, Rule 3 has the lowest priority.

Consider the arrival of the following three packets:

  • Packet 1 has source IP address = [1.2.3.4], destination IP address = [192.168.0.36], and protocol = [6]
  • Packet 2 has source IP address = [1.2.4.4], destination IP address = [192.168.0.36], and protocol = [6]
  • Packet 3 has source IP address = [1.2.3.4], destination IP address = [192.168.0.36], and protocol = [8]

Observe that:

  • Packet 1 matches all of the rules
  • Packet 2 matches Rule 2 and Rule 3
  • Packet 3 only matches Rule 3

For priority reasons, Packet 1 matches Rule 1 and is dropped. Packet 2 matches Rule 2 and is forwarded to port 1. Packet 3 matches Rule 3 and is forwarded to port 0.

For more details on the rule file format, please refer to rule_ipv4.db and rule_ipv6.db files (inside <RTE_SDK>/examples/l3fwd-acl/).

Application Phases

Once the application starts, it transitions through three phases:

  • Initialization Phase - Perform the following tasks:

  • Parse command parameters. Check the validity of rule file(s) name(s), number of logical cores, receive and transmit queues. Bind ports, queues and logical cores. Check ACL search options, and so on.

  • Call Environmental Abstraction Layer (EAL) and Poll Mode Driver (PMD) functions to initialize the environment and detect possible NICs. The EAL creates several threads and sets affinity to a specific hardware thread CPU based on the configuration specified by the command line arguments.

  • Read the rule files and format the rules into the representation that the ACL library can recognize. Call the ACL library function to add the rules into the database and compile them as a trie of pattern sets. Note that application maintains a separate AC contexts for IPv4 and IPv6 rules.

  • Runtime Phase - Process the incoming packets from a port. Packets are processed in three steps:

    • Retrieval: Gets a packet from the receive queue. Each logical core may process several queues for different ports. This depends on the configuration specified by command line arguments.
    • Lookup: Checks that the packet type is supported (IPv4/IPv6) and performs a 5-tuple lookup over corresponding AC context. If an ACL rule is matched, the packets will be dropped and return back to step 1. If a route rule is matched, it indicates the packet is not in the ACL list and should be forwarded. If there is no matches for the packet, then the packet is dropped.
    • Forwarding: Forwards the packet to the corresponding port.
  • Final Phase - Perform the following tasks:

    Calls the EAL, PMD driver and ACL library to free resource, then quits.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the l3fwd-acl sub-directory.

Running the Application

The application has a number of command line options:

./build/l3fwd-acl [EAL options] -- -p PORTMASK [-P] --config(port,queue,lcore)[,(port,queue,lcore)] --rule_ipv4 FILENAME rule_ipv6 FILENAME [--scalar] [--enable-jumbo [--max-pkt-len PKTLEN]] [--no-numa]

where,

  • -p PORTMASK: Hexadecimal bitmask of ports to configure
  • -P: Sets all ports to promiscuous mode so that packets are accepted regardless of the packet’s Ethernet MAC destination address. Without this option, only packets with the Ethernet MAC destination address set to the Ethernet address of the port are accepted.
  • –config (port,queue,lcore)[,(port,queue,lcore)]: determines which queues from which ports are mapped to which cores
  • –rule_ipv4 FILENAME: Specifies the IPv4 ACL and route rules file
  • –rule_ipv6 FILENAME: Specifies the IPv6 ACL and route rules file
  • –scalar: Use a scalar function to perform rule lookup
  • –enable-jumbo: optional, enables jumbo frames
  • –max-pkt-len: optional, maximum packet length in decimal (64-9600)
  • –no-numa: optional, disables numa awareness

For example, consider a dual processor socket platform with 8 physical cores, where cores 0-7 and 16-23 appear on socket 0, while cores 8-15 and 24-31 appear on socket 1.

To enable L3 forwarding between two ports, assuming that both ports are in the same socket, using two cores, cores 1 and 2, (which are in the same socket too), use the following command:

./build/l3fwd-acl -l 1,2 -n 4 -- -p 0x3 --config="(0,0,1),(1,0,2)" --rule_ipv4="./rule_ipv4.db" -- rule_ipv6="./rule_ipv6.db" --scalar

In this command:

  • The -l option enables cores 1, 2

  • The -p option enables ports 0 and 1

  • The –config option enables one queue on each port and maps each (port,queue) pair to a specific core. The following table shows the mapping in this example:

    Port Queue lcore Description
    0 0 1 Map queue 0 from port 0 to lcore 1.
    1 0 2 Map queue 0 from port 1 to lcore 2.
  • The –rule_ipv4 option specifies the reading of IPv4 rules sets from the ./ rule_ipv4.db file.

  • The –rule_ipv6 option specifies the reading of IPv6 rules sets from the ./ rule_ipv6.db file.

  • The –scalar option specifies the performing of rule lookup with a scalar function.

Explanation

The following sections provide some explanation of the sample application code. The aspects of port, device and CPU configuration are similar to those of the L3 Forwarding Sample Application. The following sections describe aspects that are specific to L3 forwarding with access control.

Parse Rules from File

As described earlier, both ACL and route rules are assumed to be saved in the same file. The application parses the rules from the file and adds them to the database by calling the ACL library function. It ignores empty and comment lines, and parses and validates the rules it reads. If errors are detected, the application exits with messages to identify the errors encountered.

The application needs to consider the userdata and priority fields. The ACL rules save the index to the specific rules in the userdata field, while route rules save the forwarding port number. In order to differentiate the two types of rules, ACL rules add a signature in the userdata field. As for the priority field, the application assumes rules are organized in descending order of priority. Therefore, the code only decreases the priority number with each rule it parses.

Setting Up the ACL Context

For each supported AC rule format (IPv4 5-tuple, IPv6 6-tuple) application creates a separate context handler from the ACL library for each CPU socket on the board and adds parsed rules into that context.

Note, that for each supported rule type, application needs to calculate the expected offset of the fields from the start of the packet. That’s why only packets with fixed IPv4/ IPv6 header are supported. That allows to perform ACL classify straight over incoming packet buffer - no extra protocol field retrieval need to be performed.

Subsequently, the application checks whether NUMA is enabled. If it is, the application records the socket IDs of the CPU cores involved in the task.

Finally, the application creates contexts handler from the ACL library, adds rules parsed from the file into the database and build an ACL trie. It is important to note that the application creates an independent copy of each database for each socket CPU involved in the task to reduce the time for remote memory access.

L3 Forwarding in a Virtualization Environment Sample Application

The L3 Forwarding in a Virtualization Environment sample application is a simple example of packet processing using the DPDK. The application performs L3 forwarding that takes advantage of Single Root I/O Virtualization (SR-IOV) features in a virtualized environment.

Overview

The application demonstrates the use of the hash and LPM libraries in the DPDK to implement packet forwarding. The initialization and run-time paths are very similar to those of the L3 Forwarding Sample Application. The forwarding decision is taken based on information read from the input packet.

The lookup method is either hash-based or LPM-based and is selected at compile time. When the selected lookup method is hash-based, a hash object is used to emulate the flow classification stage. The hash object is used in correlation with the flow table to map each input packet to its flow at runtime.

The hash lookup key is represented by the DiffServ 5-tuple composed of the following fields read from the input packet: Source IP Address, Destination IP Address, Protocol, Source Port and Destination Port. The ID of the output interface for the input packet is read from the identified flow table entry. The set of flows used by the application is statically configured and loaded into the hash at initialization time. When the selected lookup method is LPM based, an LPM object is used to emulate the forwarding stage for IPv4 packets. The LPM object is used as the routing table to identify the next hop for each input packet at runtime.

The LPM lookup key is represented by the Destination IP Address field read from the input packet. The ID of the output interface for the input packet is the next hop returned by the LPM lookup. The set of LPM rules used by the application is statically configured and loaded into the LPM object at the initialization time.

Note

Please refer to Virtual Function Setup Instructions for virtualized test case setup.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the l3fwd-vf sub-directory.

Running the Application

The application has a number of command line options:

./build/l3fwd-vf [EAL options] -- -p PORTMASK  --config(port,queue,lcore)[,(port,queue,lcore)] [--no-numa]

where,

  • –p PORTMASK: Hexadecimal bitmask of ports to configure
  • –config (port,queue,lcore)[,(port,queue,lcore]: determines which queues from which ports are mapped to which cores
  • –no-numa: optional, disables numa awareness

For example, consider a dual processor socket platform with 8 physical cores, where cores 0-7 and 16-23 appear on socket 0, while cores 8-15 and 24-31 appear on socket 1.

To enable L3 forwarding between two ports, assuming that both ports are in the same socket, using two cores, cores 1 and 2, (which are in the same socket too), use the following command:

./build/l3fwd-vf -l 1,2 -n 4 -- -p 0x3 --config="(0,0,1),(1,0,2)"

In this command:

  • The -l option enables cores 1 and 2

  • The -p option enables ports 0 and 1

  • The –config option enables one queue on each port and maps each (port,queue) pair to a specific core. The following table shows the mapping in this example:

    Port Queue lcore Description
    0 0 1 Map queue 0 from port 0 to lcore 1
    1 0 2 Map queue 0 from port 1 to lcore 2

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The operation of this application is similar to that of the basic L3 Forwarding Sample Application. See Explanation for more information.

Load Balancer Sample Application

The Load Balancer sample application demonstrates the concept of isolating the packet I/O task from the application-specific workload. Depending on the performance target, a number of logical cores (lcores) are dedicated to handle the interaction with the NIC ports (I/O lcores), while the rest of the lcores are dedicated to performing the application processing (worker lcores). The worker lcores are totally oblivious to the intricacies of the packet I/O activity and use the NIC-agnostic interface provided by software rings to exchange packets with the I/O cores.

Overview

The architecture of the Load Balance application is presented in the following figure.

_images/load_bal_app_arch.png

Load Balancer Application Architecture

For the sake of simplicity, the diagram illustrates a specific case of two I/O RX and two I/O TX lcores off loading the packet I/O overhead incurred by four NIC ports from four worker cores, with each I/O lcore handling RX/TX for two NIC ports.

I/O RX Logical Cores

Each I/O RX lcore performs packet RX from its assigned NIC RX rings and then distributes the received packets to the worker threads. The application allows each I/O RX lcore to communicate with any of the worker threads, therefore each (I/O RX lcore, worker lcore) pair is connected through a dedicated single producer - single consumer software ring.

The worker lcore to handle the current packet is determined by reading a predefined 1-byte field from the input packet:

worker_id = packet[load_balancing_field] % n_workers

Since all the packets that are part of the same traffic flow are expected to have the same value for the load balancing field, this scheme also ensures that all the packets that are part of the same traffic flow are directed to the same worker lcore (flow affinity) in the same order they enter the system (packet ordering).

I/O TX Logical Cores

Each I/O lcore owns the packet TX for a predefined set of NIC ports. To enable each worker thread to send packets to any NIC TX port, the application creates a software ring for each (worker lcore, NIC TX port) pair, with each I/O TX core handling those software rings that are associated with NIC ports that it handles.

Worker Logical Cores

Each worker lcore reads packets from its set of input software rings and routes them to the NIC ports for transmission by dispatching them to output software rings. The routing logic is LPM based, with all the worker threads sharing the same LPM rules.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the load_balancer sub-directory.

Running the Application

To successfully run the application, the command line used to start the application has to be in sync with the traffic flows configured on the traffic generator side.

For examples of application command lines and traffic generator flows, please refer to the DPDK Test Report. For more details on how to set up and run the sample applications provided with DPDK package, please refer to the DPDK Getting Started Guide.

Explanation

Application Configuration

The application run-time configuration is done through the application command line parameters. Any parameter that is not specified as mandatory is optional, with the default value hard-coded in the main.h header file from the application folder.

The list of application command line parameters is listed below:

  1. –rx “(PORT, QUEUE, LCORE), ...”: The list of NIC RX ports and queues handled by the I/O RX lcores. This parameter also implicitly defines the list of I/O RX lcores. This is a mandatory parameter.
  2. –tx “(PORT, LCORE), ... ”: The list of NIC TX ports handled by the I/O TX lcores. This parameter also implicitly defines the list of I/O TX lcores. This is a mandatory parameter.
  3. –w “LCORE, ...”: The list of the worker lcores. This is a mandatory parameter.
  4. –lpm “IP / PREFIX => PORT; ...”: The list of LPM rules used by the worker lcores for packet forwarding. This is a mandatory parameter.
  5. –rsz “A, B, C, D”: Ring sizes:
    1. A = The size (in number of buffer descriptors) of each of the NIC RX rings read by the I/O RX lcores.
    2. B = The size (in number of elements) of each of the software rings used by the I/O RX lcores to send packets to worker lcores.
    3. C = The size (in number of elements) of each of the software rings used by the worker lcores to send packets to I/O TX lcores.
    4. D = The size (in number of buffer descriptors) of each of the NIC TX rings written by I/O TX lcores.
  6. –bsz “(A, B), (C, D), (E, F)”: Burst sizes:
    1. A = The I/O RX lcore read burst size from NIC RX.
    2. B = The I/O RX lcore write burst size to the output software rings.
    3. C = The worker lcore read burst size from the input software rings.
    4. D = The worker lcore write burst size to the output software rings.
    5. E = The I/O TX lcore read burst size from the input software rings.
    6. F = The I/O TX lcore write burst size to the NIC TX.
  7. –pos-lb POS: The position of the 1-byte field within the input packet used by the I/O RX lcores to identify the worker lcore for the current packet. This field needs to be within the first 64 bytes of the input packet.

The infrastructure of software rings connecting I/O lcores and worker lcores is built by the application as a result of the application configuration provided by the user through the application command line parameters.

A specific lcore performing the I/O RX role for a specific set of NIC ports can also perform the I/O TX role for the same or a different set of NIC ports. A specific lcore cannot perform both the I/O role (either RX or TX) and the worker role during the same session.

Example:

./load_balancer -l 3-7 -n 4 -- --rx "(0,0,3),(1,0,3)" --tx "(0,3),(1,3)" --w "4,5,6,7" --lpm "1.0.0.0/24=>0; 1.0.1.0/24=>1;" --pos-lb 29

There is a single I/O lcore (lcore 3) that handles RX and TX for two NIC ports (ports 0 and 1) that handles packets to/from four worker lcores (lcores 4, 5, 6 and 7) that are assigned worker IDs 0 to 3 (worker ID for lcore 4 is 0, for lcore 5 is 1, for lcore 6 is 2 and for lcore 7 is 3).

Assuming that all the input packets are IPv4 packets with no VLAN label and the source IP address of the current packet is A.B.C.D, the worker lcore for the current packet is determined by byte D (which is byte 29). There are two LPM rules that are used by each worker lcore to route packets to the output NIC ports.

The following table illustrates the packet flow through the system for several possible traffic flows:

Flow # Source IP Address Destination IP Address Worker ID (Worker lcore) Output NIC Port
1 0.0.0.0 1.0.0.1 0 (4) 0
2 0.0.0.1 1.0.1.2 1 (5) 1
3 0.0.0.14 1.0.0.3 2 (6) 0
4 0.0.0.15 1.0.1.4 3 (7) 1
NUMA Support

The application has built-in performance enhancements for the NUMA case:

  1. One buffer pool per each CPU socket.
  2. One LPM table per each CPU socket.
  3. Memory for the NIC RX or TX rings is allocated on the same socket with the lcore handling the respective ring.

In the case where multiple CPU sockets are used in the system, it is recommended to enable at least one lcore to fulfill the I/O role for the NIC ports that are directly attached to that CPU socket through the PCI Express* bus. It is always recommended to handle the packet I/O with lcores from the same CPU socket as the NICs.

Depending on whether the I/O RX lcore (same CPU socket as NIC RX), the worker lcore and the I/O TX lcore (same CPU socket as NIC TX) handling a specific input packet, are on the same or different CPU sockets, the following run-time scenarios are possible:

  1. AAA: The packet is received, processed and transmitted without going across CPU sockets.
  2. AAB: The packet is received and processed on socket A, but as it has to be transmitted on a NIC port connected to socket B, the packet is sent to socket B through software rings.
  3. ABB: The packet is received on socket A, but as it has to be processed by a worker lcore on socket B, the packet is sent to socket B through software rings. The packet is transmitted by a NIC port connected to the same CPU socket as the worker lcore that processed it.
  4. ABC: The packet is received on socket A, it is processed by an lcore on socket B, then it has to be transmitted out by a NIC connected to socket C. The performance price for crossing the CPU socket boundary is paid twice for this packet.

Server-Node EFD Sample Application

This sample application demonstrates the use of EFD library as a flow-level load balancer, for more information about the EFD Library please refer to the DPDK programmer’s guide.

This sample application is a variant of the client-server sample application where a specific target node is specified for every and each flow (not in a round-robin fashion as the original load balancing sample application).

Overview

The architecture of the EFD flow-based load balancer sample application is presented in the following figure.

_images/server_node_efd.svg

Using EFD as a Flow-Level Load Balancer

As shown in Fig. 26.1, the sample application consists of a front-end node (server) using the EFD library to create a load-balancing table for flows, for each flow a target backend worker node is specified. The EFD table does not store the flow key (unlike a regular hash table), and hence, it can individually load-balance millions of flows (number of targets * maximum number of flows fit in a flow table per target) while still fitting in CPU cache.

It should be noted that although they are referred to as nodes, the frontend server and worker nodes are processes running on the same platform.

Front-end Server

Upon initializing, the frontend server node (process) creates a flow distributor table (based on the EFD library) which is populated with flow information and its intended target node.

The sample application assigns a specific target node_id (process) for each of the IP destination addresses as follows:

node_id = i % num_nodes; /* Target node id is generated */
ip_dst = rte_cpu_to_be_32(i); /* Specific ip destination address is
                                 assigned to this target node */

then the pair of <key,target> is inserted into the flow distribution table.

The main loop of the server process receives a burst of packets, then for each packet, a flow key (IP destination address) is extracted. The flow distributor table is looked up and the target node id is returned. Packets are then enqueued to the specified target node id.

It should be noted that flow distributor table is not a membership test table. I.e. if the key has already been inserted the target node id will be correct, but for new keys the flow distributor table will return a value (which can be valid).

Backend Worker Nodes

Upon initializing, the worker node (process) creates a flow table (a regular hash table that stores the key default size 1M flows) which is populated with only the flow information that is serviced at this node. This flow key is essential to point out new keys that have not been inserted before.

The worker node’s main loop is simply receiving packets then doing a hash table lookup. If a match occurs then statistics are updated for flows serviced by this node. If no match is found in the local hash table then this indicates that this is a new flow, which is dropped.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the server_node_efd sub-directory.

Running the Application

The application has two binaries to be run: the front-end server and the back-end node.

The frontend server (server) has the following command line options:

./server [EAL options] -- -p PORTMASK -n NUM_NODES -f NUM_FLOWS

Where,

  • -p PORTMASK: Hexadecimal bitmask of ports to configure
  • -n NUM_NODES: Number of back-end nodes that will be used
  • -f NUM_FLOWS: Number of flows to be added in the EFD table (1 million, by default)

The back-end node (node) has the following command line options:

./node [EAL options] -- -n NODE_ID

Where,

  • -n NODE_ID: Node ID, which cannot be equal or higher than NUM_MODES

First, the server app must be launched, with the number of nodes that will be run. Once it has been started, the node instances can be run, with different NODE_ID. These instances have to be run as secondary processes, with --proc-type=secondary in the EAL options, which will attach to the primary process memory, and therefore, they can access the queues created by the primary process to distribute packets.

To successfully run the application, the command line used to start the application has to be in sync with the traffic flows configured on the traffic generator side.

For examples of application command lines and traffic generator flows, please refer to the DPDK Test Report. For more details on how to set up and run the sample applications provided with DPDK package, please refer to the DPDK Getting Started Guide for Linux and DPDK Getting Started Guide for FreeBSD.

Explanation

As described in previous sections, there are two processes in this example.

The first process, the front-end server, creates and populates the EFD table, which is used to distribute packets to nodes, which the number of flows specified in the command line (1 million, by default).

static void
create_efd_table(void)
{
    uint8_t socket_id = rte_socket_id();

    /* create table */
    efd_table = rte_efd_create("flow table", num_flows * 2, sizeof(uint32_t),
                    1 << socket_id, socket_id);

    if (efd_table == NULL)
        rte_exit(EXIT_FAILURE, "Problem creating the flow table\n");
}

static void
populate_efd_table(void)
{
    unsigned int i;
    int32_t ret;
    uint32_t ip_dst;
    uint8_t socket_id = rte_socket_id();
    uint64_t node_id;

    /* Add flows in table */
    for (i = 0; i < num_flows; i++) {
        node_id = i % num_nodes;

        ip_dst = rte_cpu_to_be_32(i);
        ret = rte_efd_update(efd_table, socket_id,
                        (void *)&ip_dst, (efd_value_t)node_id);
        if (ret < 0)
            rte_exit(EXIT_FAILURE, "Unable to add entry %u in "
                                "EFD table\n", i);
    }

    printf("EFD table: Adding 0x%x keys\n", num_flows);
}

After initialization, packets are received from the enabled ports, and the IPv4 address from the packets is used as a key to look up in the EFD table, which tells the node where the packet has to be distributed.

static void
process_packets(uint32_t port_num __rte_unused, struct rte_mbuf *pkts[],
        uint16_t rx_count, unsigned int socket_id)
{
    uint16_t i;
    uint8_t node;
    efd_value_t data[EFD_BURST_MAX];
    const void *key_ptrs[EFD_BURST_MAX];

    struct ipv4_hdr *ipv4_hdr;
    uint32_t ipv4_dst_ip[EFD_BURST_MAX];

    for (i = 0; i < rx_count; i++) {
        /* Handle IPv4 header.*/
        ipv4_hdr = rte_pktmbuf_mtod_offset(pkts[i], struct ipv4_hdr *,
                sizeof(struct ether_hdr));
        ipv4_dst_ip[i] = ipv4_hdr->dst_addr;
        key_ptrs[i] = (void *)&ipv4_dst_ip[i];
    }

    rte_efd_lookup_bulk(efd_table, socket_id, rx_count,
                (const void **) key_ptrs, data);
    for (i = 0; i < rx_count; i++) {
        node = (uint8_t) ((uintptr_t)data[i]);

        if (node >= num_nodes) {
            /*
             * Node is out of range, which means that
             * flow has not been inserted
             */
            flow_dist_stats.drop++;
            rte_pktmbuf_free(pkts[i]);
        } else {
            flow_dist_stats.distributed++;
            enqueue_rx_packet(node, pkts[i]);
        }
    }

    for (i = 0; i < num_nodes; i++)
        flush_rx_queue(i);
}

The burst of packets received is enqueued in temporary buffers (per node), and enqueued in the shared ring between the server and the node. After this, a new burst of packets is received and this process is repeated infinitely.

static void
flush_rx_queue(uint16_t node)
{
    uint16_t j;
    struct node *cl;

    if (cl_rx_buf[node].count == 0)
        return;

    cl = &nodes[node];
    if (rte_ring_enqueue_bulk(cl->rx_q, (void **)cl_rx_buf[node].buffer,
            cl_rx_buf[node].count, NULL) != cl_rx_buf[node].count){
        for (j = 0; j < cl_rx_buf[node].count; j++)
            rte_pktmbuf_free(cl_rx_buf[node].buffer[j]);
        cl->stats.rx_drop += cl_rx_buf[node].count;
    } else
        cl->stats.rx += cl_rx_buf[node].count;

    cl_rx_buf[node].count = 0;
}

The second process, the back-end node, receives the packets from the shared ring with the server and send them out, if they belong to the node.

At initialization, it attaches to the server process memory, to have access to the shared ring, parameters and statistics.

rx_ring = rte_ring_lookup(get_rx_queue_name(node_id));
if (rx_ring == NULL)
    rte_exit(EXIT_FAILURE, "Cannot get RX ring - "
            "is server process running?\n");

mp = rte_mempool_lookup(PKTMBUF_POOL_NAME);
if (mp == NULL)
    rte_exit(EXIT_FAILURE, "Cannot get mempool for mbufs\n");

mz = rte_memzone_lookup(MZ_SHARED_INFO);
if (mz == NULL)
    rte_exit(EXIT_FAILURE, "Cannot get port info structure\n");
info = mz->addr;
tx_stats = &(info->tx_stats[node_id]);
filter_stats = &(info->filter_stats[node_id]);

Then, the hash table that contains the flows that will be handled by the node is created and populated.

static struct rte_hash *
create_hash_table(const struct shared_info *info)
{
    uint32_t num_flows_node = info->num_flows / info->num_nodes;
    char name[RTE_HASH_NAMESIZE];
    struct rte_hash *h;

    /* create table */
    struct rte_hash_parameters hash_params = {
        .entries = num_flows_node * 2, /* table load = 50% */
        .key_len = sizeof(uint32_t), /* Store IPv4 dest IP address */
        .socket_id = rte_socket_id(),
        .hash_func_init_val = 0,
    };

    snprintf(name, sizeof(name), "hash_table_%d", node_id);
    hash_params.name = name;
    h = rte_hash_create(&hash_params);

    if (h == NULL)
        rte_exit(EXIT_FAILURE,
                "Problem creating the hash table for node %d\n",
                node_id);
    return h;
}

static void
populate_hash_table(const struct rte_hash *h, const struct shared_info *info)
{
    unsigned int i;
    int32_t ret;
    uint32_t ip_dst;
    uint32_t num_flows_node = 0;
    uint64_t target_node;

    /* Add flows in table */
    for (i = 0; i < info->num_flows; i++) {
        target_node = i % info->num_nodes;
        if (target_node != node_id)
            continue;

        ip_dst = rte_cpu_to_be_32(i);

        ret = rte_hash_add_key(h, (void *) &ip_dst);
        if (ret < 0)
            rte_exit(EXIT_FAILURE, "Unable to add entry %u "
                    "in hash table\n", i);
        else
            num_flows_node++;

    }

    printf("Hash table: Adding 0x%x keys\n", num_flows_node);
}

After initialization, packets are dequeued from the shared ring (from the server) and, like in the server process, the IPv4 address from the packets is used as a key to look up in the hash table. If there is a hit, packet is stored in a buffer, to be eventually transmitted in one of the enabled ports. If key is not there, packet is dropped, since the flow is not handled by the node.

static inline void
handle_packets(struct rte_hash *h, struct rte_mbuf **bufs, uint16_t num_packets)
{
    struct ipv4_hdr *ipv4_hdr;
    uint32_t ipv4_dst_ip[PKT_READ_SIZE];
    const void *key_ptrs[PKT_READ_SIZE];
    unsigned int i;
    int32_t positions[PKT_READ_SIZE] = {0};

    for (i = 0; i < num_packets; i++) {
        /* Handle IPv4 header.*/
        ipv4_hdr = rte_pktmbuf_mtod_offset(bufs[i], struct ipv4_hdr *,
                sizeof(struct ether_hdr));
        ipv4_dst_ip[i] = ipv4_hdr->dst_addr;
        key_ptrs[i] = &ipv4_dst_ip[i];
    }
    /* Check if packets belongs to any flows handled by this node */
    rte_hash_lookup_bulk(h, key_ptrs, num_packets, positions);

    for (i = 0; i < num_packets; i++) {
        if (likely(positions[i] >= 0)) {
            filter_stats->passed++;
            transmit_packet(bufs[i]);
        } else {
            filter_stats->drop++;
            /* Drop packet, as flow is not handled by this node */
            rte_pktmbuf_free(bufs[i]);
        }
    }
}

Finally, note that both processes updates statistics, such as transmitted, received and dropped packets, which are shown and refreshed by the server app.

static void
do_stats_display(void)
{
    unsigned int i, j;
    const char clr[] = {27, '[', '2', 'J', '\0'};
    const char topLeft[] = {27, '[', '1', ';', '1', 'H', '\0'};
    uint64_t port_tx[RTE_MAX_ETHPORTS], port_tx_drop[RTE_MAX_ETHPORTS];
    uint64_t node_tx[MAX_NODES], node_tx_drop[MAX_NODES];

    /* to get TX stats, we need to do some summing calculations */
    memset(port_tx, 0, sizeof(port_tx));
    memset(port_tx_drop, 0, sizeof(port_tx_drop));
    memset(node_tx, 0, sizeof(node_tx));
    memset(node_tx_drop, 0, sizeof(node_tx_drop));

    for (i = 0; i < num_nodes; i++) {
        const struct tx_stats *tx = &info->tx_stats[i];

        for (j = 0; j < info->num_ports; j++) {
            const uint64_t tx_val = tx->tx[info->id[j]];
            const uint64_t drop_val = tx->tx_drop[info->id[j]];

            port_tx[j] += tx_val;
            port_tx_drop[j] += drop_val;
            node_tx[i] += tx_val;
            node_tx_drop[i] += drop_val;
        }
    }

    /* Clear screen and move to top left */
    printf("%s%s", clr, topLeft);

    printf("PORTS\n");
    printf("-----\n");
    for (i = 0; i < info->num_ports; i++)
        printf("Port %u: '%s'\t", (unsigned int)info->id[i],
                get_printable_mac_addr(info->id[i]));
    printf("\n\n");
    for (i = 0; i < info->num_ports; i++) {
        printf("Port %u - rx: %9"PRIu64"\t"
                "tx: %9"PRIu64"\n",
                (unsigned int)info->id[i], info->rx_stats.rx[i],
                port_tx[i]);
    }

    printf("\nSERVER\n");
    printf("-----\n");
    printf("distributed: %9"PRIu64", drop: %9"PRIu64"\n",
            flow_dist_stats.distributed, flow_dist_stats.drop);

    printf("\nNODES\n");
    printf("-------\n");
    for (i = 0; i < num_nodes; i++) {
        const unsigned long long rx = nodes[i].stats.rx;
        const unsigned long long rx_drop = nodes[i].stats.rx_drop;
        const struct filter_stats *filter = &info->filter_stats[i];

        printf("Node %2u - rx: %9llu, rx_drop: %9llu\n"
                "            tx: %9"PRIu64", tx_drop: %9"PRIu64"\n"
                "            filter_passed: %9"PRIu64", "
                "filter_drop: %9"PRIu64"\n",
                i, rx, rx_drop, node_tx[i], node_tx_drop[i],
                filter->passed, filter->drop);
    }

    printf("\n");
}

Service Cores Sample Application

The service cores sample application demonstrates the service cores capabilities of DPDK. The service cores infrastructure is part of the DPDK EAL, and allows any DPDK component to register a service. A service is a work item or task, that requires CPU time to perform its duty.

This sample application registers 5 dummy services. These 5 services are used to show how the service_cores API can be used to orchestrate these services to run on different service lcores. This orchestration is done by calling the service cores APIs, however the sample application introduces a “profile” concept to contain the service mapping details. Note that the profile concept is application specific, and not a part of the service cores API.

Compiling the Application

  1. Go to the example directory:

    export RTE_SDK=/path/to/rte_sdk
    cd ${RTE_SDK}/examples/service_cores
    
  2. Set the target (a default target is used if not specified). For example:

    export RTE_TARGET=x86_64-native-linuxapp-gcc
    

    See the DPDK Getting Started Guide for possible RTE_TARGET values.

  3. Build the application:

    make
    

Running the Application

To run the example, just execute the binary. Since the application dynamically adds service cores in the application code itself, there is no requirement to pass a service core-mask as an EAL argument at startup time.

$ ./build/service_cores

Explanation

The following sections provide some explanation of code focusing on registering applications from an applications point of view, and modifying the service core counts and mappings at runtime.

Registering a Service

The following code section shows how to register a service as an application. Note that the service component header must be included by the application in order to register services: rte_service_component.h, in addition to the ordinary service cores header rte_service.h which provides the runtime functions to add, remove and remap service cores.

struct rte_service_spec service = {
        .name = "service_name",
};
int ret = rte_service_component_register(services, &id);
if (ret)
        return -1;

/* set the service itself to be ready to run. In the case of
* ethdev, eventdev etc PMDs, this will be set when the
* appropriate configure or setup function is called.
*/
rte_service_component_runstate_set(id, 1);

/* Collect statistics for the service */
rte_service_set_stats_enable(id, 1);

/* The application sets the service to running state. Note that this
 * function enables the service to run - while the 'component' version
 * of this function (as above) marks the service itself as ready */
ret = rte_service_runstate_set(id, 1);
Controlling A Service Core

This section demonstrates how to add a service core. The rte_service.h header file provides the functions for dynamically adding and removing cores. The APIs to add and remove cores use lcore IDs similar to existing DPDK functions.

These are the functions to start a service core, and have it run a service:

/* the lcore ID to use as a service core */
uint32_t service_core_id = 7;
ret = rte_service_lcore_add(service_core_id);
if(ret)
        return -1;

/* service cores are in "stopped" state when added, so start it */
ret = rte_service_lcore_start(service_core_id);
if(ret)
        return -1;

/* map a service to the service core, causing it to run the service */
uint32_t service_id; /* ID of a registered service */
uint32_t enable = 1; /* 1 maps the service, 0 unmaps */
ret = rte_service_map_lcore_set(service_id, service_core_id, enable);
if(ret)
        return -1;
Removing A Service Core

To remove a service core, the steps are similar to adding but in reverse order. Note that it is not allowed to remove a service core if the service is running, and the service-core is the only core running that service (see documentation for rte_service_lcore_stop function for details).

Conclusion

The service cores infrastructure provides DPDK with two main features. The first is to abstract away hardware differences: the service core can CPU cycles to a software fallback implementation, allowing the application to be abstracted from the difference in HW / SW availability. The second feature is a flexible method of registering functions to be run, allowing the running of the functions to be scaled across multiple CPUs.

Multi-process Sample Application

This chapter describes the example applications for multi-processing that are included in the DPDK.

Example Applications

Building the Sample Applications

The multi-process example applications are built in the same way as other sample applications, and as documented in the DPDK Getting Started Guide.

To compile the sample application see Compiling the Sample Applications.

The applications are located in the multi_process sub-directory.

Note

If just a specific multi-process application needs to be built, the final make command can be run just in that application’s directory, rather than at the top-level multi-process directory.

Basic Multi-process Example

The examples/simple_mp folder in the DPDK release contains a basic example application to demonstrate how two DPDK processes can work together using queues and memory pools to share information.

Running the Application

To run the application, start one copy of the simple_mp binary in one terminal, passing at least two cores in the coremask/corelist, as follows:

./build/simple_mp -l 0-1 -n 4 --proc-type=primary

For the first DPDK process run, the proc-type flag can be omitted or set to auto, since all DPDK processes will default to being a primary instance, meaning they have control over the hugepage shared memory regions. The process should start successfully and display a command prompt as follows:

$ ./build/simple_mp -l 0-1 -n 4 --proc-type=primary
EAL: coremask set to 3
EAL: Detected lcore 0 on socket 0
EAL: Detected lcore 1 on socket 0
EAL: Detected lcore 2 on socket 0
EAL: Detected lcore 3 on socket 0
...

EAL: Requesting 2 pages of size 1073741824
EAL: Requesting 768 pages of size 2097152
EAL: Ask a virtual area of 0x40000000 bytes
EAL: Virtual area found at 0x7ff200000000 (size = 0x40000000)
...

EAL: check igb_uio module
EAL: check module finished
EAL: Master core 0 is ready (tid=54e41820)
EAL: Core 1 is ready (tid=53b32700)

Starting core 1

simple_mp >

To run the secondary process to communicate with the primary process, again run the same binary setting at least two cores in the coremask/corelist:

./build/simple_mp -l 2-3 -n 4 --proc-type=secondary

When running a secondary process such as that shown above, the proc-type parameter can again be specified as auto. However, omitting the parameter altogether will cause the process to try and start as a primary rather than secondary process.

Once the process type is specified correctly, the process starts up, displaying largely similar status messages to the primary instance as it initializes. Once again, you will be presented with a command prompt.

Once both processes are running, messages can be sent between them using the send command. At any stage, either process can be terminated using the quit command.

EAL: Master core 10 is ready (tid=b5f89820)           EAL: Master core 8 is ready (tid=864a3820)
EAL: Core 11 is ready (tid=84ffe700)                  EAL: Core 9 is ready (tid=85995700)
Starting core 11                                      Starting core 9
simple_mp > send hello_secondary                      simple_mp > core 9: Received 'hello_secondary'
simple_mp > core 11: Received 'hello_primary'         simple_mp > send hello_primary
simple_mp > quit                                      simple_mp > quit

Note

If the primary instance is terminated, the secondary instance must also be shut-down and restarted after the primary. This is necessary because the primary instance will clear and reset the shared memory regions on startup, invalidating the secondary process’s pointers. The secondary process can be stopped and restarted without affecting the primary process.

How the Application Works

The core of this example application is based on using two queues and a single memory pool in shared memory. These three objects are created at startup by the primary process, since the secondary process cannot create objects in memory as it cannot reserve memory zones, and the secondary process then uses lookup functions to attach to these objects as it starts up.

if (rte_eal_process_type() == RTE_PROC_PRIMARY){
    send_ring = rte_ring_create(_PRI_2_SEC, ring_size, SOCKET0, flags);
    recv_ring = rte_ring_create(_SEC_2_PRI, ring_size, SOCKET0, flags);
    message_pool = rte_mempool_create(_MSG_POOL, pool_size, string_size, pool_cache, priv_data_sz, NULL, NULL, NULL, NULL, SOCKET0, flags);
} else {
    recv_ring = rte_ring_lookup(_PRI_2_SEC);
    send_ring = rte_ring_lookup(_SEC_2_PRI);
    message_pool = rte_mempool_lookup(_MSG_POOL);
}

Note, however, that the named ring structure used as send_ring in the primary process is the recv_ring in the secondary process.

Once the rings and memory pools are all available in both the primary and secondary processes, the application simply dedicates two threads to sending and receiving messages respectively. The receive thread simply dequeues any messages on the receive ring, prints them, and frees the buffer space used by the messages back to the memory pool. The send thread makes use of the command-prompt library to interactively request user input for messages to send. Once a send command is issued by the user, a buffer is allocated from the memory pool, filled in with the message contents, then enqueued on the appropriate rte_ring.

Symmetric Multi-process Example

The second example of DPDK multi-process support demonstrates how a set of processes can run in parallel, with each process performing the same set of packet- processing operations. (Since each process is identical in functionality to the others, we refer to this as symmetric multi-processing, to differentiate it from asymmetric multi- processing - such as a client-server mode of operation seen in the next example, where different processes perform different tasks, yet co-operate to form a packet-processing system.) The following diagram shows the data-flow through the application, using two processes.

_images/sym_multi_proc_app.png

Example Data Flow in a Symmetric Multi-process Application

As the diagram shows, each process reads packets from each of the network ports in use. RSS is used to distribute incoming packets on each port to different hardware RX queues. Each process reads a different RX queue on each port and so does not contend with any other process for that queue access. Similarly, each process writes outgoing packets to a different TX queue on each port.

Running the Application

As with the simple_mp example, the first instance of the symmetric_mp process must be run as the primary instance, though with a number of other application- specific parameters also provided after the EAL arguments. These additional parameters are:

  • -p <portmask>, where portmask is a hexadecimal bitmask of what ports on the system are to be used. For example: -p 3 to use ports 0 and 1 only.
  • –num-procs <N>, where N is the total number of symmetric_mp instances that will be run side-by-side to perform packet processing. This parameter is used to configure the appropriate number of receive queues on each network port.
  • –proc-id <n>, where n is a numeric value in the range 0 <= n < N (number of processes, specified above). This identifies which symmetric_mp instance is being run, so that each process can read a unique receive queue on each network port.

The secondary symmetric_mp instances must also have these parameters specified, and the first two must be the same as those passed to the primary instance, or errors result.

For example, to run a set of four symmetric_mp instances, running on lcores 1-4, all performing level-2 forwarding of packets between ports 0 and 1, the following commands can be used (assuming run as root):

# ./build/symmetric_mp -l 1 -n 4 --proc-type=auto -- -p 3 --num-procs=4 --proc-id=0
# ./build/symmetric_mp -l 2 -n 4 --proc-type=auto -- -p 3 --num-procs=4 --proc-id=1
# ./build/symmetric_mp -l 3 -n 4 --proc-type=auto -- -p 3 --num-procs=4 --proc-id=2
# ./build/symmetric_mp -l 4 -n 4 --proc-type=auto -- -p 3 --num-procs=4 --proc-id=3

Note

In the above example, the process type can be explicitly specified as primary or secondary, rather than auto. When using auto, the first process run creates all the memory structures needed for all processes - irrespective of whether it has a proc-id of 0, 1, 2 or 3.

Note

For the symmetric multi-process example, since all processes work in the same manner, once the hugepage shared memory and the network ports are initialized, it is not necessary to restart all processes if the primary instance dies. Instead, that process can be restarted as a secondary, by explicitly setting the proc-type to secondary on the command line. (All subsequent instances launched will also need this explicitly specified, as auto-detection will detect no primary processes running and therefore attempt to re-initialize shared memory.)

How the Application Works

The initialization calls in both the primary and secondary instances are the same for the most part, calling the rte_eal_init(), 1 G and 10 G driver initialization and then rte_pci_probe() functions. Thereafter, the initialization done depends on whether the process is configured as a primary or secondary instance.

In the primary instance, a memory pool is created for the packet mbufs and the network ports to be used are initialized - the number of RX and TX queues per port being determined by the num-procs parameter passed on the command-line. The structures for the initialized network ports are stored in shared memory and therefore will be accessible by the secondary process as it initializes.

if (num_ports & 1)
   rte_exit(EXIT_FAILURE, "Application must use an even number of ports\n");

for(i = 0; i < num_ports; i++){
    if(proc_type == RTE_PROC_PRIMARY)
        if (smp_port_init(ports[i], mp, (uint16_t)num_procs) < 0)
            rte_exit(EXIT_FAILURE, "Error initializing ports\n");
}

In the secondary instance, rather than initializing the network ports, the port information exported by the primary process is used, giving the secondary process access to the hardware and software rings for each network port. Similarly, the memory pool of mbufs is accessed by doing a lookup for it by name:

mp = (proc_type == RTE_PROC_SECONDARY) ? rte_mempool_lookup(_SMP_MBUF_POOL) : rte_mempool_create(_SMP_MBUF_POOL, NB_MBUFS, MBUF_SIZE, ... )

Once this initialization is complete, the main loop of each process, both primary and secondary, is exactly the same - each process reads from each port using the queue corresponding to its proc-id parameter, and writes to the corresponding transmit queue on the output port.

Client-Server Multi-process Example

The third example multi-process application included with the DPDK shows how one can use a client-server type multi-process design to do packet processing. In this example, a single server process performs the packet reception from the ports being used and distributes these packets using round-robin ordering among a set of client processes, which perform the actual packet processing. In this case, the client applications just perform level-2 forwarding of packets by sending each packet out on a different network port.

The following diagram shows the data-flow through the application, using two client processes.

_images/client_svr_sym_multi_proc_app.png

Example Data Flow in a Client-Server Symmetric Multi-process Application

Running the Application

The server process must be run initially as the primary process to set up all memory structures for use by the clients. In addition to the EAL parameters, the application- specific parameters are:

  • -p <portmask >, where portmask is a hexadecimal bitmask of what ports on the system are to be used. For example: -p 3 to use ports 0 and 1 only.
  • -n <num-clients>, where the num-clients parameter is the number of client processes that will process the packets received by the server application.

Note

In the server process, a single thread, the master thread, that is, the lowest numbered lcore in the coremask/corelist, performs all packet I/O. If a coremask/corelist is specified with more than a single lcore bit set in it, an additional lcore will be used for a thread to periodically print packet count statistics.

Since the server application stores configuration data in shared memory, including the network ports to be used, the only application parameter needed by a client process is its client instance ID. Therefore, to run a server application on lcore 1 (with lcore 2 printing statistics) along with two client processes running on lcores 3 and 4, the following commands could be used:

# ./mp_server/build/mp_server -l 1-2 -n 4 -- -p 3 -n 2
# ./mp_client/build/mp_client -l 3 -n 4 --proc-type=auto -- -n 0
# ./mp_client/build/mp_client -l 4 -n 4 --proc-type=auto -- -n 1

Note

If the server application dies and needs to be restarted, all client applications also need to be restarted, as there is no support in the server application for it to run as a secondary process. Any client processes that need restarting can be restarted without affecting the server process.

How the Application Works

The server process performs the network port and data structure initialization much as the symmetric multi-process application does when run as primary. One additional enhancement in this sample application is that the server process stores its port configuration data in a memory zone in hugepage shared memory. This eliminates the need for the client processes to have the portmask parameter passed into them on the command line, as is done for the symmetric multi-process application, and therefore eliminates mismatched parameters as a potential source of errors.

In the same way that the server process is designed to be run as a primary process instance only, the client processes are designed to be run as secondary instances only. They have no code to attempt to create shared memory objects. Instead, handles to all needed rings and memory pools are obtained via calls to rte_ring_lookup() and rte_mempool_lookup(). The network ports for use by the processes are obtained by loading the network port drivers and probing the PCI bus, which will, as in the symmetric multi-process example, automatically get access to the network ports using the settings already configured by the primary/server process.

Once all applications are initialized, the server operates by reading packets from each network port in turn and distributing those packets to the client queues (software rings, one for each client process) in round-robin order. On the client side, the packets are read from the rings in as big of bursts as possible, then routed out to a different network port. The routing used is very simple. All packets received on the first NIC port are transmitted back out on the second port and vice versa. Similarly, packets are routed between the 3rd and 4th network ports and so on. The sending of packets is done by writing the packets directly to the network ports; they are not transferred back via the server process.

In both the server and the client processes, outgoing packets are buffered before being sent, so as to allow the sending of multiple packets in a single burst to improve efficiency. For example, the client process will buffer packets to send, until either the buffer is full or until we receive no further packets from the server.

Master-slave Multi-process Example

The fourth example of DPDK multi-process support demonstrates a master-slave model that provide the capability of application recovery if a slave process crashes or meets unexpected conditions. In addition, it also demonstrates the floating process, which can run among different cores in contrast to the traditional way of binding a process/thread to a specific CPU core, using the local cache mechanism of mempool structures.

This application performs the same functionality as the L2 Forwarding sample application, therefore this chapter does not cover that part but describes functionality that is introduced in this multi-process example only. Please refer to L2 Forwarding Sample Application (in Real and Virtualized Environments) for more information.

Unlike previous examples where all processes are started from the command line with input arguments, in this example, only one process is spawned from the command line and that process creates other processes. The following section describes this in more detail.

Master-slave Process Models

The process spawned from the command line is called the master process in this document. A process created by the master is called a slave process. The application has only one master process, but could have multiple slave processes.

Once the master process begins to run, it tries to initialize all the resources such as memory, CPU cores, driver, ports, and so on, as the other examples do. Thereafter, it creates slave processes, as shown in the following figure.

_images/master_slave_proc.png

Master-slave Process Workflow

The master process calls the rte_eal_mp_remote_launch() EAL function to launch an application function for each pinned thread through the pipe. Then, it waits to check if any slave processes have exited. If so, the process tries to re-initialize the resources that belong to that slave and launch them in the pinned thread entry again. The following section describes the recovery procedures in more detail.

For each pinned thread in EAL, after reading any data from the pipe, it tries to call the function that the application specified. In this master specified function, a fork() call creates a slave process that performs the L2 forwarding task. Then, the function waits until the slave exits, is killed or crashes. Thereafter, it notifies the master of this event and returns. Finally, the EAL pinned thread waits until the new function is launched.

After discussing the master-slave model, it is necessary to mention another issue, global and static variables.

For multiple-thread cases, all global and static variables have only one copy and they can be accessed by any thread if applicable. So, they can be used to sync or share data among threads.

In the previous examples, each process has separate global and static variables in memory and are independent of each other. If it is necessary to share the knowledge, some communication mechanism should be deployed, such as, memzone, ring, shared memory, and so on. The global or static variables are not a valid approach to share data among processes. For variables in this example, on the one hand, the slave process inherits all the knowledge of these variables after being created by the master. On the other hand, other processes cannot know if one or more processes modifies them after slave creation since that is the nature of a multiple process address space. But this does not mean that these variables cannot be used to share or sync data; it depends on the use case. The following are the possible use cases:

  1. The master process starts and initializes a variable and it will never be changed after slave processes created. This case is OK.
  2. After the slave processes are created, the master or slave cores need to change a variable, but other processes do not need to know the change. This case is also OK.
  3. After the slave processes are created, the master or a slave needs to change a variable. In the meantime, one or more other process needs to be aware of the change. In this case, global and static variables cannot be used to share knowledge. Another communication mechanism is needed. A simple approach without lock protection can be a heap buffer allocated by rte_malloc or mem zone.
Slave Process Recovery Mechanism

Before talking about the recovery mechanism, it is necessary to know what is needed before a new slave instance can run if a previous one exited.

When a slave process exits, the system returns all the resources allocated for this process automatically. However, this does not include the resources that were allocated by the DPDK. All the hardware resources are shared among the processes, which include memzone, mempool, ring, a heap buffer allocated by the rte_malloc library, and so on. If the new instance runs and the allocated resource is not returned, either resource allocation failed or the hardware resource is lost forever.

When a slave process runs, it may have dependencies on other processes. They could have execution sequence orders; they could share the ring to communicate; they could share the same port for reception and forwarding; they could use lock structures to do exclusive access in some critical path. What happens to the dependent process(es) if the peer leaves? The consequence are varied since the dependency cases are complex. It depends on what the processed had shared. However, it is necessary to notify the peer(s) if one slave exited. Then, the peer(s) will be aware of that and wait until the new instance begins to run.

Therefore, to provide the capability to resume the new slave instance if the previous one exited, it is necessary to provide several mechanisms:

  1. Keep a resource list for each slave process. Before a slave process run, the master should prepare a resource list. After it exits, the master could either delete the allocated resources and create new ones, or re-initialize those for use by the new instance.
  2. Set up a notification mechanism for slave process exit cases. After the specific slave leaves, the master should be notified and then help to create a new instance. This mechanism is provided in Section Master-slave Process Models.
  3. Use a synchronization mechanism among dependent processes. The master should have the capability to stop or kill slave processes that have a dependency on the one that has exited. Then, after the new instance of exited slave process begins to run, the dependency ones could resume or run from the start. The example sends a STOP command to slave processes dependent on the exited one, then they will exit. Thereafter, the master creates new instances for the exited slave processes.

The following diagram describes slave process recovery.

_images/slave_proc_recov.png

Slave Process Recovery Process Flow

Floating Process Support

When the DPDK application runs, there is always a -c option passed in to indicate the cores that are enabled. Then, the DPDK creates a thread for each enabled core. By doing so, it creates a 1:1 mapping between the enabled core and each thread. The enabled core always has an ID, therefore, each thread has a unique core ID in the DPDK execution environment. With the ID, each thread can easily access the structures or resources exclusively belonging to it without using function parameter passing. It can easily use the rte_lcore_id() function to get the value in every function that is called.

For threads/processes not created in that way, either pinned to a core or not, they will not own a unique ID and the rte_lcore_id() function will not work in the correct way. However, sometimes these threads/processes still need the unique ID mechanism to do easy access on structures or resources. For example, the DPDK mempool library provides a local cache mechanism (refer to Local Cache) for fast element allocation and freeing. If using a non-unique ID or a fake one, a race condition occurs if two or more threads/ processes with the same core ID try to use the local cache.

Therefore, unused core IDs from the passing of parameters with the -c option are used to organize the core ID allocation array. Once the floating process is spawned, it tries to allocate a unique core ID from the array and release it on exit.

A natural way to spawn a floating process is to use the fork() function and allocate a unique core ID from the unused core ID array. However, it is necessary to write new code to provide a notification mechanism for slave exit and make sure the process recovery mechanism can work with it.

To avoid producing redundant code, the Master-Slave process model is still used to spawn floating processes, then cancel the affinity to specific cores. Besides that, clear the core ID assigned to the DPDK spawning a thread that has a 1:1 mapping with the core mask. Thereafter, get a new core ID from the unused core ID allocation array.

Run the Application

This example has a command line similar to the L2 Forwarding sample application with a few differences.

To run the application, start one copy of the l2fwd_fork binary in one terminal. Unlike the L2 Forwarding example, this example requires at least three cores since the master process will wait and be accountable for slave process recovery. The command is as follows:

#./build/l2fwd_fork -l 2-4 -n 4 -- -p 3 -f

This example provides another -f option to specify the use of floating process. If not specified, the example will use a pinned process to perform the L2 forwarding task.

To verify the recovery mechanism, proceed as follows: First, check the PID of the slave processes:

#ps -fe | grep l2fwd_fork
root 5136 4843 29 11:11 pts/1 00:00:05 ./build/l2fwd_fork
root 5145 5136 98 11:11 pts/1 00:00:11 ./build/l2fwd_fork
root 5146 5136 98 11:11 pts/1 00:00:11 ./build/l2fwd_fork

Then, kill one of the slaves:

#kill -9 5145

After 1 or 2 seconds, check whether the slave has resumed:

#ps -fe | grep l2fwd_fork
root 5136 4843 3 11:11 pts/1 00:00:06 ./build/l2fwd_fork
root 5247 5136 99 11:14 pts/1 00:00:01 ./build/l2fwd_fork
root 5248 5136 99 11:14 pts/1 00:00:01 ./build/l2fwd_fork

It can also monitor the traffic generator statics to see whether slave processes have resumed.

Explanation

As described in previous sections, not all global and static variables need to change to be accessible in multiple processes; it depends on how they are used. In this example, the statics info on packets dropped/forwarded/received count needs to be updated by the slave process, and the master needs to see the update and print them out. So, it needs to allocate a heap buffer using rte_zmalloc. In addition, if the -f option is specified, an array is needed to store the allocated core ID for the floating process so that the master can return it after a slave has exited accidentally.

static int
l2fwd_malloc_shared_struct(void)
{
    port_statistics = rte_zmalloc("port_stat", sizeof(struct l2fwd_port_statistics) * RTE_MAX_ETHPORTS, 0);

    if (port_statistics == NULL)
        return -1;

    /* allocate mapping_id array */

    if (float_proc) {
        int i;

        mapping_id = rte_malloc("mapping_id", sizeof(unsigned) * RTE_MAX_LCORE, 0);
        if (mapping_id == NULL)
            return -1;

        for (i = 0 ;i < RTE_MAX_LCORE; i++)
            mapping_id[i] = INVALID_MAPPING_ID;

    }
    return 0;
}

For each slave process, packets are received from one port and forwarded to another port that another slave is operating on. If the other slave exits accidentally, the port it is operating on may not work normally, so the first slave cannot forward packets to that port. There is a dependency on the port in this case. So, the master should recognize the dependency. The following is the code to detect this dependency:

for (portid = 0; portid < nb_ports; portid++) {
    /* skip ports that are not enabled */

    if ((l2fwd_enabled_port_mask & (1 << portid)) == 0)
        continue;

    /* Find pair ports' lcores */

    find_lcore = find_pair_lcore = 0;
    pair_port = l2fwd_dst_ports[portid];

    for (i = 0; i < RTE_MAX_LCORE; i++) {
        if (!rte_lcore_is_enabled(i))
            continue;

        for (j = 0; j < lcore_queue_conf[i].n_rx_port;j++) {
            if (lcore_queue_conf[i].rx_port_list[j] == portid) {
                lcore = i;
                find_lcore = 1;
                break;
            }

            if (lcore_queue_conf[i].rx_port_list[j] == pair_port) {
                pair_lcore = i;
                find_pair_lcore = 1;
                break;
            }
        }

        if (find_lcore && find_pair_lcore)
            break;
    }

    if (!find_lcore || !find_pair_lcore)
        rte_exit(EXIT_FAILURE, "Not find port=%d pair\\n", portid);

    printf("lcore %u and %u paired\\n", lcore, pair_lcore);

    lcore_resource[lcore].pair_id = pair_lcore;
    lcore_resource[pair_lcore].pair_id = lcore;
}

Before launching the slave process, it is necessary to set up the communication channel between the master and slave so that the master can notify the slave if its peer process with the dependency exited. In addition, the master needs to register a callback function in the case where a specific slave exited.

for (i = 0; i < RTE_MAX_LCORE; i++) {
    if (lcore_resource[i].enabled) {
        /* Create ring for master and slave communication */

        ret = create_ms_ring(i);
        if (ret != 0)
            rte_exit(EXIT_FAILURE, "Create ring for lcore=%u failed",i);

        if (flib_register_slave_exit_notify(i,slave_exit_cb) != 0)
            rte_exit(EXIT_FAILURE, "Register master_trace_slave_exit failed");
    }
}

After launching the slave process, the master waits and prints out the port statics periodically. If an event indicating that a slave process exited is detected, it sends the STOP command to the peer and waits until it has also exited. Then, it tries to clean up the execution environment and prepare new resources. Finally, the new slave instance is launched.

while (1) {
    sleep(1);
    cur_tsc = rte_rdtsc();
    diff_tsc = cur_tsc - prev_tsc;

    /* if timer is enabled */

    if (timer_period > 0) {
        /* advance the timer */
        timer_tsc += diff_tsc;

        /* if timer has reached its timeout */
        if (unlikely(timer_tsc >= (uint64_t) timer_period)) {
            print_stats();

            /* reset the timer */
            timer_tsc = 0;
        }
    }

    prev_tsc = cur_tsc;

    /* Check any slave need restart or recreate */

    rte_spinlock_lock(&res_lock);

    for (i = 0; i < RTE_MAX_LCORE; i++) {
        struct lcore_resource_struct *res = &lcore_resource[i];
        struct lcore_resource_struct *pair = &lcore_resource[res->pair_id];

        /* If find slave exited, try to reset pair */

        if (res->enabled && res->flags && pair->enabled) {
            if (!pair->flags) {
                master_sendcmd_with_ack(pair->lcore_id, CMD_STOP);
                rte_spinlock_unlock(&res_lock);
                sleep(1);
                rte_spinlock_lock(&res_lock);
                if (pair->flags)
                    continue;
            }

            if (reset_pair(res->lcore_id, pair->lcore_id) != 0)
                rte_exit(EXIT_FAILURE, "failed to reset slave");

            res->flags = 0;
            pair->flags = 0;
        }
    }
    rte_spinlock_unlock(&res_lock);
}

When the slave process is spawned and starts to run, it checks whether the floating process option is applied. If so, it clears the affinity to a specific core and also sets the unique core ID to 0. Then, it tries to allocate a new core ID. Since the core ID has changed, the resource allocated by the master cannot work, so it remaps the resource to the new core ID slot.

static int
l2fwd_launch_one_lcore( attribute ((unused)) void *dummy)
{
    unsigned lcore_id = rte_lcore_id();

    if (float_proc) {
        unsigned flcore_id;

        /* Change it to floating process, also change it's lcore_id */

        clear_cpu_affinity();

        RTE_PER_LCORE(_lcore_id) = 0;

        /* Get a lcore_id */

        if (flib_assign_lcore_id() < 0 ) {
            printf("flib_assign_lcore_id failed\n");
            return -1;
        }

        flcore_id = rte_lcore_id();

        /* Set mapping id, so master can return it after slave exited */

        mapping_id[lcore_id] = flcore_id;
        printf("Org lcore_id = %u, cur lcore_id = %u\n",lcore_id, flcore_id);
        remapping_slave_resource(lcore_id, flcore_id);
    }

    l2fwd_main_loop();

    /* return lcore_id before return */
    if (float_proc) {
        flib_free_lcore_id(rte_lcore_id());
        mapping_id[lcore_id] = INVALID_MAPPING_ID;
    }
    return 0;
}

QoS Metering Sample Application

The QoS meter sample application is an example that demonstrates the use of DPDK to provide QoS marking and metering, as defined by RFC2697 for Single Rate Three Color Marker (srTCM) and RFC 2698 for Two Rate Three Color Marker (trTCM) algorithm.

Overview

The application uses a single thread for reading the packets from the RX port, metering, marking them with the appropriate color (green, yellow or red) and writing them to the TX port.

A policing scheme can be applied before writing the packets to the TX port by dropping or changing the color of the packet in a static manner depending on both the input and output colors of the packets that are processed by the meter.

The operation mode can be selected as compile time out of the following options:

  • Simple forwarding
  • srTCM color blind
  • srTCM color aware
  • srTCM color blind
  • srTCM color aware

Please refer to RFC2697 and RFC2698 for details about the srTCM and trTCM configurable parameters (CIR, CBS and EBS for srTCM; CIR, PIR, CBS and PBS for trTCM).

The color blind modes are functionally equivalent with the color-aware modes when all the incoming packets are colored as green.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the qos_meter sub-directory.

Running the Application

The application execution command line is as below:

./qos_meter [EAL options] -- -p PORTMASK

The application is constrained to use a single core in the EAL core mask and 2 ports only in the application port mask (first port from the port mask is used for RX and the other port in the core mask is used for TX).

Refer to DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

Selecting one of the metering modes is done with these defines:

#define APP_MODE_FWD   0
#define APP_MODE_SRTCM_COLOR_BLIND  1
#define APP_MODE_SRTCM_COLOR_AWARE  2
#define APP_MODE_TRTCM_COLOR_BLIND  3
#define APP_MODE_TRTCM_COLOR_AWARE  4

#define APP_MODE  APP_MODE_SRTCM_COLOR_BLIND

To simplify debugging (for example, by using the traffic generator RX side MAC address based packet filtering feature), the color is defined as the LSB byte of the destination MAC address.

The traffic meter parameters are configured in the application source code with following default values:

struct rte_meter_srtcm_params app_srtcm_params[] = {

    {.cir = 1000000 * 46, .cbs = 2048, .ebs = 2048},

};

struct rte_meter_trtcm_params app_trtcm_params[] = {

    {.cir = 1000000 * 46, .pir = 1500000 * 46, .cbs = 2048, .pbs = 2048},

};

Assuming the input traffic is generated at line rate and all packets are 64 bytes Ethernet frames (IPv4 packet size of 46 bytes) and green, the expected output traffic should be marked as shown in the following table:

Output Traffic Marking
Mode Green (Mpps) Yellow (Mpps) Red (Mpps)
srTCM blind 1 1 12.88
srTCM color 1 1 12.88
trTCM blind 1 0.5 13.38
trTCM color 1 0.5 13.38
FWD 14.88 0 0

To set up the policing scheme as desired, it is necessary to modify the main.h source file, where this policy is implemented as a static structure, as follows:

int policer_table[e_RTE_METER_COLORS][e_RTE_METER_COLORS] =
{
   { GREEN, RED, RED},
   { DROP, YELLOW, RED},
   { DROP, DROP, RED}
};

Where rows indicate the input color, columns indicate the output color, and the value that is stored in the table indicates the action to be taken for that particular case.

There are four different actions:

  • GREEN: The packet’s color is changed to green.
  • YELLOW: The packet’s color is changed to yellow.
  • RED: The packet’s color is changed to red.
  • DROP: The packet is dropped.

In this particular case:

  • Every packet which input and output color are the same, keeps the same color.
  • Every packet which color has improved is dropped (this particular case can’t happen, so these values will not be used).
  • For the rest of the cases, the color is changed to red.

QoS Scheduler Sample Application

The QoS sample application demonstrates the use of the DPDK to provide QoS scheduling.

Overview

The architecture of the QoS scheduler application is shown in the following figure.

_images/qos_sched_app_arch.png

QoS Scheduler Application Architecture

There are two flavors of the runtime execution for this application, with two or three threads per each packet flow configuration being used. The RX thread reads packets from the RX port, classifies the packets based on the double VLAN (outer and inner) and the lower two bytes of the IP destination address and puts them into the ring queue. The worker thread dequeues the packets from the ring and calls the QoS scheduler enqueue/dequeue functions. If a separate TX core is used, these are sent to the TX ring. Otherwise, they are sent directly to the TX port. The TX thread, if present, reads from the TX ring and write the packets to the TX port.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the qos_sched sub-directory.

Note

This application is intended as a linuxapp only.

Note

To get statistics on the sample app using the command line interface as described in the next section, DPDK must be compiled defining CONFIG_RTE_SCHED_COLLECT_STATS, which can be done by changing the configuration file for the specific target to be compiled.

Running the Application

Note

In order to run the application, a total of at least 4 G of huge pages must be set up for each of the used sockets (depending on the cores in use).

The application has a number of command line options:

./qos_sched [EAL options] -- <APP PARAMS>

Mandatory application parameters include:

  • –pfc “RX PORT, TX PORT, RX LCORE, WT LCORE, TX CORE”: Packet flow configuration. Multiple pfc entities can be configured in the command line, having 4 or 5 items (if TX core defined or not).

Optional application parameters include:

  • -i: It makes the application to start in the interactive mode. In this mode, the application shows a command line that can be used for obtaining statistics while scheduling is taking place (see interactive mode below for more information).
  • –mst n: Master core index (the default value is 1).
  • –rsz “A, B, C”: Ring sizes:
  • A = Size (in number of buffer descriptors) of each of the NIC RX rings read by the I/O RX lcores (the default value is 128).
  • B = Size (in number of elements) of each of the software rings used by the I/O RX lcores to send packets to worker lcores (the default value is 8192).
  • C = Size (in number of buffer descriptors) of each of the NIC TX rings written by worker lcores (the default value is 256)
  • –bsz “A, B, C, D”: Burst sizes
  • A = I/O RX lcore read burst size from the NIC RX (the default value is 64)
  • B = I/O RX lcore write burst size to the output software rings, worker lcore read burst size from input software rings,QoS enqueue size (the default value is 64)
  • C = QoS dequeue size (the default value is 32)
  • D = Worker lcore write burst size to the NIC TX (the default value is 64)
  • –msz M: Mempool size (in number of mbufs) for each pfc (default 2097152)
  • –rth “A, B, C”: The RX queue threshold parameters
  • A = RX prefetch threshold (the default value is 8)
  • B = RX host threshold (the default value is 8)
  • C = RX write-back threshold (the default value is 4)
  • –tth “A, B, C”: TX queue threshold parameters
  • A = TX prefetch threshold (the default value is 36)
  • B = TX host threshold (the default value is 0)
  • C = TX write-back threshold (the default value is 0)
  • –cfg FILE: Profile configuration to load

Refer to DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

The profile configuration file defines all the port/subport/pipe/traffic class/queue parameters needed for the QoS scheduler configuration.

The profile file has the following format:

; port configuration [port]

frame overhead = 24
number of subports per port = 1
number of pipes per subport = 4096
queue sizes = 64 64 64 64

; Subport configuration

[subport 0]
tb rate = 1250000000; Bytes per second
tb size = 1000000; Bytes
tc 0 rate = 1250000000;     Bytes per second
tc 1 rate = 1250000000;     Bytes per second
tc 2 rate = 1250000000;     Bytes per second
tc 3 rate = 1250000000;     Bytes per second
tc period = 10;             Milliseconds
tc oversubscription period = 10;     Milliseconds

pipe 0-4095 = 0;        These pipes are configured with pipe profile 0

; Pipe configuration

[pipe profile 0]
tb rate = 305175; Bytes per second
tb size = 1000000; Bytes

tc 0 rate = 305175; Bytes per second
tc 1 rate = 305175; Bytes per second
tc 2 rate = 305175; Bytes per second
tc 3 rate = 305175; Bytes per second
tc period = 40; Milliseconds

tc 0 oversubscription weight = 1
tc 1 oversubscription weight = 1
tc 2 oversubscription weight = 1
tc 3 oversubscription weight = 1

tc 0 wrr weights = 1 1 1 1
tc 1 wrr weights = 1 1 1 1
tc 2 wrr weights = 1 1 1 1
tc 3 wrr weights = 1 1 1 1

; RED params per traffic class and color (Green / Yellow / Red)

[red]
tc 0 wred min = 48 40 32
tc 0 wred max = 64 64 64
tc 0 wred inv prob = 10 10 10
tc 0 wred weight = 9 9 9

tc 1 wred min = 48 40 32
tc 1 wred max = 64 64 64
tc 1 wred inv prob = 10 10 10
tc 1 wred weight = 9 9 9

tc 2 wred min = 48 40 32
tc 2 wred max = 64 64 64
tc 2 wred inv prob = 10 10 10
tc 2 wred weight = 9 9 9

tc 3 wred min = 48 40 32
tc 3 wred max = 64 64 64
tc 3 wred inv prob = 10 10 10
tc 3 wred weight = 9 9 9
Interactive mode

These are the commands that are currently working under the command line interface:

  • Control Commands
  • –quit: Quits the application.
  • General Statistics
    • stats app: Shows a table with in-app calculated statistics.
    • stats port X subport Y: For a specific subport, it shows the number of packets that went through the scheduler properly and the number of packets that were dropped. The same information is shown in bytes. The information is displayed in a table separating it in different traffic classes.
    • stats port X subport Y pipe Z: For a specific pipe, it shows the number of packets that went through the scheduler properly and the number of packets that were dropped. The same information is shown in bytes. This information is displayed in a table separating it in individual queues.
  • Average queue size

All of these commands work the same way, averaging the number of packets throughout a specific subset of queues.

Two parameters can be configured for this prior to calling any of these commands:

  • qavg n X: n is the number of times that the calculation will take place. Bigger numbers provide higher accuracy. The default value is 10.
  • qavg period X: period is the number of microseconds that will be allowed between each calculation. The default value is 100.

The commands that can be used for measuring average queue size are:

  • qavg port X subport Y: Show average queue size per subport.
  • qavg port X subport Y tc Z: Show average queue size per subport for a specific traffic class.
  • qavg port X subport Y pipe Z: Show average queue size per pipe.
  • qavg port X subport Y pipe Z tc A: Show average queue size per pipe for a specific traffic class.
  • qavg port X subport Y pipe Z tc A q B: Show average queue size of a specific queue.
Example

The following is an example command with a single packet flow configuration:

./qos_sched -l 1,5,7 -n 4 -- --pfc "3,2,5,7" --cfg ./profile.cfg

This example uses a single packet flow configuration which creates one RX thread on lcore 5 reading from port 3 and a worker thread on lcore 7 writing to port 2.

Another example with 2 packet flow configurations using different ports but sharing the same core for QoS scheduler is given below:

./qos_sched -l 1,2,6,7 -n 4 -- --pfc "3,2,2,6,7" --pfc "1,0,2,6,7" --cfg ./profile.cfg

Note that independent cores for the packet flow configurations for each of the RX, WT and TX thread are also supported, providing flexibility to balance the work.

The EAL coremask/corelist is constrained to contain the default mastercore 1 and the RX, WT and TX cores only.

Explanation

The Port/Subport/Pipe/Traffic Class/Queue are the hierarchical entities in a typical QoS application:

  • A subport represents a predefined group of users.
  • A pipe represents an individual user/subscriber.
  • A traffic class is the representation of a different traffic type with a specific loss rate, delay and jitter requirements; such as data voice, video or data transfers.
  • A queue hosts packets from one or multiple connections of the same type belonging to the same user.

The traffic flows that need to be configured are application dependent. This application classifies based on the QinQ double VLAN tags and the IP destination address as indicated in the following table.

Entity Types
Level Name Siblings per Parent QoS Functional Description Selected By
Port
Ethernet port Physical port
Subport Config (8) Traffic shaped (token bucket) Outer VLAN tag
Pipe Config (4k) Traffic shaped (token bucket) Inner VLAN tag
Traffic Class 4 TCs of the same pipe services in strict priority Destination IP address (0.0.X.0)
Queue 4 Queue of the same TC serviced in WRR Destination IP address (0.0.0.X)

Please refer to the “QoS Scheduler” chapter in the DPDK Programmer’s Guide for more information about these parameters.

Quota and Watermark Sample Application

The Quota and Watermark sample application is a simple example of packet processing using Data Plane Development Kit (DPDK) that showcases the use of a quota as the maximum number of packets enqueue/dequeue at a time and low and high thresholds, or watermarks, to signal low and high ring usage respectively.

Additionally, it shows how the thresholds can be used to feedback congestion notifications to data producers by temporarily stopping processing overloaded rings and sending Ethernet flow control frames.

This sample application is split in two parts:

  • qw - The core quota and watermark sample application
  • qwctl - A command line tool to alter quota and watermarks while qw is running

Overview

The Quota and Watermark sample application performs forwarding for each packet that is received on a given port. The destination port is the adjacent port from the enabled port mask, that is, if the first four ports are enabled (port mask 0xf), ports 0 and 1 forward into each other, and ports 2 and 3 forward into each other. The MAC addresses of the forwarded Ethernet frames are not affected.

Internally, packets are pulled from the ports by the master logical core and put on a variable length processing pipeline, each stage of which being connected by rings, as shown in Fig. 31.1.

_images/pipeline_overview.png

Pipeline Overview

An adjustable quota value controls how many packets are being moved through the pipeline per enqueue and dequeue. Adjustable threshold values associated with the rings control a back-off mechanism that tries to prevent the pipeline from being overloaded by:

  • Stopping enqueuing on rings for which the usage has crossed the high watermark threshold
  • Sending Ethernet pause frames
  • Only resuming enqueuing on a ring once its usage goes below a global low watermark threshold

This mechanism allows congestion notifications to go up the ring pipeline and eventually lead to an Ethernet flow control frame being send to the source.

On top of serving as an example of quota and watermark usage, this application can be used to benchmark ring based processing pipelines performance using a traffic- generator, as shown in Fig. 31.2.

_images/ring_pipeline_perf_setup.png

Ring-based Processing Pipeline Performance Setup

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the quota_watermark sub-directory.

Running the Application

The core application, qw, has to be started first.

Once it is up and running, one can alter quota and watermarks while it runs using the control application, qwctl.

Running the Core Application

The application requires a single command line option:

./qw/build/qw [EAL options] -- -p PORTMASK

where,

-p PORTMASK: A hexadecimal bitmask of the ports to configure

To run the application in a linuxapp environment with four logical cores and ports 0 and 2, issue the following command:

./qw/build/qw -l 0-3 -n 4 -- -p 5

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Running the Control Application

The control application requires a number of command line options:

./qwctl/build/qwctl [EAL options] --proc-type=secondary

The –proc-type=secondary option is necessary for the EAL to properly initialize the control application to use the same huge pages as the core application and thus be able to access its rings.

To run the application in a linuxapp environment on logical core 0, issue the following command:

./qwctl/build/qwctl -l 0 -n 4 --proc-type=secondary

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

qwctl is an interactive command line that let the user change variables in a running instance of qw. The help command gives a list of available commands:

$ qwctl > help

Code Overview

The following sections provide a quick guide to the application’s source code.

Core Application - qw
EAL and Drivers Setup

The EAL arguments are parsed at the beginning of the main() function:

ret = rte_eal_init(argc, argv);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Cannot initialize EAL\n");

argc -= ret;
argv += ret;

Then, a call to init_dpdk(), defined in init.c, is made to initialize the poll mode drivers:

void
init_dpdk(void)
{
    int ret;

    /* Bind the drivers to usable devices */

    ret = rte_pci_probe();
    if (ret < 0)
        rte_exit(EXIT_FAILURE, "rte_pci_probe(): error %d\n", ret);

    if (rte_eth_dev_count() < 2)
        rte_exit(EXIT_FAILURE, "Not enough Ethernet port available\n");
}

To fully understand this code, it is recommended to study the chapters that relate to the Poll Mode Driver in the DPDK Getting Started Guide and the DPDK API Reference.

Shared Variables Setup

The quota and high and low watermark shared variables are put into an rte_memzone using a call to setup_shared_variables():

void
setup_shared_variables(void)
{
       const struct rte_memzone *qw_memzone;

       qw_memzone = rte_memzone_reserve(QUOTA_WATERMARK_MEMZONE_NAME,
                      3 * sizeof(int), rte_socket_id(), 0);
       if (qw_memzone == NULL)
               rte_exit(EXIT_FAILURE, "%s\n", rte_strerror(rte_errno));

       quota = qw_memzone->addr;
       low_watermark = (unsigned int *) qw_memzone->addr + 1;
       high_watermark = (unsigned int *) qw_memzone->addr + 2;
}

These three variables are initialized to a default value in main() and can be changed while qw is running using the qwctl control program.

Application Arguments

The qw application only takes one argument: a port mask that specifies which ports should be used by the application. At least two ports are needed to run the application and there should be an even number of ports given in the port mask.

The port mask parsing is done in parse_qw_args(), defined in args.c.

Mbuf Pool Initialization

Once the application’s arguments are parsed, an mbuf pool is created. It contains a set of mbuf objects that are used by the driver and the application to store network packets:

/* Create a pool of mbuf to store packets */
mbuf_pool = rte_pktmbuf_pool_create("mbuf_pool", MBUF_PER_POOL, 32, 0,
                                     MBUF_DATA_SIZE, rte_socket_id());

if (mbuf_pool == NULL)
    rte_panic("%s\n", rte_strerror(rte_errno));

The rte_mempool is a generic structure used to handle pools of objects. In this case, it is necessary to create a pool that will be used by the driver.

The number of allocated pkt mbufs is MBUF_PER_POOL, with a data room size of MBUF_DATA_SIZE each. A per-lcore cache of 32 mbufs is kept. The memory is allocated in on the master lcore’s socket, but it is possible to extend this code to allocate one mbuf pool per socket.

The rte_pktmbuf_pool_create() function uses the default mbuf pool and mbuf initializers, respectively rte_pktmbuf_pool_init() and rte_pktmbuf_init(). An advanced application may want to use the mempool API to create the mbuf pool with more control.

Ports Configuration and Pairing

Each port in the port mask is configured and a corresponding ring is created in the master lcore’s array of rings. This ring is the first in the pipeline and will hold the packets directly coming from the port.

for (port_id = 0; port_id < RTE_MAX_ETHPORTS; port_id++)
    if (is_bit_set(port_id, portmask)) {
        configure_eth_port(port_id);
        init_ring(master_lcore_id, port_id);
    }

pair_ports();

The configure_eth_port() and init_ring() functions are used to configure a port and a ring respectively and are defined in init.c. They make use of the DPDK APIs defined in rte_eth.h and rte_ring.h.

pair_ports() builds the port_pairs[] array so that its key-value pairs are a mapping between reception and transmission ports. It is defined in init.c.

Logical Cores Assignment

The application uses the master logical core to poll all the ports for new packets and enqueue them on a ring associated with the port.

Each logical core except the last runs pipeline_stage() after a ring for each used port is initialized on that core. pipeline_stage() on core X dequeues packets from core X-1’s rings and enqueue them on its own rings. See Fig. 31.3.

/* Start pipeline_stage() on all the available slave lcore but the last */

for (lcore_id = 0 ; lcore_id < last_lcore_id; lcore_id++) {
    if (rte_lcore_is_enabled(lcore_id) && lcore_id != master_lcore_id) {
        for (port_id = 0; port_id < RTE_MAX_ETHPORTS; port_id++)
            if (is_bit_set(port_id, portmask))
                init_ring(lcore_id, port_id);

            rte_eal_remote_launch(pipeline_stage, NULL, lcore_id);
    }
}

The last available logical core runs send_stage(), which is the last stage of the pipeline dequeuing packets from the last ring in the pipeline and sending them out on the destination port setup by pair_ports().

/* Start send_stage() on the last slave core */

rte_eal_remote_launch(send_stage, NULL, last_lcore_id);
Receive, Process and Transmit Packets
_images/threads_pipelines.png

Threads and Pipelines

In the receive_stage() function running on the master logical core, the main task is to read ingress packets from the RX ports and enqueue them on the port’s corresponding first ring in the pipeline. This is done using the following code:

lcore_id = rte_lcore_id();

/* Process each port round robin style */

for (port_id = 0; port_id < RTE_MAX_ETHPORTS; port_id++) {
        if (!is_bit_set(port_id, portmask))
                continue;

        ring = rings[lcore_id][port_id];

        if (ring_state[port_id] != RING_READY) {
                if (rte_ring_count(ring) > *low_watermark)
                        continue;
                else
                        ring_state[port_id] = RING_READY;
        }

        /* Enqueue received packets on the RX ring */
        nb_rx_pkts = rte_eth_rx_burst(port_id, 0, pkts,
                        (uint16_t) *quota);
        ret = rte_ring_enqueue_bulk(ring, (void *) pkts,
                        nb_rx_pkts, &free);
        if (RING_SIZE - free > *high_watermark) {
                ring_state[port_id] = RING_OVERLOADED;
                send_pause_frame(port_id, 1337);
        }

        if (ret == 0) {

                /*
                 * Return  mbufs to the pool,
                 * effectively dropping packets
                 */
                for (i = 0; i < nb_rx_pkts; i++)
                        rte_pktmbuf_free(pkts[i]);
        }
}

For each port in the port mask, the corresponding ring’s pointer is fetched into ring and that ring’s state is checked:

  • If it is in the RING_READY state, *quota packets are grabbed from the port and put on the ring. Should this operation make the ring’s usage cross its high watermark, the ring is marked as overloaded and an Ethernet flow control frame is sent to the source.
  • If it is not in the RING_READY state, this port is ignored until the ring’s usage crosses the *low_watermark value.

The pipeline_stage() function’s task is to process and move packets from the preceding pipeline stage. This thread is running on most of the logical cores to create and arbitrarily long pipeline.

lcore_id = rte_lcore_id();

previous_lcore_id = get_previous_lcore_id(lcore_id);

for (port_id = 0; port_id < RTE_MAX_ETHPORTS; port_id++) {
        if (!is_bit_set(port_id, portmask))
                continue;

        tx = rings[lcore_id][port_id];
        rx = rings[previous_lcore_id][port_id];

        if (ring_state[port_id] != RING_READY) {
                if (rte_ring_count(tx) > *low_watermark)
                        continue;
                else
                        ring_state[port_id] = RING_READY;
        }

        /* Dequeue up to quota mbuf from rx */
        nb_dq_pkts = rte_ring_dequeue_burst(rx, pkts,
                        *quota, NULL);
        if (unlikely(nb_dq_pkts < 0))
                continue;

        /* Enqueue them on tx */
        ret = rte_ring_enqueue_bulk(tx, pkts,
                        nb_dq_pkts, &free);
        if (RING_SIZE - free > *high_watermark)
                ring_state[port_id] = RING_OVERLOADED;

        if (ret == 0) {

                /*
                 * Return  mbufs to the pool,
                 * effectively dropping packets
                 */
                for (i = 0; i < nb_dq_pkts; i++)
                        rte_pktmbuf_free(pkts[i]);
        }
}

The thread’s logic works mostly like receive_stage(), except that packets are moved from ring to ring instead of port to ring.

In this example, no actual processing is done on the packets, but pipeline_stage() is an ideal place to perform any processing required by the application.

Finally, the send_stage() function’s task is to read packets from the last ring in a pipeline and send them on the destination port defined in the port_pairs[] array. It is running on the last available logical core only.

lcore_id = rte_lcore_id();

previous_lcore_id = get_previous_lcore_id(lcore_id);

for (port_id = 0; port_id < RTE_MAX_ETHPORTS; port_id++) {
    if (!is_bit_set(port_id, portmask)) continue;

    dest_port_id = port_pairs[port_id];
    tx = rings[previous_lcore_id][port_id];

    if (rte_ring_empty(tx)) continue;

    /* Dequeue packets from tx and send them */

    nb_dq_pkts = rte_ring_dequeue_burst(tx, (void *) tx_pkts, *quota);
    nb_tx_pkts = rte_eth_tx_burst(dest_port_id, 0, tx_pkts, nb_dq_pkts);
}

For each port in the port mask, up to *quota packets are pulled from the last ring in its pipeline and sent on the destination port paired with the current port.

Control Application - qwctl

The qwctl application uses the rte_cmdline library to provide the user with an interactive command line that can be used to modify and inspect parameters in a running qw application. Those parameters are the global quota and low_watermark value as well as each ring’s built-in high watermark.

Command Definitions

The available commands are defined in commands.c.

It is advised to use the cmdline sample application user guide as a reference for everything related to the rte_cmdline library.

Accessing Shared Variables

The setup_shared_variables() function retrieves the shared variables quota and low_watermark from the rte_memzone previously created by qw.

static void
setup_shared_variables(void)
{
    const struct rte_memzone *qw_memzone;

    qw_memzone = rte_memzone_lookup(QUOTA_WATERMARK_MEMZONE_NAME);
    if (qw_memzone == NULL)
        rte_exit(EXIT_FAILURE, "Couldn't find memzone\n");

    quota = qw_memzone->addr;

    low_watermark = (unsigned int *) qw_memzone->addr + 1;
    high_watermark = (unsigned int *) qw_memzone->addr + 2;
}

Timer Sample Application

The Timer sample application is a simple application that demonstrates the use of a timer in a DPDK application. This application prints some messages from different lcores regularly, demonstrating the use of timers.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the timer sub-directory.

Running the Application

To run the example in linuxapp environment:

$ ./build/timer -l 0-3 -n 4

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The following sections provide some explanation of the code.

Initialization and Main Loop

In addition to EAL initialization, the timer subsystem must be initialized, by calling the rte_timer_subsystem_init() function.

/* init EAL */

ret = rte_eal_init(argc, argv);
if (ret < 0)
    rte_panic("Cannot init EAL\n");

/* init RTE timer library */

rte_timer_subsystem_init();

After timer creation (see the next paragraph), the main loop is executed on each slave lcore using the well-known rte_eal_remote_launch() and also on the master.

/* call lcore_mainloop() on every slave lcore  */

RTE_LCORE_FOREACH_SLAVE(lcore_id) {
    rte_eal_remote_launch(lcore_mainloop, NULL, lcore_id);
}

/* call it on master lcore too */

(void) lcore_mainloop(NULL);

The main loop is very simple in this example:

while (1) {
    /*
     *   Call the timer handler on each core: as we don't
     *   need a very precise timer, so only call
     *   rte_timer_manage() every ~10ms (at 2 GHz). In a real
     *   application, this will enhance performances as
     *   reading the HPET timer is not efficient.
    */

    cur_tsc = rte_rdtsc();

    diff_tsc = cur_tsc - prev_tsc;

    if (diff_tsc > TIMER_RESOLUTION_CYCLES) {
        rte_timer_manage();
        prev_tsc = cur_tsc;
    }
}

As explained in the comment, it is better to use the TSC register (as it is a per-lcore register) to check if the rte_timer_manage() function must be called or not. In this example, the resolution of the timer is 10 milliseconds.

Managing Timers

In the main() function, the two timers are initialized. This call to rte_timer_init() is necessary before doing any other operation on the timer structure.

/* init timer structures */

rte_timer_init(&timer0);
rte_timer_init(&timer1);

Then, the two timers are configured:

  • The first timer (timer0) is loaded on the master lcore and expires every second. Since the PERIODICAL flag is provided, the timer is reloaded automatically by the timer subsystem. The callback function is timer0_cb().
  • The second timer (timer1) is loaded on the next available lcore every 333 ms. The SINGLE flag means that the timer expires only once and must be reloaded manually if required. The callback function is timer1_cb().
/* load timer0, every second, on master lcore, reloaded automatically */

hz = rte_get_hpet_hz();

lcore_id = rte_lcore_id();

rte_timer_reset(&timer0, hz, PERIODICAL, lcore_id, timer0_cb, NULL);

/* load timer1, every second/3, on next lcore, reloaded manually */

lcore_id = rte_get_next_lcore(lcore_id, 0, 1);

rte_timer_reset(&timer1, hz/3, SINGLE, lcore_id, timer1_cb, NULL);

The callback for the first timer (timer0) only displays a message until a global counter reaches 20 (after 20 seconds). In this case, the timer is stopped using the rte_timer_stop() function.

/* timer0 callback */

static void
timer0_cb( attribute ((unused)) struct rte_timer *tim, __attribute ((unused)) void *arg)
{
    static unsigned counter = 0;

    unsigned lcore_id = rte_lcore_id();

    printf("%s() on lcore %u\n", FUNCTION , lcore_id);

    /* this timer is automatically reloaded until we decide to stop it, when counter reaches 20. */

    if ((counter ++) == 20)
        rte_timer_stop(tim);
}

The callback for the second timer (timer1) displays a message and reloads the timer on the next lcore, using the rte_timer_reset() function:

/* timer1 callback */

static void
timer1_cb( attribute ((unused)) struct rte_timer *tim, _attribute ((unused)) void *arg)
{
    unsigned lcore_id = rte_lcore_id();
    uint64_t hz;

    printf("%s() on lcore %u\\n", FUNCTION , lcore_id);

    /* reload it on another lcore */

    hz = rte_get_hpet_hz();

    lcore_id = rte_get_next_lcore(lcore_id, 0, 1);

    rte_timer_reset(&timer1, hz/3, SINGLE, lcore_id, timer1_cb, NULL);
}

Packet Ordering Application

The Packet Ordering sample app simply shows the impact of reordering a stream. It’s meant to stress the library with different configurations for performance.

Overview

The application uses at least three CPU cores:

  • RX core (maser core) receives traffic from the NIC ports and feeds Worker cores with traffic through SW queues.
  • Worker core (slave core) basically do some light work on the packet. Currently it modifies the output port of the packet for configurations with more than one port enabled.
  • TX Core (slave core) receives traffic from Worker cores through software queues, inserts out-of-order packets into reorder buffer, extracts ordered packets from the reorder buffer and sends them to the NIC ports for transmission.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the packet_ordering sub-directory.

Running the Application

Refer to DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Application Command Line

The application execution command line is:

./test-pipeline [EAL options] -- -p PORTMASK [--disable-reorder]

The -c EAL CPU_COREMASK option has to contain at least 3 CPU cores. The first CPU core in the core mask is the master core and would be assigned to RX core, the last to TX core and the rest to Worker cores.

The PORTMASK parameter must contain either 1 or even enabled port numbers. When setting more than 1 port, traffic would be forwarded in pairs. For example, if we enable 4 ports, traffic from port 0 to 1 and from 1 to 0, then the other pair from 2 to 3 and from 3 to 2, having [0,1] and [2,3] pairs.

The disable-reorder long option does, as its name implies, disable the reordering of traffic, which should help evaluate reordering performance impact.

VMDQ and DCB Forwarding Sample Application

The VMDQ and DCB Forwarding sample application is a simple example of packet processing using the DPDK. The application performs L2 forwarding using VMDQ and DCB to divide the incoming traffic into queues. The traffic splitting is performed in hardware by the VMDQ and DCB features of the Intel® 82599 and X710/XL710 Ethernet Controllers.

Overview

This sample application can be used as a starting point for developing a new application that is based on the DPDK and uses VMDQ and DCB for traffic partitioning.

The VMDQ and DCB filters work on MAC and VLAN traffic to divide the traffic into input queues on the basis of the Destination MAC address, VLAN ID and VLAN user priority fields. VMDQ filters split the traffic into 16 or 32 groups based on the Destination MAC and VLAN ID. Then, DCB places each packet into one of queues within that group, based upon the VLAN user priority field.

All traffic is read from a single incoming port (port 0) and output on port 1, without any processing being performed. With Intel® 82599 NIC, for example, the traffic is split into 128 queues on input, where each thread of the application reads from multiple queues. When run with 8 threads, that is, with the -c FF option, each thread receives and forwards packets from 16 queues.

As supplied, the sample application configures the VMDQ feature to have 32 pools with 4 queues each as indicated in Fig. 34.1. The Intel® 82599 10 Gigabit Ethernet Controller NIC also supports the splitting of traffic into 16 pools of 8 queues. While the Intel® X710 or XL710 Ethernet Controller NICs support many configurations of VMDQ pools of 4 or 8 queues each. For simplicity, only 16 or 32 pools is supported in this sample. And queues numbers for each VMDQ pool can be changed by setting CONFIG_RTE_LIBRTE_I40E_QUEUE_NUM_PER_VM in config/common_* file. The nb-pools, nb-tcs and enable-rss parameters can be passed on the command line, after the EAL parameters:

./build/vmdq_dcb [EAL options] -- -p PORTMASK --nb-pools NP --nb-tcs TC --enable-rss

where, NP can be 16 or 32, TC can be 4 or 8, rss is disabled by default.

_images/vmdq_dcb_example.svg

Packet Flow Through the VMDQ and DCB Sample Application

In Linux* user space, the application can display statistics with the number of packets received on each queue. To have the application display the statistics, send a SIGHUP signal to the running application process.

The VMDQ and DCB Forwarding sample application is in many ways simpler than the L2 Forwarding application (see L2 Forwarding Sample Application (in Real and Virtualized Environments)) as it performs unidirectional L2 forwarding of packets from one port to a second port. No command-line options are taken by this application apart from the standard EAL command-line options.

Note

Since VMD queues are being used for VMM, this application works correctly when VTd is disabled in the BIOS or Linux* kernel (intel_iommu=off).

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the vmdq_dcb sub-directory.

Running the Application

To run the example in a linuxapp environment:

user@target:~$ ./build/vmdq_dcb -l 0-3 -n 4 -- -p 0x3 --nb-pools 32 --nb-tcs 4

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The following sections provide some explanation of the code.

Initialization

The EAL, driver and PCI configuration is performed largely as in the L2 Forwarding sample application, as is the creation of the mbuf pool. See L2 Forwarding Sample Application (in Real and Virtualized Environments). Where this example application differs is in the configuration of the NIC port for RX.

The VMDQ and DCB hardware feature is configured at port initialization time by setting the appropriate values in the rte_eth_conf structure passed to the rte_eth_dev_configure() API. Initially in the application, a default structure is provided for VMDQ and DCB configuration to be filled in later by the application.

/* empty vmdq+dcb configuration structure. Filled in programmatically */
static const struct rte_eth_conf vmdq_dcb_conf_default = {
    .rxmode = {
        .mq_mode        = ETH_MQ_RX_VMDQ_DCB,
        .split_hdr_size = 0,
        .header_split   = 0, /**< Header Split disabled */
        .hw_ip_checksum = 0, /**< IP checksum offload disabled */
        .hw_vlan_filter = 0, /**< VLAN filtering disabled */
        .jumbo_frame    = 0, /**< Jumbo Frame Support disabled */
    },
    .txmode = {
        .mq_mode = ETH_MQ_TX_VMDQ_DCB,
    },
    /*
     * should be overridden separately in code with
     * appropriate values
     */
    .rx_adv_conf = {
        .vmdq_dcb_conf = {
            .nb_queue_pools = ETH_32_POOLS,
            .enable_default_pool = 0,
            .default_pool = 0,
            .nb_pool_maps = 0,
            .pool_map = {{0, 0},},
            .dcb_tc = {0},
        },
        .dcb_rx_conf = {
            .nb_tcs = ETH_4_TCS,
            /** Traffic class each UP mapped to. */
            .dcb_tc = {0},
        },
        .vmdq_rx_conf = {
            .nb_queue_pools = ETH_32_POOLS,
            .enable_default_pool = 0,
            .default_pool = 0,
            .nb_pool_maps = 0,
            .pool_map = {{0, 0},},
        },
    },
    .tx_adv_conf = {
        .vmdq_dcb_tx_conf = {
            .nb_queue_pools = ETH_32_POOLS,
            .dcb_tc = {0},
        },
    },
};

The get_eth_conf() function fills in an rte_eth_conf structure with the appropriate values, based on the global vlan_tags array, and dividing up the possible user priority values equally among the individual queues (also referred to as traffic classes) within each pool. With Intel® 82599 NIC, if the number of pools is 32, then the user priority fields are allocated 2 to a queue. If 16 pools are used, then each of the 8 user priority fields is allocated to its own queue within the pool. With Intel® X710/XL710 NICs, if number of tcs is 4, and number of queues in pool is 8, then the user priority fields are allocated 2 to one tc, and a tc has 2 queues mapping to it, then RSS will determine the destination queue in 2. For the VLAN IDs, each one can be allocated to possibly multiple pools of queues, so the pools parameter in the rte_eth_vmdq_dcb_conf structure is specified as a bitmask value. For destination MAC, each VMDQ pool will be assigned with a MAC address. In this sample, each VMDQ pool is assigned to the MAC like 52:54:00:12:<port_id>:<pool_id>, that is, the MAC of VMDQ pool 2 on port 1 is 52:54:00:12:01:02.

const uint16_t vlan_tags[] = {
    0, 1, 2, 3, 4, 5, 6, 7,
    8, 9, 10, 11, 12, 13, 14, 15,
    16, 17, 18, 19, 20, 21, 22, 23,
    24, 25, 26, 27, 28, 29, 30, 31
};

/* pool mac addr template, pool mac addr is like: 52 54 00 12 port# pool# */
static struct ether_addr pool_addr_template = {
    .addr_bytes = {0x52, 0x54, 0x00, 0x12, 0x00, 0x00}
};

/* Builds up the correct configuration for vmdq+dcb based on the vlan tags array
 * given above, and the number of traffic classes available for use. */
static inline int
get_eth_conf(struct rte_eth_conf *eth_conf)
{
    struct rte_eth_vmdq_dcb_conf conf;
    struct rte_eth_vmdq_rx_conf  vmdq_conf;
    struct rte_eth_dcb_rx_conf   dcb_conf;
    struct rte_eth_vmdq_dcb_tx_conf tx_conf;
    uint8_t i;

    conf.nb_queue_pools = (enum rte_eth_nb_pools)num_pools;
    vmdq_conf.nb_queue_pools = (enum rte_eth_nb_pools)num_pools;
    tx_conf.nb_queue_pools = (enum rte_eth_nb_pools)num_pools;
    conf.nb_pool_maps = num_pools;
    vmdq_conf.nb_pool_maps = num_pools;
    conf.enable_default_pool = 0;
    vmdq_conf.enable_default_pool = 0;
    conf.default_pool = 0; /* set explicit value, even if not used */
    vmdq_conf.default_pool = 0;

    for (i = 0; i < conf.nb_pool_maps; i++) {
        conf.pool_map[i].vlan_id = vlan_tags[i];
        vmdq_conf.pool_map[i].vlan_id = vlan_tags[i];
        conf.pool_map[i].pools = 1UL << i ;
        vmdq_conf.pool_map[i].pools = 1UL << i;
    }
    for (i = 0; i < ETH_DCB_NUM_USER_PRIORITIES; i++){
        conf.dcb_tc[i] = i % num_tcs;
        dcb_conf.dcb_tc[i] = i % num_tcs;
        tx_conf.dcb_tc[i] = i % num_tcs;
    }
    dcb_conf.nb_tcs = (enum rte_eth_nb_tcs)num_tcs;
    (void)(rte_memcpy(eth_conf, &vmdq_dcb_conf_default, sizeof(*eth_conf)));
    (void)(rte_memcpy(&eth_conf->rx_adv_conf.vmdq_dcb_conf, &conf,
              sizeof(conf)));
    (void)(rte_memcpy(&eth_conf->rx_adv_conf.dcb_rx_conf, &dcb_conf,
              sizeof(dcb_conf)));
    (void)(rte_memcpy(&eth_conf->rx_adv_conf.vmdq_rx_conf, &vmdq_conf,
              sizeof(vmdq_conf)));
    (void)(rte_memcpy(&eth_conf->tx_adv_conf.vmdq_dcb_tx_conf, &tx_conf,
              sizeof(tx_conf)));
    if (rss_enable) {
        eth_conf->rxmode.mq_mode= ETH_MQ_RX_VMDQ_DCB_RSS;
        eth_conf->rx_adv_conf.rss_conf.rss_hf = ETH_RSS_IP |
                            ETH_RSS_UDP |
                            ETH_RSS_TCP |
                            ETH_RSS_SCTP;
    }
    return 0;
}

......

/* Set mac for each pool.*/
for (q = 0; q < num_pools; q++) {
    struct ether_addr mac;
    mac = pool_addr_template;
    mac.addr_bytes[4] = port;
    mac.addr_bytes[5] = q;
    printf("Port %u vmdq pool %u set mac %02x:%02x:%02x:%02x:%02x:%02x\n",
        port, q,
        mac.addr_bytes[0], mac.addr_bytes[1],
        mac.addr_bytes[2], mac.addr_bytes[3],
        mac.addr_bytes[4], mac.addr_bytes[5]);
    retval = rte_eth_dev_mac_addr_add(port, &mac,
            q + vmdq_pool_base);
    if (retval) {
        printf("mac addr add failed at pool %d\n", q);
        return retval;
    }
}

Once the network port has been initialized using the correct VMDQ and DCB values, the initialization of the port’s RX and TX hardware rings is performed similarly to that in the L2 Forwarding sample application. See L2 Forwarding Sample Application (in Real and Virtualized Environments) for more information.

Statistics Display

When run in a linuxapp environment, the VMDQ and DCB Forwarding sample application can display statistics showing the number of packets read from each RX queue. This is provided by way of a signal handler for the SIGHUP signal, which simply prints to standard output the packet counts in grid form. Each row of the output is a single pool with the columns being the queue number within that pool.

To generate the statistics output, use the following command:

user@host$ sudo killall -HUP vmdq_dcb_app

Please note that the statistics output will appear on the terminal where the vmdq_dcb_app is running, rather than the terminal from which the HUP signal was sent.

Vhost Sample Application

The vhost sample application demonstrates integration of the Data Plane Development Kit (DPDK) with the Linux* KVM hypervisor by implementing the vhost-net offload API. The sample application performs simple packet switching between virtual machines based on Media Access Control (MAC) address or Virtual Local Area Network (VLAN) tag. The splitting of Ethernet traffic from an external switch is performed in hardware by the Virtual Machine Device Queues (VMDQ) and Data Center Bridging (DCB) features of the Intel® 82599 10 Gigabit Ethernet Controller.

Testing steps

This section shows the steps how to test a typical PVP case with this vhost-switch sample, whereas packets are received from the physical NIC port first and enqueued to the VM’s Rx queue. Through the guest testpmd’s default forwarding mode (io forward), those packets will be put into the Tx queue. The vhost-switch example, in turn, gets the packets and puts back to the same physical NIC port.

Build

To compile the sample application see Compiling the Sample Applications.

The application is located in the vhost sub-directory.

Note

In this example, you need build DPDK both on the host and inside guest.

Start the vswitch example
./vhost-switch -l 0-3 -n 4 --socket-mem 1024  \
     -- --socket-file /tmp/sock0 --client \
     ...

Check the Parameters section for the explanations on what do those parameters mean.

Start the VM
qemu-system-x86_64 -machine accel=kvm -cpu host \
    -m $mem -object memory-backend-file,id=mem,size=$mem,mem-path=/dev/hugepages,share=on \
            -mem-prealloc -numa node,memdev=mem \
    \
    -chardev socket,id=char1,path=/tmp/sock0,server \
    -netdev type=vhost-user,id=hostnet1,chardev=char1  \
    -device virtio-net-pci,netdev=hostnet1,id=net1,mac=52:54:00:00:00:14 \
    ...

Note

For basic vhost-user support, QEMU 2.2 (or above) is required. For some specific features, a higher version might be need. Such as QEMU 2.7 (or above) for the reconnect feature.

Run testpmd inside guest

Make sure you have DPDK built inside the guest. Also make sure the corresponding virtio-net PCI device is bond to a uio driver, which could be done by:

modprobe uio_pci_generic
$RTE_SDK/usertools/dpdk-devbind.py -b=uio_pci_generic 0000:00:04.0

Then start testpmd for packet forwarding testing.

./x86_64-native-gcc/app/testpmd -l 0-1 -- -i
> start tx_first

Inject packets

While a virtio-net is connected to vhost-switch, a VLAN tag starts with 1000 is assigned to it. So make sure configure your packet generator with the right MAC and VLAN tag, you should be able to see following log from the vhost-switch console. It means you get it work:

VHOST_DATA: (0) mac 52:54:00:00:00:14 and vlan 1000 registered

Parameters

–socket-file path Specifies the vhost-user socket file path.

–client DPDK vhost-user will act as the client mode when such option is given. In the client mode, QEMU will create the socket file. Otherwise, DPDK will create it. Put simply, it’s the server to create the socket file.

–vm2vm mode The vm2vm parameter sets the mode of packet switching between guests in the host.

  • 0 disables vm2vm, impling that VM’s packets will always go to the NIC port.
  • 1 means a normal mac lookup packet routing.
  • 2 means hardware mode packet forwarding between guests, it allows packets go to the NIC port, hardware L2 switch will determine which guest the packet should forward to or need send to external, which bases on the packet destination MAC address and VLAN tag.

–mergeable 0|1 Set 0/1 to disable/enable the mergeable Rx feature. It’s disabled by default.

–stats interval The stats parameter controls the printing of virtio-net device statistics. The parameter specifies an interval (in unit of seconds) to print statistics, with an interval of 0 seconds disabling statistics.

–rx-retry 0|1 The rx-retry option enables/disables enqueue retries when the guests Rx queue is full. This feature resolves a packet loss that is observed at high data rates, by allowing it to delay and retry in the receive path. This option is enabled by default.

–rx-retry-num num The rx-retry-num option specifies the number of retries on an Rx burst, it takes effect only when rx retry is enabled. The default value is 4.

–rx-retry-delay msec The rx-retry-delay option specifies the timeout (in micro seconds) between retries on an RX burst, it takes effect only when rx retry is enabled. The default value is 15.

–dequeue-zero-copy Dequeue zero copy will be enabled when this option is given.

–vlan-strip 0|1 VLAN strip option is removed, because different NICs have different behaviors when disabling VLAN strip. Such feature, which heavily depends on hardware, should be removed from this example to reduce confusion. Now, VLAN strip is enabled and cannot be disabled.

Common Issues

  • QEMU fails to allocate memory on hugetlbfs, with an error like the following:

    file_ram_alloc: can't mmap RAM pages: Cannot allocate memory
    

    When running QEMU the above error indicates that it has failed to allocate memory for the Virtual Machine on the hugetlbfs. This is typically due to insufficient hugepages being free to support the allocation request. The number of free hugepages can be checked as follows:

    cat /sys/kernel/mm/hugepages/hugepages-<pagesize>/nr_hugepages
    

    The command above indicates how many hugepages are free to support QEMU’s allocation request.

  • vhost-user will not work with QEMU without the -mem-prealloc option

    The current implementation works properly only when the guest memory is pre-allocated.

  • vhost-user will not work with a QEMU version without shared memory mapping:

    Make sure share=on QEMU option is given.

  • Failed to build DPDK in VM

    Make sure “-cpu host” QEMU option is given.

Vhost_scsi Sample Application

The vhost_scsi sample application implemented a simple SCSI block device, which used as the backend of Qemu vhost-user-scsi device. Users can extend the exist example to use other type of block device(e.g. AIO) besides memory based block device. Similar with vhost-user-net device, the sample application used domain socket to communicate with Qemu, and the virtio ring was processed by vhost_scsi sample application.

The sample application reuse lots codes from SPDK(Storage Performance Development Kit, https://github.com/spdk/spdk) vhost-user-scsi target, for DPDK vhost library used in storage area, user can take SPDK as reference as well.

Testing steps

This section shows the steps how to start a VM with the block device as fast data path for critical application.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the examples sub-directory.

You will also need to build DPDK both on the host and inside the guest

Start the vhost_scsi example
./vhost_scsi -m 1024
Start the VM
qemu-system-x86_64 -machine accel=kvm \
    -m $mem -object memory-backend-file,id=mem,size=$mem,\
    mem-path=/dev/hugepages,share=on -numa node,memdev=mem \
    -drive file=os.img,if=none,id=disk \
    -device ide-hd,drive=disk,bootindex=0 \
    -chardev socket,id=char0,path=/tmp/vhost.socket \
    -device vhost-user-scsi-pci,chardev=char0,bootindex=2 \
    ...

Note

You must check whether your Qemu can support “vhost-user-scsi” or not, Qemu v2.10 or newer version is required.

Vhost_scsi Common Issues

  • vhost_scsi can not start with block size 512 Bytes:

    Currently DPDK vhost library was designed for NET device(althrough the APIs are generic now), for 512 Bytes block device, Qemu BIOS(x86 BIOS Enhanced Disk Device) will enumerate all block device and do some IOs to those block devices with 512 Bytes sector size. DPDK vhost library can not process such scenarios(both BIOS and OS will enumerate the block device), so as a workaround, the vhost_scsi example application hardcoded the block size with 4096 Bytes.

  • vhost_scsi can only support the block device as fast data disk(non OS image):

    Make sure bootindex=2 Qemu option is given to vhost-user-scsi-pci device.

Netmap Compatibility Sample Application

Introduction

The Netmap compatibility library provides a minimal set of APIs to give programs written against the Netmap APIs the ability to be run, with minimal changes to their source code, using the DPDK to perform the actual packet I/O.

Since Netmap applications use regular system calls, like open(), ioctl() and mmap() to communicate with the Netmap kernel module performing the packet I/O, the compat_netmap library provides a set of similar APIs to use in place of those system calls, effectively turning a Netmap application into a DPDK application.

The provided library is currently minimal and doesn’t support all the features that Netmap supports, but is enough to run simple applications, such as the bridge example detailed below.

Knowledge of Netmap is required to understand the rest of this section. Please refer to the Netmap distribution for details about Netmap.

Available APIs

The library provides the following drop-in replacements for system calls usually used in Netmap applications:

  • rte_netmap_close()
  • rte_netmap_ioctl()
  • rte_netmap_open()
  • rte_netmap_mmap()
  • rte_netmap_poll()

They use the same signature as their libc counterparts, and can be used as drop-in replacements in most cases.

Caveats

Given the difference between the way Netmap and the DPDK approach packet I/O, there are caveats and limitations to be aware of when trying to use the compat_netmap library, the most important of these are listed below. These may change as the library is updated:

  • Any system call that can potentially affect file descriptors cannot be used with a descriptor returned by the rte_netmap_open() function.

Note that:

  • The rte_netmap_mmap() function merely returns the address of a DPDK memzone. The address, length, flags, offset, and other arguments are ignored.
  • The rte_netmap_poll() function only supports infinite (negative) or zero time outs. It effectively turns calls to the poll() system call made in a Netmap application into polling of the DPDK ports, changing the semantics of the usual POSIX defined poll.
  • Not all of Netmap’s features are supported: host rings, slot flags and so on are not supported or are simply not relevant in the DPDK model.
  • The Netmap manual page states that “a device obtained through /dev/netmap also supports the ioctl supported by network devices”. This is not the case with this compatibility layer.
  • The Netmap kernel module exposes a sysfs interface to change some internal parameters, such as the size of the shared memory region. This interface is not available when using this compatibility layer.

Porting Netmap Applications

Porting Netmap applications typically involves two major steps:

  • Changing the system calls to use their compat_netmap library counterparts.
  • Adding further DPDK initialization code.

Since the compat_netmap functions have the same signature as the usual libc calls, the change is trivial in most cases.

The usual DPDK initialization code involving rte_eal_init() and rte_pci_probe() has to be added to the Netmap application in the same way it is used in all other DPDK sample applications. Please refer to the DPDK Programmer’s Guide and example source code for details about initialization.

In addition of the regular DPDK initialization code, the ported application needs to call initialization functions for the compat_netmap library, namely rte_netmap_init() and rte_netmap_init_port().

These two initialization functions take compat_netmap specific data structures as parameters: struct rte_netmap_conf and struct rte_netmap_port_conf. The structures’ fields are Netmap related and are self-explanatory for developers familiar with Netmap. They are defined in $RTE_SDK/examples/netmap_compat/lib/compat_netmap.h.

The bridge application is an example largely based on the bridge example shipped with the Netmap distribution. It shows how a minimal Netmap application with minimal and straightforward source code changes can be run on top of the DPDK. Please refer to $RTE_SDK/examples/netmap_compat/bridge/bridge.c for an example of a ported application.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the netmap_compat sub-directory.

Running the “bridge” Sample Application

The application requires a single command line option:

./build/bridge [EAL options] -- -i INTERFACE_A [-i INTERFACE_B]

where,

  • -i INTERFACE: Interface (DPDK port number) to use.

    If a single -i parameter is given, the interface will send back all the traffic it receives. If two -i parameters are given, the two interfaces form a bridge, where traffic received on one interface is replicated and sent to the other interface.

For example, to run the application in a linuxapp environment using port 0 and 2:

./build/bridge [EAL options] -- -i 0 -i 2

Refer to the DPDK Getting Started Guide for Linux for general information on running applications and the Environment Abstraction Layer (EAL) options.

Note that unlike a traditional bridge or the l2fwd sample application, no MAC address changes are done on the frames. Do not forget to take this into account when configuring a traffic generators and testing this sample application.

Internet Protocol (IP) Pipeline Application

Application overview

The Internet Protocol (IP) Pipeline application is intended to be a vehicle for rapid development of packet processing applications running on multi-core CPUs.

The application provides a library of reusable functional blocks called pipelines. These pipelines can be seen as prefabricated blocks that can be instantiated and inter-connected through packet queues to create complete applications (super-pipelines).

Pipelines are created and inter-connected through the application configuration file. By using different configuration files, different applications are effectively created, therefore this application can be seen as an application generator. The configuration of each pipeline can be updated at run-time through the application Command Line Interface (CLI).

Main application components are:

A Library of reusable pipelines

  • Each pipeline represents a functional block, e.g. flow classification, firewall, routing, master, etc.
  • Each pipeline type can be instantiated several times in the same application, which each instance configured separately and mapped to a single CPU core. Each CPU core can run one or several pipeline instances, which can be of same or different type.
  • Pipeline instances are inter-connected through packet queues (for packet processing) and message queues (for run-time configuration).
  • Pipelines are implemented using DPDK Packet Framework.
  • More pipeline types can always be built and added to the existing pipeline types.

The Configuration file

  • The configuration file defines the application structure. By using different configuration files, different applications are created.
  • All the application resources are created and configured through the application configuration file: pipeline instances, buffer pools, links (i.e. network interfaces), hardware device RX/TX queues, software queues, traffic manager devices, EAL startup arguments, etc.
  • The configuration file syntax is “define by reference”, meaning that resources are defined as they are referenced. First time a resource name is detected, it is registered with default parameters. Optionally, the resource parameters can be further refined through a configuration file section dedicated to that resource.
  • Command Line Interface (CLI)

Global CLI commands: link configuration, etc.

  • Common pipeline CLI commands: ping (keep-alive), statistics, etc.
  • Pipeline type specific CLI commands: used to configure instances of specific pipeline type. These commands are registered with the application when the pipeline type is registered. For example, the commands for routing pipeline instances include: route add, route delete, route list, etc.
  • CLI commands can be grouped into scripts that can be invoked at initialization and at runtime.

Design goals

Rapid development

This application enables rapid development through quick connectivity of standard components called pipelines. These components are built using DPDK Packet Framework and encapsulate packet processing features at different levels: ports, tables, actions, pipelines and complete applications.

Pipeline instances are instantiated, configured and inter-connected through low complexity configuration files loaded during application initialization. Each pipeline instance is mapped to a single CPU core, with each CPU core able to run one or multiple pipeline instances of same or different types. By loading a different configuration file, a different application is effectively started.

Flexibility

Each packet processing application is typically represented as a chain of functional stages which is often called the functional pipeline of the application. These stages are mapped to CPU cores to create chains of CPU cores (pipeline model), clusters of CPU cores (run-to-completion model) or chains of clusters of CPU cores (hybrid model).

This application allows all the above programming models. By applying changes to the configuration file, the application provides the flexibility to reshuffle its building blocks in different ways until the configuration providing the best performance is identified.

Move pipelines around

The mapping of pipeline instances to CPU cores can be reshuffled through the configuration file. One or several pipeline instances can be mapped to the same CPU core.

_images/ip_pipelines_1.svg

Example of moving pipeline instances across different CPU cores

Move tables around

There is some degree of flexibility for moving tables from one pipeline instance to another. Based on the configuration arguments passed to each pipeline instance in the configuration file, specific tables can be enabled or disabled. This way, a specific table can be “moved” from pipeline instance A to pipeline instance B by simply disabling its associated functionality for pipeline instance A while enabling it for pipeline instance B.

Due to requirement to have simple syntax for the configuration file, moving tables across different pipeline instances is not as flexible as the mapping of pipeline instances to CPU cores, or mapping actions to pipeline tables. Complete flexibility in moving tables from one pipeline to another could be achieved through a complex pipeline description language that would detail the structural elements of the pipeline (ports, tables and actions) and their connectivity, resulting in complex syntax for the configuration file, which is not acceptable. Good configuration file readability through simple syntax is preferred.

Example: the IP routing pipeline can run the routing function only (with ARP function run by a different pipeline instance), or it can run both the routing and ARP functions as part of the same pipeline instance.

_images/ip_pipelines_2.svg

Example of moving tables across different pipeline instances

Move actions around

When it makes sense, packet processing actions can be moved from one pipeline instance to another. Based on the configuration arguments passed to each pipeline instance in the configuration file, specific actions can be enabled or disabled. This way, a specific action can be “moved” from pipeline instance A to pipeline instance B by simply disabling its associated functionality for pipeline instance A while enabling it for pipeline instance B.

Example: The flow actions of accounting, traffic metering, application identification, NAT, etc can be run as part of the flow classification pipeline instance or split across several flow actions pipeline instances, depending on the number of flow instances and their compute requirements.

_images/ip_pipelines_3.svg

Example of moving actions across different tables and pipeline instances

Performance

Performance of the application is the highest priority requirement. Flexibility is not provided at the expense of performance.

The purpose of flexibility is to provide an incremental development methodology that allows monitoring the performance evolution:

  • Apply incremental changes in the configuration (e.g. mapping on pipeline instances to CPU cores) in order to identify the configuration providing the best performance for a given application;
  • Add more processing incrementally (e.g. by enabling more actions for specific pipeline instances) until the application is feature complete while checking the performance impact at each step.
Debug capabilities

The application provides a significant set of debug capabilities:

  • Command Line Interface (CLI) support for statistics polling: pipeline instance ping (keep-alive checks), pipeline instance statistics per input port/output port/table, link statistics, etc;
  • Logging: Turn on/off application log messages based on priority level;

Running the application

The application startup command line is:

ip_pipeline [-f CONFIG_FILE] [-s SCRIPT_FILE] -p PORT_MASK [-l LOG_LEVEL]

The application startup arguments are:

-f CONFIG_FILE

  • Optional: Yes
  • Default: ./config/ip_pipeline.cfg
  • Argument: Path to the configuration file to be loaded by the application. Please refer to the Configuration file syntax for details on how to write the configuration file.

-s SCRIPT_FILE

  • Optional: Yes
  • Default: Not present
  • Argument: Path to the CLI script file to be run by the master pipeline at application startup. No CLI script file will be run at startup if this argument is not present.

-p PORT_MASK

  • Optional: No
  • Default: N/A
  • Argument: Hexadecimal mask of NIC port IDs to be used by the application. First port enabled in this mask will be referenced as LINK0 as part of the application configuration file, next port as LINK1, etc.

-l LOG_LEVEL

  • Optional: Yes
  • Default: 1 (High priority)
  • Argument: Log level to determine which application messages are to be printed to standard output. Available log levels are: 0 (None), 1 (High priority), 2 (Low priority). Only application messages whose priority is higher than or equal to the application log level will be printed.

Application stages

Configuration

During this stage, the application configuration file is parsed and its content is loaded into the application data structures. In case of any configuration file parse error, an error message is displayed and the application is terminated. Please refer to the Configuration file syntax for a description of the application configuration file format.

Configuration checking

In the absence of any parse errors, the loaded content of application data structures is checked for overall consistency. In case of any configuration check error, an error message is displayed and the application is terminated.

Initialization

During this stage, the application resources are initialized and the handles to access them are saved into the application data structures. In case of any initialization error, an error message is displayed and the application is terminated.

The typical resources to be initialized are: pipeline instances, buffer pools, links (i.e. network interfaces), hardware device RX/TX queues, software queues, traffic management devices, etc.

Run-time

Each CPU core runs the pipeline instances assigned to it in time sharing mode and in round robin order:

  1. Packet processing task: The pipeline run-time code is typically a packet processing task built on top of DPDK Packet Framework rte_pipeline library, which reads bursts of packets from the pipeline input ports, performs table lookups and executes the identified actions for all tables in the pipeline, with packet eventually written to pipeline output ports or dropped.
  2. Message handling task: Each CPU core will also periodically execute the message handling code of each of the pipelines mapped to it. The pipeline message handling code is processing the messages that are pending in the pipeline input message queues, which are typically sent by the master CPU core for the on-the-fly pipeline configuration: check that pipeline is still alive (ping), add/delete entries in the pipeline tables, get statistics, etc. The frequency of executing the message handling code is usually much smaller than the frequency of executing the packet processing work.

Please refer to the PIPELINE section for more details about the application pipeline module encapsulation.

Configuration file syntax

Syntax overview

The syntax of the configuration file is designed to be simple, which favors readability. The configuration file is parsed using the DPDK library librte_cfgfile, which supports simple INI file format for configuration files.

As result, the configuration file is split into several sections, with each section containing one or more entries. The scope of each entry is its section, and each entry specifies a variable that is assigned a specific value. Any text after the ; character is considered a comment and is therefore ignored.

The following are application specific: number of sections, name of each section, number of entries of each section, name of the variables used for each section entry, the value format (e.g. signed/unsigned integer, string, etc) and range of each section entry variable.

Generic example of configuration file section:

[<section_name>]

<variable_name_1> = <value_1>

; ...

<variable_name_N> = <value_N>
Application resources present in the configuration file
Application resource names in the configuration file
Resource type Format Examples
Pipeline PIPELINE<ID> PIPELINE0, PIPELINE1
Mempool MEMPOOL<ID> MEMPOOL0, MEMPOOL1
Link (network interface) LINK<ID> LINK0, LINK1
Link RX queue RXQ<LINK_ID>.<QUEUE_ID> RXQ0.0, RXQ1.5
Link TX queue TXQ<LINK_ID>.<QUEUE_ID> TXQ0.0, TXQ1.5
Software queue SWQ<ID> SWQ0, SWQ1
Traffic Manager TM<LINK_ID> TM0, TM1
KNI (kernel NIC interface) KNI<LINK_ID> KNI0, KNI1
Source SOURCE<ID> SOURCE0, SOURCE1
Sink SINK<ID> SINK0, SINK1
Message queue MSGQ<ID> MSGQ-REQ-PIPELINE<ID> MSGQ-RSP-PIPELINE<ID> MSGQ-REQ-CORE-<CORE_ID> MSGQ-RSP-CORE-<CORE_ID> MSGQ0, MSGQ1, MSGQ-REQ-PIPELINE2, MSGQ-RSP-PIPELINE2, MSGQ-REQ-CORE-s0c1, MSGQ-RSP-CORE-s0c1

LINK instances are created implicitly based on the PORT_MASK application startup argument. LINK0 is the first port enabled in the PORT_MASK, port 1 is the next one, etc. The LINK ID is different than the DPDK PMD-level NIC port ID, which is the actual position in the bitmask mentioned above. For example, if bit 5 is the first bit set in the bitmask, then LINK0 is having the PMD ID of 5. This mechanism creates a contiguous LINK ID space and isolates the configuration file against changes in the board PCIe slots where NICs are plugged in.

RXQ, TXQ, TM and KNI instances have the LINK ID as part of their name. For example, RXQ2.1, TXQ2.1 and TM2 are all associated with LINK2.

Rules to parse the configuration file

The main rules used to parse the configuration file are:

  1. Application resource name determines the type of resource based on the name prefix.

    Example: all software queues need to start with SWQ prefix, so SWQ0 and SWQ5 are valid software queue names.

  2. An application resource is defined by creating a configuration file section with its name. The configuration file section allows fine tuning on any of the resource parameters. Some resource parameters are mandatory, in which case it is required to have them specified as part of the section, while some others are optional, in which case they get assigned their default value when not present.

    Example: section SWQ0 defines a software queue named SWQ0, whose parameters are detailed as part of this section.

  3. An application resource can also be defined by referencing it. Referencing a resource takes place by simply using its name as part of the value assigned to a variable in any configuration file section. In this case, the resource is registered with all its parameters having their default values. Optionally, a section with the resource name can be added to the configuration file to fine tune some or all of the resource parameters.

    Example: in section PIPELINE3, variable pktq_in includes SWQ5 as part of its list, which results in defining a software queue named SWQ5; when there is no SWQ5 section present in the configuration file, SWQ5 gets registered with default parameters.

PIPELINE section
Configuration file PIPELINE section (1/2)
Section Description Optional Range Default value
type Pipeline type. Defines the functionality to be executed. NO See “List of pipeline types” N/A
core CPU core to run the current pipeline. YES See “CPU Core notation” CPU socket 0, core 0, hyper-thread 0
pktq_in Packet queues to serve as input ports for the current pipeline instance. The acceptable packet queue types are: RXQ, SWQ, TM and SOURCE. First device in this list is used as pipeline input port 0, second as pipeline input port 1, etc. YES List of input packet queue IDs Empty list
pktq_out Packet queues to serve as output ports for the current pipeline instance. The acceptable packet queue types are: TXQ, SWQ, TM and SINK. First device in this list is used as pipeline output port 0, second as pipeline output port 1, etc. YES List of output packet queue IDs. Empty list
Configuration file PIPELINE section (2/2)
Section Description Optional Range Default value
msgq_in Input message queues. These queues contain request messages that need to be handled by the current pipeline instance. The type and format of request messages is defined by the pipeline type. For each pipeline instance, there is an input message queue defined implicitly, whose name is: MSGQ-REQ-<PIPELINE_ID>. This message queue should not be mentioned as part of msgq_in list. YES List of message queue IDs Empty list
msgq_out Output message queues. These queues are used by the current pipeline instance to write response messages as result of request messages being handled. The type and format of response messages is defined by the pipeline type. For each pipeline instance, there is an output message queue defined implicitly, whose name is: MSGQ-RSP-<PIPELINE_ID>. This message queue should not be mentioned as part of msgq_out list. YES List of message queue IDs Empty list
timer_period Time period, measured in milliseconds, for handling the input message queues. YES milliseconds 1 ms
<any other> Arguments to be passed to the current pipeline instance. Format of the arguments, their type, whether each argument is optional or mandatory and its default value (when optional) are defined by the pipeline type. The value of the arguments is applicable to the current pipeline instance only. Depends on pipeline type Depends on pipeline type Depends on pipeline type
CPU core notation

The CPU Core notation is:

<CPU core> ::= [s|S<CPU socket ID>][c|C]<CPU core ID>[h|H]

For example:

CPU socket 0, core 0, hyper-thread 0: 0, c0, s0c0

CPU socket 0, core 0, hyper-thread 1: 0h, c0h, s0c0h

CPU socket 3, core 9, hyper-thread 1: s3c9h
MEMPOOL section
Configuration file MEMPOOL section
Section Description Optional Type Default value
buffer_size Buffer size (in bytes) for the current buffer pool. YES uint32_t 2048 + sizeof(struct rte_mbuf) + HEADROOM
pool_size Number of buffers in the current buffer pool. YES uint32_t 32K
cache_size Per CPU thread cache size (in number of buffers) for the current buffer pool. YES uint32_t 256
cpu CPU socket ID where to allocate memory for the current buffer pool. YES uint32_t 0
RXQ section
Configuration file RXQ section
Section Description Optional Type Default value
mempool Mempool to use for buffer allocation for current NIC RX queue. The mempool ID has to be associated with a valid instance defined in the mempool entry of the global section. YES uint32_t MEMPOOL0
Size NIC RX queue size (number of descriptors) YES uint32_t 128
burst Read burst size (number of descriptors) YES uint32_t 32
TXQ section
Configuration file TXQ section
Section Description Optional Type Default value
size NIC TX queue size (number of descriptors) YES uint32_t power of 2 > 0 512
burst Write burst size (number of descriptors) YES uint32_t power of 2 0 < burst < size 32
dropless When dropless is set to NO, packets can be dropped if not enough free slots are currently available in the queue, so the write operation to the queue is non- blocking. When dropless is set to YES, packets cannot be dropped if not enough free slots are currently available in the queue, so the write operation to the queue is blocking, as the write operation is retried until enough free slots become available and all the packets are successfully written to the queue. YES YES/NO NO
n_retries Number of retries. Valid only when dropless is set to YES. When set to 0, it indicates unlimited number of retries. YES uint32_t 0
SWQ section
Configuration file SWQ section
Section Description Optional Type Default value
size Queue size (number of packets) YES uint32_t power of 2 256
burst_read Read burst size (number of packets) YES uint32_t power of 2 0 < burst < size 32
burst_write Write burst size (number of packets) YES uint32_t power of 2 0 < burst < size 32
dropless When dropless is set to NO, packets can be dropped if not enough free slots are currently available in the queue, so the write operation to the queue is non- blocking. When dropless is set to YES, packets cannot be dropped if not enough free slots are currently available in the queue, so the write operation to the queue is blocking, as the write operation is retried until enough free slots become available and all the packets are successfully written to the queue. YES YES/NO NO
n_retries Number of retries. Valid only when dropless is set to YES. When set to 0, it indicates unlimited number of retries. YES uint32_t 0
cpu CPU socket ID where to allocate memory for this SWQ. YES uint32_t 0
TM section
Configuration file TM section
Section Description Optional Type Default value
Cfg File name to parse for the TM configuration to be applied. The syntax of this file is described in the examples/qos_sched DPDK application documentation. YES string tm_profile
burst_read Read burst size (number of packets) YES uint32_t 64
burst_write Write burst size (number of packets) YES uint32_t 32
KNI section
Configuration file KNI section
Section Description Optional Type Default value
core CPU core to run the KNI kernel thread. When core config is set, the KNI kernel thread will be bound to the particular core. When core config is not set, the KNI kernel thread will be scheduled by the OS. YES See “CPU Core notation” Not set
mempool Mempool to use for buffer allocation for current KNI port. The mempool ID has to be associated with a valid instance defined in the mempool entry of the global section. YES uint32_t MEMPOOL0
burst_read Read burst size (number of packets) YES uint32_t power of 2 0 < burst < size 32
burst_write Write burst size (number of packets) YES uint32_t power of 2 0 < burst < size 32
dropless When dropless is set to NO, packets can be dropped if not enough free slots are currently available in the queue, so the write operation to the queue is non- blocking. When dropless is set to YES, packets cannot be dropped if not enough free slots are currently available in the queue, so the write operation to the queue is blocking, as the write operation is retried until enough free slots become available and all the packets are successfully written to the queue. YES YES/NO NO
n_retries Number of retries. Valid only when dropless is set to YES. When set to 0, it indicates unlimited number of retries. YES uint64_t 0
SOURCE section
Configuration file SOURCE section
Section Description Optional Type Default value
Mempool Mempool to use for buffer allocation. YES uint32_t MEMPOOL0
Burst Read burst size (number of packets)   uint32_t 32
SINK section

Currently, there are no parameters to be passed to a sink device, so SINK section is not allowed.

MSGQ section
Configuration file MSGQ section
Section Description Optional Type Default value
size Queue size (number of packets) YES uint32_t != 0 power of 2 64
cpu CPU socket ID where to allocate memory for the current queue. YES uint32_t 0
EAL section

The application generates the EAL parameters rather than reading them from the command line.

The CPU core mask parameter is generated based on the core entry of all PIPELINE sections. All the other EAL parameters can be set from this section of the application configuration file.

Library of pipeline types

Pipeline module

A pipeline is a self-contained module that implements a packet processing function and is typically implemented on top of the DPDK Packet Framework librte_pipeline library. The application provides a run-time mechanism to register different pipeline types.

Depending on the required configuration, each registered pipeline type (pipeline class) is instantiated one or several times, with each pipeline instance (pipeline object) assigned to one of the available CPU cores. Each CPU core can run one or more pipeline instances, which might be of same or different types. For more information of the CPU core threading model, please refer to the Run-time section.

Pipeline type

Each pipeline type is made up of a back-end and a front-end. The back-end represents the packet processing engine of the pipeline, typically implemented using the DPDK Packet Framework libraries, which reads packets from the input packet queues, handles them and eventually writes them to the output packet queues or drops them. The front-end represents the run-time configuration interface of the pipeline, which is exposed as CLI commands. The front-end communicates with the back-end through message queues.

Pipeline back-end
Field name Field type Description
f_init Function pointer Function to initialize the back-end of the current pipeline instance. Typical work implemented by this function for the current pipeline instance: Memory allocation; Parse the pipeline type specific arguments; Initialize the pipeline input ports, output ports and tables, interconnect input ports to tables; Set the message handlers.
f_free Function pointer Function to free the resources allocated by the back-end of the current pipeline instance.
f_run Function pointer Set to NULL for pipelines implemented using the DPDK library librte_pipeline (typical case), and to non-NULL otherwise. This mechanism is made available to support quick integration of legacy code. This function is expected to provide the packet processing related code to be called as part of the CPU thread dispatch loop, so this function is not allowed to contain an infinite loop.
f_timer Function pointer Function to read the pipeline input message queues, handle the request messages, create response messages and write the response queues. The format of request and response messages is defined by each pipeline type, with the exception of some requests which are mandatory for all pipelines (e.g. ping, statistics).
f_track Function pointer See section Tracking pipeline output port to physical link
Pipeline front-end
Field name Field type Description
f_init Function pointer Function to initialize the front-end of the current pipeline instance.
f_free Function pointer Function to free the resources allocated by the front-end of the current pipeline instance.
cmds Array of CLI commands Array of CLI commands to be registered to the application CLI for the current pipeline type. Even though the CLI is executed by a different pipeline (typically, this is the master pipeline), from modularity perspective is more efficient to keep the message client side (part of the front-end) together with the message server side (part of the back-end).
Table copies

Fast table copy: pipeline table used by pipeline for the packet processing task, updated through messages, table data structures are optimized for lookup operation.

Slow table copy: used by the configuration layer, typically updated through CLI commands, kept in sync with the fast copy (its update triggers the fast copy update). Required for executing advanced table queries without impacting the packet processing task, therefore the slow copy is typically organized using different criteria than the fast copy.

Examples:

  • Flow classification: Search through current set of flows (e.g. list all flows with a specific source IP address);
  • Firewall: List rules in descending order of priority;
  • Routing table: List routes sorted by prefix depth and their type (local, remote, default);
  • ARP: List entries sorted per output interface.
Packet meta-data

Packet meta-data field offsets provided as argument to pipeline instances are essentially defining the data structure for the packet meta-data used by the current application use-case. It is very useful to put it in the configuration file as a comment in order to facilitate the readability of the configuration file.

The reason to use field offsets for defining the data structure for the packet meta-data is due to the C language limitation of not being able to define data structures at run-time. Feature to consider: have the configuration file parser automatically generate and print the data structure defining the packet meta-data for the current application use-case.

Packet meta-data typically contains:

  1. Pure meta-data: intermediate data per packet that is computed internally, passed between different tables of the same pipeline instance (e.g. lookup key for the ARP table is obtained from the routing table), or between different pipeline instances (e.g. flow ID, traffic metering color, etc);
  2. Packet fields: typically, packet header fields that are read directly from the packet, or read from the packet and saved (duplicated) as a working copy at a different location within the packet meta-data (e.g. Diffserv 5-tuple, IP destination address, etc).

Several strategies are used to design the packet meta-data, as described in the next subsections.

Store packet meta-data in a different cache line as the packet headers

This approach is able to support protocols with variable header length, like MPLS, where the offset of IP header from the start of the packet (and, implicitly, the offset of the IP header in the packet buffer) is not fixed. Since the pipelines typically require the specification of a fixed offset to the packet fields (e.g. Diffserv 5-tuple, used by the flow classification pipeline, or the IP destination address, used by the IP routing pipeline), the workaround is to have the packet RX pipeline copy these fields at fixed offsets within the packet meta-data.

As this approach duplicates some of the packet fields, it requires accessing more cache lines per packet for filling in selected packet meta-data fields (on RX), as well as flushing selected packet meta-data fields into the packet (on TX).

Example:

; struct app_pkt_metadata {
;   uint32_t ip_da;
;      uint32_t hash;
;      uint32_t flow_id;
;      uint32_t color;
; } __attribute__((__packed__));
;

[PIPELINE1]
; Packet meta-data offsets
ip_da_offset = 0;   Used by: routing
hash_offset = 4;    Used by: RX, flow classification
flow_id_offset = 8; Used by: flow classification, flow actions
color_offset = 12;  Used by: flow actions, routing
Overlay the packet meta-data in the same cache line with the packet headers

This approach is minimizing the number of cache line accessed per packet by storing the packet metadata in the same cache line with the packet headers. To enable this strategy, either some headroom is reserved for meta-data at the beginning of the packet headers cache line (e.g. if 16 bytes are needed for meta-data, then the packet headroom can be set to 128+16 bytes, so that NIC writes the first byte of the packet at offset 16 from the start of the first packet cache line), or meta-data is reusing the space of some packet headers that are discarded from the packet (e.g. input Ethernet header).

Example:

; struct app_pkt_metadata {
;     uint8_t headroom[RTE_PKTMBUF_HEADROOM]; /* 128 bytes (default) */
;     union {
;         struct {
;             struct ether_hdr ether; /* 14 bytes */
;             struct qinq_hdr qinq; /* 8 bytes */
;         };
;         struct {
;             uint32_t hash;
;             uint32_t flow_id;
;             uint32_t color;
;         };
;     };
;     struct ipv4_hdr ip; /* 20 bytes */
; } __attribute__((__packed__));
;
[PIPELINE2]
; Packet meta-data offsets
qinq_offset = 142;    Used by: RX, flow classification
ip_da_offset = 166;   Used by: routing
hash_offset = 128;    Used by: RX, flow classification
flow_id_offset = 132; Used by: flow classification, flow actions
color_offset = 136;   Used by: flow actions, routing
List of pipeline types
List of pipeline types provided with the application
Name Table(s) Actions Messages

Pass-through

Note: depending on port type, can be used for RX, TX, IP fragmentation, IP reassembly or Traffic Management

Passthrough
  1. Pkt metadata build
  2. Flow hash
  3. Pkt checks
  4. Load balancing
  1. Ping
  2. Stats
Flow classification

Exact match

  • Key = byte array (source: pkt metadata)
  • Data = action dependent
  1. Flow ID
  2. Flow stats
  3. Metering
  4. Network Address
  5. Translation (NAT)
  1. Ping
  2. Stats
  3. Flow stats
  4. Action stats
  5. Flow add/ update/ delete
  6. Default flow add/ update/ delete
  7. Action update
Flow actions

Array

  • Key = Flow ID (source: pkt metadata)
  • Data = action dependent
  1. Flow stats
  2. Metering
  3. Network Address
  4. Translation (NAT)
  1. Ping
  2. Stats
  3. Action stats
  4. Action update
Firewall

ACL

  • Key = n-tuple (source: pkt headers)
  • Data = none
  1. Allow/Drop
  1. Ping
  2. Stats
  3. Rule add/ update/ delete
  4. Default rule add/ update/ delete
IP routing

LPM (IPv4 or IPv6, depending on pipeline type)

  • Key = IP destination (source: pkt metadata)
  • Data = Dependent on actions and next hop type

Hash table (for ARP, only

when ARP is enabled)

  • Key = (Port ID, next hop IP address) (source: pkt meta-data)
  • Data: MAC address
  1. TTL decrement and
  2. IPv4 checksum
  3. update
  4. Header
  5. encapsulation
  6. (based on next hop
  7. type)
  1. Ping
  2. Stats
  3. Route add/ update/ delete
  4. Default route add/ update/ delete
  5. ARP entry add/ update/ delete
  6. Default ARP entry add/ update/ delete

Command Line Interface (CLI)

Global CLI commands
Global CLI commands
Command Description Syntax
run Run CLI commands script file. run <file> <file> = path to file with CLI commands to execute
quit Gracefully terminate the application. quit
CLI commands common for all pipeline types
CLI commands mandatory for all pipelines
Command Description Syntax
ping Check whether specific pipeline instance is alive. The master pipeline sends a ping request message to given pipeline instance and waits for a response message back. Timeout message is displayed when the response message is not received before the timer expires. p <pipeline ID> ping
stats Display statistics for specific pipeline input port, output port or table. p <pipeline ID> stats port in <port in ID> p <pipeline ID> stats port out <port out ID> p <pipeline ID> stats table <table ID>
input port enable Enable given input port for specific pipeline instance. p <pipeline ID> port in <port ID> enable
input port disable Disable given input port for specific pipeline instance. p <pipeline ID> port in <port ID> disable
Pipeline type specific CLI commands

The pipeline specific CLI commands are part of the pipeline type front-end.

Test Pipeline Application

The Test Pipeline application illustrates the use of the DPDK Packet Framework tool suite. Its purpose is to demonstrate the performance of single-table DPDK pipelines.

Overview

The application uses three CPU cores:

  • Core A (“RX core”) receives traffic from the NIC ports and feeds core B with traffic through SW queues.
  • Core B (“Pipeline core”) implements a single-table DPDK pipeline whose type is selectable through specific command line parameter. Core B receives traffic from core A through software queues, processes it according to the actions configured in the table entries that are hit by the input packets and feeds it to core C through another set of software queues.
  • Core C (“TX core”) receives traffic from core B through software queues and sends it to the NIC ports for transmission.
_images/test_pipeline_app.png

Test Pipeline Application

Compiling the Application

To compile the sample application see Compiling the Sample Applications

The application is located in the $RTE_SDK/test/test-pipline directory.

Running the Application

Application Command Line

The application execution command line is:

./test-pipeline [EAL options] -- -p PORTMASK --TABLE_TYPE

The -c or -l EAL CPU coremask/corelist option has to contain exactly 3 CPU cores. The first CPU core in the core mask is assigned for core A, the second for core B and the third for core C.

The PORTMASK parameter must contain 2 or 4 ports.

Table Types and Behavior

Table 39.1 describes the table types used and how they are populated.

The hash tables are pre-populated with 16 million keys. For hash tables, the following parameters can be selected:

  • Configurable key size implementation or fixed (specialized) key size implementation (e.g. hash-8-ext or hash-spec-8-ext). The key size specialized implementations are expected to provide better performance for 8-byte and 16-byte key sizes, while the key-size-non-specialized implementation is expected to provide better performance for larger key sizes;
  • Key size (e.g. hash-spec-8-ext or hash-spec-16-ext). The available options are 8, 16 and 32 bytes;
  • Table type (e.g. hash-spec-16-ext or hash-spec-16-lru). The available options are ext (extendable bucket) or lru (least recently used).
Table Types
# TABLE_TYPE Description of Core B Table Pre-added Table Entries
1 none Core B is not implementing a DPDK pipeline. Core B is implementing a pass-through from its input set of software queues to its output set of software queues. N/A
2 stub Stub table. Core B is implementing the same pass-through functionality as described for the “none” option by using the DPDK Packet Framework by using one stub table for each input NIC port. N/A
3 hash-[spec]-8-lru LRU hash table with 8-byte key size and 16 million entries.

16 million entries are successfully added to the hash table with the following key format:

[4-byte index, 4 bytes of 0]

The action configured for all table entries is “Sendto output port”, with the output port index uniformly distributed for the range of output ports.

The default table rule (used in the case of a lookup miss) is to drop the packet.

At run time, core A is creating the following lookup key and storing it into the packet meta data for core B to use for table lookup:

[destination IPv4 address, 4 bytes of 0]

4 hash-[spec]-8-ext Extendable bucket hash table with 8-byte key size and 16 million entries. Same as hash-[spec]-8-lru table entries, above.
5 hash-[spec]-16-lru LRU hash table with 16-byte key size and 16 million entries.

16 million entries are successfully added to the hash table with the following key format:

[4-byte index, 12 bytes of 0]

The action configured for all table entries is “Send to output port”, with the output port index uniformly distributed for the range of output ports.

The default table rule (used in the case of a lookup miss) is to drop the packet.

At run time, core A is creating the following lookup key and storing it into the packet meta data for core B to use for table lookup:

[destination IPv4 address, 12 bytes of 0]

6 hash-[spec]-16-ext Extendable bucket hash table with 16-byte key size and 16 million entries. Same as hash-[spec]-16-lru table entries, above.
7 hash-[spec]-32-lru LRU hash table with 32-byte key size and 16 million entries.

16 million entries are successfully added to the hash table with the following key format:

[4-byte index, 28 bytes of 0].

The action configured for all table entries is “Send to output port”, with the output port index uniformly distributed for the range of output ports.

The default table rule (used in the case of a lookup miss) is to drop the packet.

At run time, core A is creating the following lookup key and storing it into the packet meta data for Lpmcore B to use for table lookup:

[destination IPv4 address, 28 bytes of 0]

8 hash-[spec]-32-ext Extendable bucket hash table with 32-byte key size and 16 million entries. Same as hash-[spec]-32-lru table entries, above.
9 lpm Longest Prefix Match (LPM) IPv4 table.

In the case of two ports, two routes are added to the table:

[0.0.0.0/9 => send to output port 0]

[0.128.0.0/9 => send to output port 1]

In case of four ports, four entries are added to the table:

[0.0.0.0/10 => send to output port 0]

[0.64.0.0/10 => send to output port 1]

[0.128.0.0/10 => send to output port 2]

[0.192.0.0/10 => send to output port 3]

The default table rule (used in the case of a lookup miss) is to drop the packet.

At run time, core A is storing the IPv4 destination within the packet meta data to be later used by core B as the lookup key.

10 acl Access Control List (ACL) table

In the case of two ports, two ACL rules are added to the table:

[priority = 0 (highest),

IPv4 source = ANY,

IPv4 destination = 0.0.0.0/9,

L4 protocol = ANY,

TCP source port = ANY,

TCP destination port = ANY

=> send to output port 0]

[priority = 0 (highest),

IPv4 source = ANY,

IPv4 destination = 0.128.0.0/9,

L4 protocol = ANY,

TCP source port = ANY,

TCP destination port = ANY

=> send to output port 0].

The default table rule (used in the case of a lookup miss) is to drop the packet.

Input Traffic

Regardless of the table type used for the core B pipeline, the same input traffic can be used to hit all table entries with uniform distribution, which results in uniform distribution of packets sent out on the set of output NIC ports. The profile for input traffic is TCP/IPv4 packets with:

  • destination IP address as A.B.C.D with A fixed to 0 and B, C,D random
  • source IP address fixed to 0.0.0.0
  • destination TCP port fixed to 0
  • source TCP port fixed to 0

Eventdev Pipeline SW PMD Sample Application

The eventdev pipeline sample application is a sample app that demonstrates the usage of the eventdev API using the software PMD. It shows how an application can configure a pipeline and assign a set of worker cores to perform the processing required.

The application has a range of command line arguments allowing it to be configured for various numbers worker cores, stages,queue depths and cycles per stage of work. This is useful for performance testing as well as quickly testing a particular pipeline configuration.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the examples sub-directory.

Running the Application

The application has a lot of command line options. This allows specification of the eventdev PMD to use, and a number of attributes of the processing pipeline options.

An example eventdev pipeline running with the software eventdev PMD using these settings is shown below:

  • -r1: core mask 0x1 for RX
  • -t1: core mask 0x1 for TX
  • -e4: core mask 0x4 for the software scheduler
  • -w FF00: core mask for worker cores, 8 cores from 8th to 16th
  • -s4: 4 atomic stages
  • -n0: process infinite packets (run forever)
  • -c32: worker dequeue depth of 32
  • -W1000: do 1000 cycles of work per packet in each stage
  • -D: dump statistics on exit
./build/eventdev_pipeline_sw_pmd --vdev event_sw0 -- -r1 -t1 -e4 -w FF00 -s4 -n0 -c32 -W1000 -D

The application has some sanity checking built-in, so if there is a function (eg; the RX core) which doesn’t have a cpu core mask assigned, the application will print an error message:

Core part of pipeline was not assigned any cores. This will stall the
pipeline, please check core masks (use -h for details on setting core masks):
        rx: 0
        tx: 1

Configuration of the eventdev is covered in detail in the programmers guide, see the Event Device Library section.

Observing the Application

At runtime the eventdev pipeline application prints out a summary of the configuration, and some runtime statistics like packets per second. On exit the worker statistics are printed, along with a full dump of the PMD statistics if required. The following sections show sample output for each of the output types.

Configuration

This provides an overview of the pipeline, scheduling type at each stage, and parameters to options such as how many flows to use and what eventdev PMD is in use. See the following sample output for details:

Config:
      ports: 2
      workers: 8
      packets: 0
      priorities: 1
      Queue-prio: 0
      qid0 type: atomic
      Cores available: 44
      Cores used: 10
      Eventdev 0: event_sw
Stages:
      Stage 0, Type Atomic    Priority = 128
      Stage 1, Type Atomic    Priority = 128
      Stage 2, Type Atomic    Priority = 128
      Stage 3, Type Atomic    Priority = 128
Runtime

At runtime, the statistics of the consumer are printed, stating the number of packets received, runtime in milliseconds, average mpps, and current mpps.

# consumer RX= xxxxxxx, time yyyy ms, avg z.zzz mpps [current w.www mpps]
Shutdown

At shutdown, the application prints the number of packets received and transmitted, and an overview of the distribution of work across worker cores.

Signal 2 received, preparing to exit...
  worker 12 thread done. RX=4966581 TX=4966581
  worker 13 thread done. RX=4963329 TX=4963329
  worker 14 thread done. RX=4953614 TX=4953614
  worker 0 thread done. RX=0 TX=0
  worker 11 thread done. RX=4970549 TX=4970549
  worker 10 thread done. RX=4986391 TX=4986391
  worker 9 thread done. RX=4970528 TX=4970528
  worker 15 thread done. RX=4974087 TX=4974087
  worker 8 thread done. RX=4979908 TX=4979908
  worker 2 thread done. RX=0 TX=0

Port Workload distribution:
worker 0 :      12.5 % (4979876 pkts)
worker 1 :      12.5 % (4970497 pkts)
worker 2 :      12.5 % (4986359 pkts)
worker 3 :      12.5 % (4970517 pkts)
worker 4 :      12.5 % (4966566 pkts)
worker 5 :      12.5 % (4963297 pkts)
worker 6 :      12.5 % (4953598 pkts)
worker 7 :      12.5 % (4974055 pkts)

To get a full dump of the state of the eventdev PMD, pass the -D flag to this application. When the app is terminated using Ctrl+C, the rte_event_dev_dump() function is called, resulting in a dump of the statistics that the PMD provides. The statistics provided depend on the PMD used, see the Event Device Drivers section for a list of eventdev PMDs.

Distributor Sample Application

The distributor sample application is a simple example of packet distribution to cores using the Data Plane Development Kit (DPDK).

Overview

The distributor application performs the distribution of packets that are received on an RX_PORT to different cores. When processed by the cores, the destination port of a packet is the port from the enabled port mask adjacent to the one on which the packet was received, that is, if the first four ports are enabled (port mask 0xf), ports 0 and 1 RX/TX into each other, and ports 2 and 3 RX/TX into each other.

This application can be used to benchmark performance using the traffic generator as shown in the figure below.

_images/dist_perf.svg

Performance Benchmarking Setup (Basic Environment)

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the distributor sub-directory.

Running the Application

  1. The application has a number of command line options:

    ./build/distributor_app [EAL options] -- -p PORTMASK
    

    where,

    • -p PORTMASK: Hexadecimal bitmask of ports to configure
  2. To run the application in linuxapp environment with 10 lcores, 4 ports, issue the command:

    $ ./build/distributor_app -l 1-9,22 -n 4 -- -p f
    
  3. Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

Explanation

The distributor application consists of four types of threads: a receive thread (lcore_rx()), a distributor thread (lcore_dist()), a set of worker threads (lcore_worker()), and a transmit thread(lcore_tx()). How these threads work together is shown in Fig. 41.2 below. The main() function launches threads of these four types. Each thread has a while loop which will be doing processing and which is terminated only upon SIGINT or ctrl+C.

The receive thread receives the packets using rte_eth_rx_burst() and will enqueue them to an rte_ring. The distributor thread will dequeue the packets from the ring and assign them to workers (using rte_distributor_process() API). This assignment is based on the tag (or flow ID) of the packet - indicated by the hash field in the mbuf. For IP traffic, this field is automatically filled by the NIC with the “usr” hash value for the packet, which works as a per-flow tag. The distributor thread communicates with the worker threads using a cache-line swapping mechanism, passing up to 8 mbuf pointers at a time (one cache line) to each worker.

More than one worker thread can exist as part of the application, and these worker threads do simple packet processing by requesting packets from the distributor, doing a simple XOR operation on the input port mbuf field (to indicate the output port which will be used later for packet transmission) and then finally returning the packets back to the distributor thread.

The distributor thread will then call the distributor api rte_distributor_returned_pkts() to get the processed packets, and will enqueue them to another rte_ring for transfer to the TX thread for transmission on the output port. The transmit thread will dequeue the packets from the ring and transmit them on the output port specified in packet mbuf.

Users who wish to terminate the running of the application have to press ctrl+C (or send SIGINT to the app). Upon this signal, a signal handler provided in the application will terminate all running threads gracefully and print final statistics to the user.

_images/dist_app.svg

Distributor Sample Application Layout

Debug Logging Support

Debug logging is provided as part of the application; the user needs to uncomment the line “#define DEBUG” defined in start of the application in main.c to enable debug logs.

Statistics

The main function will print statistics on the console every second. These statistics include the number of packets enqueued and dequeued at each stage in the application, and also key statistics per worker, including how many packets of each burst size (1-8) were sent to each worker thread.

Application Initialization

Command line parsing is done in the same way as it is done in the L2 Forwarding Sample Application. See Command Line Arguments.

Mbuf pool initialization is done in the same way as it is done in the L2 Forwarding Sample Application. See Mbuf Pool Initialization.

Driver Initialization is done in same way as it is done in the L2 Forwarding Sample Application. See Driver Initialization.

RX queue initialization is done in the same way as it is done in the L2 Forwarding Sample Application. See RX Queue Initialization.

TX queue initialization is done in the same way as it is done in the L2 Forwarding Sample Application. See TX Queue Initialization.

VM Power Management Application

Introduction

Applications running in Virtual Environments have an abstract view of the underlying hardware on the Host, in particular applications cannot see the binding of virtual to physical hardware. When looking at CPU resourcing, the pinning of Virtual CPUs(vCPUs) to Host Physical CPUs(pCPUS) is not apparent to an application and this pinning may change over time. Furthermore, Operating Systems on virtual machines do not have the ability to govern their own power policy; the Machine Specific Registers (MSRs) for enabling P-State transitions are not exposed to Operating Systems running on Virtual Machines(VMs).

The Virtual Machine Power Management solution shows an example of how a DPDK application can indicate its processing requirements using VM local only information(vCPU/lcore) to a Host based Monitor which is responsible for accepting requests for frequency changes for a vCPU, translating the vCPU to a pCPU via libvirt and affecting the change in frequency.

The solution is comprised of two high-level components:

  1. Example Host Application

    Using a Command Line Interface(CLI) for VM->Host communication channel management allows adding channels to the Monitor, setting and querying the vCPU to pCPU pinning, inspecting and manually changing the frequency for each CPU. The CLI runs on a single lcore while the thread responsible for managing VM requests runs on a second lcore.

    VM requests arriving on a channel for frequency changes are passed to the librte_power ACPI cpufreq sysfs based library. The Host Application relies on both qemu-kvm and libvirt to function.

  2. librte_power for Virtual Machines

    Using an alternate implementation for the librte_power API, requests for frequency changes are forwarded to the host monitor rather than the APCI cpufreq sysfs interface used on the host.

    The l3fwd-power application will use this implementation when deployed on a VM (see L3 Forwarding with Power Management Sample Application).

_images/vm_power_mgr_highlevel.svg

Highlevel Solution

Overview

VM Power Management employs qemu-kvm to provide communications channels between the host and VMs in the form of Virtio-Serial which appears as a paravirtualized serial device on a VM and can be configured to use various backends on the host. For this example each Virtio-Serial endpoint on the host is configured as AF_UNIX file socket, supporting poll/select and epoll for event notification. In this example each channel endpoint on the host is monitored via epoll for EPOLLIN events. Each channel is specified as qemu-kvm arguments or as libvirt XML for each VM, where each VM can have a number of channels up to a maximum of 64 per VM, in this example each DPDK lcore on a VM has exclusive access to a channel.

To enable frequency changes from within a VM, a request via the librte_power interface is forwarded via Virtio-Serial to the host, each request contains the vCPU and power command(scale up/down/min/max). The API for host and guest librte_power is consistent across environments, with the selection of VM or Host Implementation determined at automatically at runtime based on the environment.

Upon receiving a request, the host translates the vCPU to a pCPU via the libvirt API before forwarding to the host librte_power.

_images/vm_power_mgr_vm_request_seq.svg

VM request to scale frequency

Performance Considerations

While Haswell Microarchitecture allows for independent power control for each core, earlier Microarchtectures do not offer such fine grained control. When deployed on pre-Haswell platforms greater care must be taken in selecting which cores are assigned to a VM, for instance a core will not scale down until its sibling is similarly scaled.

Configuration

BIOS

Enhanced Intel SpeedStep® Technology must be enabled in the platform BIOS if the power management feature of DPDK is to be used. Otherwise, the sys file folder /sys/devices/system/cpu/cpu0/cpufreq will not exist, and the CPU frequency-based power management cannot be used. Consult the relevant BIOS documentation to determine how these settings can be accessed.

Host Operating System

The Host OS must also have the apci_cpufreq module installed, in some cases the intel_pstate driver may be the default Power Management environment. To enable acpi_cpufreq and disable intel_pstate, add the following to the grub Linux command line:

intel_pstate=disable

Upon rebooting, load the acpi_cpufreq module:

modprobe acpi_cpufreq
Hypervisor Channel Configuration

Virtio-Serial channels are configured via libvirt XML:

<name>{vm_name}</name>
<controller type='virtio-serial' index='0'>
  <address type='pci' domain='0x0000' bus='0x00' slot='0x06' function='0x0'/>
</controller>
<channel type='unix'>
  <source mode='bind' path='/tmp/powermonitor/{vm_name}.{channel_num}'/>
  <target type='virtio' name='virtio.serial.port.poweragent.{vm_channel_num}'/>
  <address type='virtio-serial' controller='0' bus='0' port='{N}'/>
</channel>

Where a single controller of type virtio-serial is created and up to 32 channels can be associated with a single controller and multiple controllers can be specified. The convention is to use the name of the VM in the host path {vm_name} and to increment {channel_num} for each channel, likewise the port value {N} must be incremented for each channel.

Each channel on the host will appear in path, the directory /tmp/powermonitor/ must first be created and given qemu permissions

mkdir /tmp/powermonitor/
chown qemu:qemu /tmp/powermonitor

Note that files and directories within /tmp are generally removed upon rebooting the host and the above steps may need to be carried out after each reboot.

The serial device as it appears on a VM is configured with the target element attribute name and must be in the form of virtio.serial.port.poweragent.{vm_channel_num}, where vm_channel_num is typically the lcore channel to be used in DPDK VM applications.

Each channel on a VM will be present at /dev/virtio-ports/virtio.serial.port.poweragent.{vm_channel_num}

Compiling and Running the Host Application

Compiling

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the vm_power_manager sub-directory.

Running

The application does not have any specific command line options other than EAL:

./build/vm_power_mgr [EAL options]

The application requires exactly two cores to run, one core is dedicated to the CLI, while the other is dedicated to the channel endpoint monitor, for example to run on cores 0 & 1 on a system with 4 memory channels:

./build/vm_power_mgr -l 0-1 -n 4

After successful initialization the user is presented with VM Power Manager CLI:

vm_power>

Virtual Machines can now be added to the VM Power Manager:

vm_power> add_vm {vm_name}

When a {vm_name} is specified with the add_vm command a lookup is performed with libvirt to ensure that the VM exists, {vm_name} is used as an unique identifier to associate channels with a particular VM and for executing operations on a VM within the CLI. VMs do not have to be running in order to add them.

A number of commands can be issued via the CLI in relation to VMs:

Remove a Virtual Machine identified by {vm_name} from the VM Power Manager.

rm_vm {vm_name}

Add communication channels for the specified VM, the virtio channels must be enabled in the VM configuration(qemu/libvirt) and the associated VM must be active. {list} is a comma-separated list of channel numbers to add, using the keyword ‘all’ will attempt to add all channels for the VM:

add_channels {vm_name} {list}|all

Enable or disable the communication channels in {list}(comma-separated) for the specified VM, alternatively list can be replaced with keyword ‘all’. Disabled channels will still receive packets on the host, however the commands they specify will be ignored. Set status to ‘enabled’ to begin processing requests again:

set_channel_status {vm_name} {list}|all enabled|disabled

Print to the CLI the information on the specified VM, the information lists the number of vCPUS, the pinning to pCPU(s) as a bit mask, along with any communication channels associated with each VM, along with the status of each channel:

show_vm {vm_name}

Set the binding of Virtual CPU on VM with name {vm_name} to the Physical CPU mask:

set_pcpu_mask {vm_name} {vcpu} {pcpu}

Set the binding of Virtual CPU on VM to the Physical CPU:

set_pcpu {vm_name} {vcpu} {pcpu}

Manual control and inspection can also be carried in relation CPU frequency scaling:

Get the current frequency for each core specified in the mask:

show_cpu_freq_mask {mask}

Set the current frequency for the cores specified in {core_mask} by scaling each up/down/min/max:

set_cpu_freq {core_mask} up|down|min|max

Get the current frequency for the specified core:

show_cpu_freq {core_num}

Set the current frequency for the specified core by scaling up/down/min/max:

set_cpu_freq {core_num} up|down|min|max

Compiling and Running the Guest Applications

For compiling and running l3fwd-power, see L3 Forwarding with Power Management Sample Application.

A guest CLI is also provided for validating the setup.

For both l3fwd-power and guest CLI, the channels for the VM must be monitored by the host application using the add_channels command on the host.

Compiling
  1. export RTE_SDK=/path/to/rte_sdk
  2. cd ${RTE_SDK}/examples/vm_power_manager/guest_cli
  3. make
Running

The application does not have any specific command line options other than EAL:

./build/vm_power_mgr [EAL options]

The application for example purposes uses a channel for each lcore enabled, for example to run on cores 0,1,2,3 on a system with 4 memory channels:

./build/guest_vm_power_mgr -l 0-3 -n 4

After successful initialization the user is presented with VM Power Manager Guest CLI:

vm_power(guest)>

To change the frequency of a lcore, use the set_cpu_freq command. Where {core_num} is the lcore and channel to change frequency by scaling up/down/min/max.

set_cpu_freq {core_num} up|down|min|max

TEP termination Sample Application

The TEP (Tunnel End point) termination sample application simulates a VXLAN Tunnel Endpoint (VTEP) termination in DPDK, which is used to demonstrate the offload and filtering capabilities of Intel® XL710 10/40 Gigabit Ethernet Controller for VXLAN packet. This sample uses the basic virtio devices management mechanism from vhost example, and also uses the us-vHost interface and tunnel filtering mechanism to direct a specified traffic to a specific VM. In addition, this sample is also designed to show how tunneling protocols can be handled.

Background

With virtualization, overlay networks allow a network structure to be built or imposed across physical nodes which is abstracted away from the actual underlining physical network connections. This allows network isolation, QOS, etc to be provided on a per client basis.

_images/overlay_networking.svg

Overlay Networking.

In a typical setup, the network overlay tunnel is terminated at the Virtual/Tunnel End Point (VEP/TEP). The TEP is normally located at the physical host level ideally in the software switch. Due to processing constraints and the inevitable bottleneck that the switch becomes, the ability to offload overlay support features becomes an important requirement. Intel® XL710 10/40 Gigabit Ethernet network card provides hardware filtering and offload capabilities to support overlay networks implementations such as MAC in UDP and MAC in GRE.

Sample Code Overview

The DPDK TEP termination sample code demonstrates the offload and filtering capabilities of Intel® XL710 10/40 Gigabit Ethernet Controller for VXLAN packet.

The sample code is based on vhost library. The vhost library is developed for user space Ethernet switch to easily integrate with vhost functionality.

The sample will support the followings:

  • Tunneling packet recognition.

  • The port of UDP tunneling is configurable

  • Directing incoming traffic to the correct queue based on the tunnel filter type. The supported filter type are listed below.

    • Inner MAC and VLAN and tenant ID
    • Inner MAC and tenant ID, and Outer MAC
    • Inner MAC and tenant ID

    The tenant ID will be assigned from a static internal table based on the us-vhost device ID. Each device will receive a unique device ID. The inner MAC will be learned by the first packet transmitted from a device.

  • Decapsulation of RX VXLAN traffic. This is a software only operation.

  • Encapsulation of TX VXLAN traffic. This is a software only operation.

  • Inner IP and inner L4 checksum offload.

  • TSO offload support for tunneling packet.

The following figure shows the framework of the TEP termination sample application based on DPDK vhost lib.

_images/tep_termination_arch.svg

TEP termination Framework Overview

Supported Distributions

The example in this section have been validated with the following distributions:

  • Fedora* 18
  • Fedora* 19
  • Fedora* 20

Compiling the Sample Code

To enable vhost, turn on vhost library in the configure file config/common_linuxapp.

CONFIG_RTE_LIBRTE_VHOST=y

Then following the to compile the sample application shown in Compiling the Sample Applications.

Running the Sample Code

  1. Go to the examples directory:

    export RTE_SDK=/path/to/rte_sdk
    cd ${RTE_SDK}/examples/tep_termination
    
  2. Run the tep_termination sample code:

    user@target:~$ ./build/app/tep_termination -l 0-3 -n 4 --huge-dir /mnt/huge --
                    -p 0x1 --dev-basename tep-termination --nb-devices 4
                    --udp-port 4789 --filter-type 1
    

Note

Please note the huge-dir parameter instructs the DPDK to allocate its memory from the 2 MB page hugetlbfs.

Parameters

The same parameters with the vhost sample.

Refer to Parameters for detailed explanation.

Number of Devices.

The nb-devices option specifies the number of virtIO device. The default value is 2.

user@target:~$ ./build/app/tep_termination -l 0-3 -n 4 --huge-dir /mnt/huge --
                --nb-devices 2

Tunneling UDP port.

The udp-port option is used to specify the destination UDP number for UDP tunneling packet. The default value is 4789.

user@target:~$ ./build/app/tep_termination -l 0-3 -n 4 --huge-dir /mnt/huge --
                --nb-devices 2 --udp-port 4789

Filter Type.

The filter-type option is used to specify which filter type is used to filter UDP tunneling packet to a specified queue. The default value is 1, which means the filter type of inner MAC and tenant ID is used.

user@target:~$ ./build/app/tep_termination -l 0-3 -n 4 --huge-dir /mnt/huge --
            --nb-devices 2 --udp-port 4789 --filter-type 1

TX Checksum.

The tx-checksum option is used to enable or disable the inner header checksum offload. The default value is 0, which means the checksum offload is disabled.

user@target:~$ ./build/app/tep_termination -l 0-3 -n 4 --huge-dir /mnt/huge --
            --nb-devices 2 --tx-checksum

TCP segment size.

The tso-segsz option specifies the TCP segment size for TSO offload for tunneling packet. The default value is 0, which means TSO offload is disabled.

user@target:~$ ./build/app/tep_termination -l 0-3 -n 4 --huge-dir /mnt/huge --
            --tx-checksum --tso-segsz 800

Decapsulation option.

The decap option is used to enable or disable decapsulation operation for received VXLAN packet. The default value is 1.

user@target:~$ ./build/app/tep_termination -l 0-3 -n 4 --huge-dir /mnt/huge --
            --nb-devices 4 --udp-port 4789 --decap 1

Encapsulation option.

The encap option is used to enable or disable encapsulation operation for transmitted packet. The default value is 1.

user@target:~$ ./build/app/tep_termination -l 0-3 -n 4 --huge-dir /mnt/huge --
            --nb-devices 4 --udp-port 4789 --encap 1

Running the Virtual Machine (QEMU)

Refer to Start the VM.

Running DPDK in the Virtual Machine

Refer to Run testpmd inside guest.

Passing Traffic to the Virtual Machine Device

For a virtio-net device to receive traffic, the traffic’s Layer 2 header must include both the virtio-net device’s MAC address. The DPDK sample code behaves in a similar manner to a learning switch in that it learns the MAC address of the virtio-net devices from the first transmitted packet. On learning the MAC address, the DPDK vhost sample code prints a message with the MAC address and tenant ID virtio-net device. For example:

DATA: (0) MAC_ADDRESS cc:bb:bb:bb:bb:bb and VNI 1000 registered

The above message indicates that device 0 has been registered with MAC address cc:bb:bb:bb:bb:bb and VNI 1000. Any packets received on the NIC with these values are placed on the devices receive queue.

PTP Client Sample Application

The PTP (Precision Time Protocol) client sample application is a simple example of using the DPDK IEEE1588 API to communicate with a PTP master clock to synchronize the time on the NIC and, optionally, on the Linux system.

Note, PTP is a time syncing protocol and cannot be used within DPDK as a time-stamping mechanism. See the following for an explanation of the protocol: Precision Time Protocol.

Limitations

The PTP sample application is intended as a simple reference implementation of a PTP client using the DPDK IEEE1588 API. In order to keep the application simple the following assumptions are made:

  • The first discovered master is the master for the session.
  • Only L2 PTP packets are supported.
  • Only the PTP v2 protocol is supported.
  • Only the slave clock is implemented.

How the Application Works

_images/ptpclient.svg

PTP Synchronization Protocol

The PTP synchronization in the sample application works as follows:

  • Master sends Sync message - the slave saves it as T2.
  • Master sends Follow Up message and sends time of T1.
  • Slave sends Delay Request frame to PTP Master and stores T3.
  • Master sends Delay Response T4 time which is time of received T3.

The adjustment for slave can be represented as:

adj = -[(T2-T1)-(T4 - T3)]/2

If the command line parameter -T 1 is used the application also synchronizes the PTP PHC clock with the Linux kernel clock.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the ptpclient sub-directory.

Note

To compile the application edit the config/common_linuxapp configuration file to enable IEEE1588 and then recompile DPDK:

CONFIG_RTE_LIBRTE_IEEE1588=y

Running the Application

To run the example in a linuxapp environment:

./build/ptpclient -l 1 -n 4 -- -p 0x1 -T 0

Refer to DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

  • -p portmask: Hexadecimal portmask.
  • -T 0: Update only the PTP slave clock.
  • -T 1: Update the PTP slave clock and synchronize the Linux Kernel to the PTP clock.

Code Explanation

The following sections provide an explanation of the main components of the code.

All DPDK library functions used in the sample code are prefixed with rte_ and are explained in detail in the DPDK API Documentation.

The Main Function

The main() function performs the initialization and calls the execution threads for each lcore.

The first task is to initialize the Environment Abstraction Layer (EAL). The argc and argv arguments are provided to the rte_eal_init() function. The value returned is the number of parsed arguments:

int ret = rte_eal_init(argc, argv);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Error with EAL initialization\n");

And than we parse application specific arguments

argc -= ret;
argv += ret;

ret = ptp_parse_args(argc, argv);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Error with PTP initialization\n");

The main() also allocates a mempool to hold the mbufs (Message Buffers) used by the application:

mbuf_pool = rte_pktmbuf_pool_create("MBUF_POOL", NUM_MBUFS * nb_ports,
       MBUF_CACHE_SIZE, 0, RTE_MBUF_DEFAULT_BUF_SIZE, rte_socket_id());

Mbufs are the packet buffer structure used by DPDK. They are explained in detail in the “Mbuf Library” section of the DPDK Programmer’s Guide.

The main() function also initializes all the ports using the user defined port_init() function with portmask provided by user:

for (portid = 0; portid < nb_ports; portid++)
    if ((ptp_enabled_port_mask & (1 << portid)) != 0) {

        if (port_init(portid, mbuf_pool) == 0) {
            ptp_enabled_ports[ptp_enabled_port_nb] = portid;
            ptp_enabled_port_nb++;
        } else {
            rte_exit(EXIT_FAILURE, "Cannot init port %"PRIu8 "\n",
                    portid);
        }
    }

Once the initialization is complete, the application is ready to launch a function on an lcore. In this example lcore_main() is called on a single lcore.

lcore_main();

The lcore_main() function is explained below.

The Lcores Main

As we saw above the main() function calls an application function on the available lcores.

The main work of the application is done within the loop:

for (portid = 0; portid < ptp_enabled_port_nb; portid++) {

    portid = ptp_enabled_ports[portid];
    nb_rx = rte_eth_rx_burst(portid, 0, &m, 1);

    if (likely(nb_rx == 0))
        continue;

    if (m->ol_flags & PKT_RX_IEEE1588_PTP)
        parse_ptp_frames(portid, m);

    rte_pktmbuf_free(m);
}

Packets are received one by one on the RX ports and, if required, PTP response packets are transmitted on the TX ports.

If the offload flags in the mbuf indicate that the packet is a PTP packet then the packet is parsed to determine which type:

if (m->ol_flags & PKT_RX_IEEE1588_PTP)
     parse_ptp_frames(portid, m);

All packets are freed explicitly using rte_pktmbuf_free().

The forwarding loop can be interrupted and the application closed using Ctrl-C.

PTP parsing

The parse_ptp_frames() function processes PTP packets, implementing slave PTP IEEE1588 L2 functionality.

void
parse_ptp_frames(uint16_t portid, struct rte_mbuf *m) {
    struct ptp_header *ptp_hdr;
    struct ether_hdr *eth_hdr;
    uint16_t eth_type;

    eth_hdr = rte_pktmbuf_mtod(m, struct ether_hdr *);
    eth_type = rte_be_to_cpu_16(eth_hdr->ether_type);

    if (eth_type == PTP_PROTOCOL) {
        ptp_data.m = m;
        ptp_data.portid = portid;
        ptp_hdr = (struct ptp_header *)(rte_pktmbuf_mtod(m, char *)
                    + sizeof(struct ether_hdr));

        switch (ptp_hdr->msgtype) {
        case SYNC:
            parse_sync(&ptp_data);
            break;
        case FOLLOW_UP:
            parse_fup(&ptp_data);
            break;
        case DELAY_RESP:
            parse_drsp(&ptp_data);
            print_clock_info(&ptp_data);
            break;
        default:
            break;
        }
    }
}

There are 3 types of packets on the RX path which we must parse to create a minimal implementation of the PTP slave client:

  • SYNC packet.
  • FOLLOW UP packet
  • DELAY RESPONSE packet.

When we parse the FOLLOW UP packet we also create and send a DELAY_REQUEST packet. Also when we parse the DELAY RESPONSE packet, and all conditions are met we adjust the PTP slave clock.

Performance Thread Sample Application

The performance thread sample application is a derivative of the standard L3 forwarding application that demonstrates different threading models.

Overview

For a general description of the L3 forwarding applications capabilities please refer to the documentation of the standard application in L3 Forwarding Sample Application.

The performance thread sample application differs from the standard L3 forwarding example in that it divides the TX and RX processing between different threads, and makes it possible to assign individual threads to different cores.

Three threading models are considered:

  1. When there is one EAL thread per physical core.
  2. When there are multiple EAL threads per physical core.
  3. When there are multiple lightweight threads per EAL thread.

Since DPDK release 2.0 it is possible to launch applications using the --lcores EAL parameter, specifying cpu-sets for a physical core. With the performance thread sample application its is now also possible to assign individual RX and TX functions to different cores.

As an alternative to dividing the L3 forwarding work between different EAL threads the performance thread sample introduces the possibility to run the application threads as lightweight threads (L-threads) within one or more EAL threads.

In order to facilitate this threading model the example includes a primitive cooperative scheduler (L-thread) subsystem. More details of the L-thread subsystem can be found in The L-thread subsystem.

Note: Whilst theoretically possible it is not anticipated that multiple L-thread schedulers would be run on the same physical core, this mode of operation should not be expected to yield useful performance and is considered invalid.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the performance-thread/l3fwd-thread sub-directory.

Running the Application

The application has a number of command line options:

./build/l3fwd-thread [EAL options] --
    -p PORTMASK [-P]
    --rx(port,queue,lcore,thread)[,(port,queue,lcore,thread)]
    --tx(lcore,thread)[,(lcore,thread)]
    [--enable-jumbo] [--max-pkt-len PKTLEN]]  [--no-numa]
    [--hash-entry-num] [--ipv6] [--no-lthreads] [--stat-lcore lcore]
    [--parse-ptype]

Where:

  • -p PORTMASK: Hexadecimal bitmask of ports to configure.
  • -P: optional, sets all ports to promiscuous mode so that packets are accepted regardless of the packet’s Ethernet MAC destination address. Without this option, only packets with the Ethernet MAC destination address set to the Ethernet address of the port are accepted.
  • --rx (port,queue,lcore,thread)[,(port,queue,lcore,thread)]: the list of NIC RX ports and queues handled by the RX lcores and threads. The parameters are explained below.
  • --tx (lcore,thread)[,(lcore,thread)]: the list of TX threads identifying the lcore the thread runs on, and the id of RX thread with which it is associated. The parameters are explained below.
  • --enable-jumbo: optional, enables jumbo frames.
  • --max-pkt-len: optional, maximum packet length in decimal (64-9600).
  • --no-numa: optional, disables numa awareness.
  • --hash-entry-num: optional, specifies the hash entry number in hex to be setup.
  • --ipv6: optional, set it if running ipv6 packets.
  • --no-lthreads: optional, disables l-thread model and uses EAL threading model. See below.
  • --stat-lcore: optional, run CPU load stats collector on the specified lcore.
  • --parse-ptype: optional, set to use software to analyze packet type. Without this option, hardware will check the packet type.

The parameters of the --rx and --tx options are:

  • --rx parameters

    port RX port
    queue RX queue that will be read on the specified RX port
    lcore Core to use for the thread
    thread Thread id (continuously from 0 to N)
  • --tx parameters

    lcore Core to use for L3 route match and transmit
    thread Id of RX thread to be associated with this TX thread

The l3fwd-thread application allows you to start packet processing in two threading models: L-Threads (default) and EAL Threads (when the --no-lthreads parameter is used). For consistency all parameters are used in the same way for both models.

Running with L-threads

When the L-thread model is used (default option), lcore and thread parameters in --rx/--tx are used to affinitize threads to the selected scheduler.

For example, the following places every l-thread on different lcores:

l3fwd-thread -l 0-7 -n 2 -- -P -p 3 \
             --rx="(0,0,0,0)(1,0,1,1)" \
             --tx="(2,0)(3,1)"

The following places RX l-threads on lcore 0 and TX l-threads on lcore 1 and 2 and so on:

l3fwd-thread -l 0-7 -n 2 -- -P -p 3 \
             --rx="(0,0,0,0)(1,0,0,1)" \
             --tx="(1,0)(2,1)"
Running with EAL threads

When the --no-lthreads parameter is used, the L-threading model is turned off and EAL threads are used for all processing. EAL threads are enumerated in the same way as L-threads, but the --lcores EAL parameter is used to affinitize threads to the selected cpu-set (scheduler). Thus it is possible to place every RX and TX thread on different lcores.

For example, the following places every EAL thread on different lcores:

l3fwd-thread -l 0-7 -n 2 -- -P -p 3 \
             --rx="(0,0,0,0)(1,0,1,1)" \
             --tx="(2,0)(3,1)" \
             --no-lthreads

To affinitize two or more EAL threads to one cpu-set, the EAL --lcores parameter is used.

The following places RX EAL threads on lcore 0 and TX EAL threads on lcore 1 and 2 and so on:

l3fwd-thread -l 0-7 -n 2 --lcores="(0,1)@0,(2,3)@1" -- -P -p 3 \
             --rx="(0,0,0,0)(1,0,1,1)" \
             --tx="(2,0)(3,1)" \
             --no-lthreads
Examples

For selected scenarios the command line configuration of the application for L-threads and its corresponding EAL threads command line can be realized as follows:

  1. Start every thread on different scheduler (1:1):

    l3fwd-thread -l 0-7 -n 2 -- -P -p 3 \
                 --rx="(0,0,0,0)(1,0,1,1)" \
                 --tx="(2,0)(3,1)"
    

    EAL thread equivalent:

    l3fwd-thread -l 0-7 -n 2 -- -P -p 3 \
                 --rx="(0,0,0,0)(1,0,1,1)" \
                 --tx="(2,0)(3,1)" \
                 --no-lthreads
    
  2. Start all threads on one core (N:1).

    Start 4 L-threads on lcore 0:

    l3fwd-thread -l 0-7 -n 2 -- -P -p 3 \
                 --rx="(0,0,0,0)(1,0,0,1)" \
                 --tx="(0,0)(0,1)"
    

    Start 4 EAL threads on cpu-set 0:

    l3fwd-thread -l 0-7 -n 2 --lcores="(0-3)@0" -- -P -p 3 \
                 --rx="(0,0,0,0)(1,0,0,1)" \
                 --tx="(2,0)(3,1)" \
                 --no-lthreads
    
  3. Start threads on different cores (N:M).

    Start 2 L-threads for RX on lcore 0, and 2 L-threads for TX on lcore 1:

    l3fwd-thread -l 0-7 -n 2 -- -P -p 3 \
                 --rx="(0,0,0,0)(1,0,0,1)" \
                 --tx="(1,0)(1,1)"
    

    Start 2 EAL threads for RX on cpu-set 0, and 2 EAL threads for TX on cpu-set 1:

    l3fwd-thread -l 0-7 -n 2 --lcores="(0-1)@0,(2-3)@1" -- -P -p 3 \
                 --rx="(0,0,0,0)(1,0,1,1)" \
                 --tx="(2,0)(3,1)" \
                 --no-lthreads
    

Explanation

To a great extent the sample application differs little from the standard L3 forwarding application, and readers are advised to familiarize themselves with the material covered in the L3 Forwarding Sample Application documentation before proceeding.

The following explanation is focused on the way threading is handled in the performance thread example.

Mode of operation with EAL threads

The performance thread sample application has split the RX and TX functionality into two different threads, and the RX and TX threads are interconnected via software rings. With respect to these rings the RX threads are producers and the TX threads are consumers.

On initialization the TX and RX threads are started according to the command line parameters.

The RX threads poll the network interface queues and post received packets to a TX thread via a corresponding software ring.

The TX threads poll software rings, perform the L3 forwarding hash/LPM match, and assemble packet bursts before performing burst transmit on the network interface.

As with the standard L3 forward application, burst draining of residual packets is performed periodically with the period calculated from elapsed time using the timestamps counter.

The diagram below illustrates a case with two RX threads and three TX threads.

_images/performance_thread_1.svg
Mode of operation with L-threads

Like the EAL thread configuration the application has split the RX and TX functionality into different threads, and the pairs of RX and TX threads are interconnected via software rings.

On initialization an L-thread scheduler is started on every EAL thread. On all but the master EAL thread only a a dummy L-thread is initially started. The L-thread started on the master EAL thread then spawns other L-threads on different L-thread schedulers according the the command line parameters.

The RX threads poll the network interface queues and post received packets to a TX thread via the corresponding software ring.

The ring interface is augmented by means of an L-thread condition variable that enables the TX thread to be suspended when the TX ring is empty. The RX thread signals the condition whenever it posts to the TX ring, causing the TX thread to be resumed.

Additionally the TX L-thread spawns a worker L-thread to take care of polling the software rings, whilst it handles burst draining of the transmit buffer.

The worker threads poll the software rings, perform L3 route lookup and assemble packet bursts. If the TX ring is empty the worker thread suspends itself by waiting on the condition variable associated with the ring.

Burst draining of residual packets, less than the burst size, is performed by the TX thread which sleeps (using an L-thread sleep function) and resumes periodically to flush the TX buffer.

This design means that L-threads that have no work, can yield the CPU to other L-threads and avoid having to constantly poll the software rings.

The diagram below illustrates a case with two RX threads and three TX functions (each comprising a thread that processes forwarding and a thread that periodically drains the output buffer of residual packets).

_images/performance_thread_2.svg
CPU load statistics

It is possible to display statistics showing estimated CPU load on each core. The statistics indicate the percentage of CPU time spent: processing received packets (forwarding), polling queues/rings (waiting for work), and doing any other processing (context switch and other overhead).

When enabled statistics are gathered by having the application threads set and clear flags when they enter and exit pertinent code sections. The flags are then sampled in real time by a statistics collector thread running on another core. This thread displays the data in real time on the console.

This feature is enabled by designating a statistics collector core, using the --stat-lcore parameter.

The L-thread subsystem

The L-thread subsystem resides in the examples/performance-thread/common directory and is built and linked automatically when building the l3fwd-thread example.

The subsystem provides a simple cooperative scheduler to enable arbitrary functions to run as cooperative threads within a single EAL thread. The subsystem provides a pthread like API that is intended to assist in reuse of legacy code written for POSIX pthreads.

The following sections provide some detail on the features, constraints, performance and porting considerations when using L-threads.

Comparison between L-threads and POSIX pthreads

The fundamental difference between the L-thread and pthread models is the way in which threads are scheduled. The simplest way to think about this is to consider the case of a processor with a single CPU. To run multiple threads on a single CPU, the scheduler must frequently switch between the threads, in order that each thread is able to make timely progress. This is the basis of any multitasking operating system.

This section explores the differences between the pthread model and the L-thread model as implemented in the provided L-thread subsystem. If needed a theoretical discussion of preemptive vs cooperative multi-threading can be found in any good text on operating system design.

Scheduling and context switching

The POSIX pthread library provides an application programming interface to create and synchronize threads. Scheduling policy is determined by the host OS, and may be configurable. The OS may use sophisticated rules to determine which thread should be run next, threads may suspend themselves or make other threads ready, and the scheduler may employ a time slice giving each thread a maximum time quantum after which it will be preempted in favor of another thread that is ready to run. To complicate matters further threads may be assigned different scheduling priorities.

By contrast the L-thread subsystem is considerably simpler. Logically the L-thread scheduler performs the same multiplexing function for L-threads within a single pthread as the OS scheduler does for pthreads within an application process. The L-thread scheduler is simply the main loop of a pthread, and in so far as the host OS is concerned it is a regular pthread just like any other. The host OS is oblivious about the existence of and not at all involved in the scheduling of L-threads.

The other and most significant difference between the two models is that L-threads are scheduled cooperatively. L-threads cannot not preempt each other, nor can the L-thread scheduler preempt a running L-thread (i.e. there is no time slicing). The consequence is that programs implemented with L-threads must possess frequent rescheduling points, meaning that they must explicitly and of their own volition return to the scheduler at frequent intervals, in order to allow other L-threads an opportunity to proceed.

In both models switching between threads requires that the current CPU context is saved and a new context (belonging to the next thread ready to run) is restored. With pthreads this context switching is handled transparently and the set of CPU registers that must be preserved between context switches is as per an interrupt handler.

An L-thread context switch is achieved by the thread itself making a function call to the L-thread scheduler. Thus it is only necessary to preserve the callee registers. The caller is responsible to save and restore any other registers it is using before a function call, and restore them on return, and this is handled by the compiler. For X86_64 on both Linux and BSD the System V calling convention is used, this defines registers RSP, RBP, and R12-R15 as callee-save registers (for more detailed discussion a good reference is X86 Calling Conventions).

Taking advantage of this, and due to the absence of preemption, an L-thread context switch is achieved with less than 20 load/store instructions.

The scheduling policy for L-threads is fixed, there is no prioritization of L-threads, all L-threads are equal and scheduling is based on a FIFO ready queue.

An L-thread is a struct containing the CPU context of the thread (saved on context switch) and other useful items. The ready queue contains pointers to threads that are ready to run. The L-thread scheduler is a simple loop that polls the ready queue, reads from it the next thread ready to run, which it resumes by saving the current context (the current position in the scheduler loop) and restoring the context of the next thread from its thread struct. Thus an L-thread is always resumed at the last place it yielded.

A well behaved L-thread will call the context switch regularly (at least once in its main loop) thus returning to the scheduler’s own main loop. Yielding inserts the current thread at the back of the ready queue, and the process of servicing the ready queue is repeated, thus the system runs by flipping back and forth the between L-threads and scheduler loop.

In the case of pthreads, the preemptive scheduling, time slicing, and support for thread prioritization means that progress is normally possible for any thread that is ready to run. This comes at the price of a relatively heavier context switch and scheduling overhead.

With L-threads the progress of any particular thread is determined by the frequency of rescheduling opportunities in the other L-threads. This means that an errant L-thread monopolizing the CPU might cause scheduling of other threads to be stalled. Due to the lower cost of context switching, however, voluntary rescheduling to ensure progress of other threads, if managed sensibly, is not a prohibitive overhead, and overall performance can exceed that of an application using pthreads.

Mutual exclusion

With pthreads preemption means that threads that share data must observe some form of mutual exclusion protocol.

The fact that L-threads cannot preempt each other means that in many cases mutual exclusion devices can be completely avoided.

Locking to protect shared data can be a significant bottleneck in multi-threaded applications so a carefully designed cooperatively scheduled program can enjoy significant performance advantages.

So far we have considered only the simplistic case of a single core CPU, when multiple CPUs are considered things are somewhat more complex.

First of all it is inevitable that there must be multiple L-thread schedulers, one running on each EAL thread. So long as these schedulers remain isolated from each other the above assertions about the potential advantages of cooperative scheduling hold true.

A configuration with isolated cooperative schedulers is less flexible than the pthread model where threads can be affinitized to run on any CPU. With isolated schedulers scaling of applications to utilize fewer or more CPUs according to system demand is very difficult to achieve.

The L-thread subsystem makes it possible for L-threads to migrate between schedulers running on different CPUs. Needless to say if the migration means that threads that share data end up running on different CPUs then this will introduce the need for some kind of mutual exclusion system.

Of course rte_ring software rings can always be used to interconnect threads running on different cores, however to protect other kinds of shared data structures, lock free constructs or else explicit locking will be required. This is a consideration for the application design.

In support of this extended functionality, the L-thread subsystem implements thread safe mutexes and condition variables.

The cost of affinitizing and of condition variable signaling is significantly lower than the equivalent pthread operations, and so applications using these features will see a performance benefit.

Thread local storage

As with applications written for pthreads an application written for L-threads can take advantage of thread local storage, in this case local to an L-thread. An application may save and retrieve a single pointer to application data in the L-thread struct.

For legacy and backward compatibility reasons two alternative methods are also offered, the first is modelled directly on the pthread get/set specific APIs, the second approach is modelled on the RTE_PER_LCORE macros, whereby PER_LTHREAD macros are introduced, in both cases the storage is local to the L-thread.

Constraints and performance implications when using L-threads
API compatibility

The L-thread subsystem provides a set of functions that are logically equivalent to the corresponding functions offered by the POSIX pthread library, however not all pthread functions have a corresponding L-thread equivalent, and not all features available to pthreads are implemented for L-threads.

The pthread library offers considerable flexibility via programmable attributes that can be associated with threads, mutexes, and condition variables.

By contrast the L-thread subsystem has fixed functionality, the scheduler policy cannot be varied, and L-threads cannot be prioritized. There are no variable attributes associated with any L-thread objects. L-threads, mutexes and conditional variables, all have fixed functionality. (Note: reserved parameters are included in the APIs to facilitate possible future support for attributes).

The table below lists the pthread and equivalent L-thread APIs with notes on differences and/or constraints. Where there is no L-thread entry in the table, then the L-thread subsystem provides no equivalent function.

Pthread and equivalent L-thread APIs.
Pthread function L-thread function Notes
pthread_barrier_destroy    
pthread_barrier_init    
pthread_barrier_wait    
pthread_cond_broadcast lthread_cond_broadcast See note 1
pthread_cond_destroy lthread_cond_destroy  
pthread_cond_init lthread_cond_init  
pthread_cond_signal lthread_cond_signal See note 1
pthread_cond_timedwait    
pthread_cond_wait lthread_cond_wait See note 5
pthread_create lthread_create See notes 2, 3
pthread_detach lthread_detach See note 4
pthread_equal    
pthread_exit lthread_exit  
pthread_getspecific lthread_getspecific  
pthread_getcpuclockid    
pthread_join lthread_join  
pthread_key_create lthread_key_create  
pthread_key_delete lthread_key_delete  
pthread_mutex_destroy lthread_mutex_destroy  
pthread_mutex_init lthread_mutex_init  
pthread_mutex_lock lthread_mutex_lock See note 6
pthread_mutex_trylock lthread_mutex_trylock See note 6
pthread_mutex_timedlock    
pthread_mutex_unlock lthread_mutex_unlock  
pthread_once    
pthread_rwlock_destroy    
pthread_rwlock_init    
pthread_rwlock_rdlock    
pthread_rwlock_timedrdlock    
pthread_rwlock_timedwrlock    
pthread_rwlock_tryrdlock    
pthread_rwlock_trywrlock    
pthread_rwlock_unlock    
pthread_rwlock_wrlock    
pthread_self lthread_current  
pthread_setspecific lthread_setspecific  
pthread_spin_init   See note 10
pthread_spin_destroy   See note 10
pthread_spin_lock   See note 10
pthread_spin_trylock   See note 10
pthread_spin_unlock   See note 10
pthread_cancel lthread_cancel  
pthread_setcancelstate    
pthread_setcanceltype    
pthread_testcancel    
pthread_getschedparam    
pthread_setschedparam    
pthread_yield lthread_yield See note 7
pthread_setaffinity_np lthread_set_affinity See notes 2, 3, 8
  lthread_sleep See note 9
  lthread_sleep_clks See note 9

Note 1:

Neither lthread signal nor broadcast may be called concurrently by L-threads running on different schedulers, although multiple L-threads running in the same scheduler may freely perform signal or broadcast operations. L-threads running on the same or different schedulers may always safely wait on a condition variable.

Note 2:

Pthread attributes may be used to affinitize a pthread with a cpu-set. The L-thread subsystem does not support a cpu-set. An L-thread may be affinitized only with a single CPU at any time.

Note 3:

If an L-thread is intended to run on a different NUMA node than the node that creates the thread then, when calling lthread_create() it is advantageous to specify the destination core as a parameter of lthread_create(). See Memory allocation and NUMA awareness for details.

Note 4:

An L-thread can only detach itself, and cannot detach other L-threads.

Note 5:

A wait operation on a pthread condition variable is always associated with and protected by a mutex which must be owned by the thread at the time it invokes pthread_wait(). By contrast L-thread condition variables are thread safe (for waiters) and do not use an associated mutex. Multiple L-threads (including L-threads running on other schedulers) can safely wait on a L-thread condition variable. As a consequence the performance of an L-thread condition variables is typically an order of magnitude faster than its pthread counterpart.

Note 6:

Recursive locking is not supported with L-threads, attempts to take a lock recursively will be detected and rejected.

Note 7:

lthread_yield() will save the current context, insert the current thread to the back of the ready queue, and resume the next ready thread. Yielding increases ready queue backlog, see Ready queue backlog for more details about the implications of this.

N.B. The context switch time as measured from immediately before the call to lthread_yield() to the point at which the next ready thread is resumed, can be an order of magnitude faster that the same measurement for pthread_yield.

Note 8:

lthread_set_affinity() is similar to a yield apart from the fact that the yielding thread is inserted into a peer ready queue of another scheduler. The peer ready queue is actually a separate thread safe queue, which means that threads appearing in the peer ready queue can jump any backlog in the local ready queue on the destination scheduler.

The context switch time as measured from the time just before the call to lthread_set_affinity() to just after the same thread is resumed on the new scheduler can be orders of magnitude faster than the same measurement for pthread_setaffinity_np().

Note 9:

Although there is no pthread_sleep() function, lthread_sleep() and lthread_sleep_clks() can be used wherever sleep(), usleep() or nanosleep() might ordinarily be used. The L-thread sleep functions suspend the current thread, start an rte_timer and resume the thread when the timer matures. The rte_timer_manage() entry point is called on every pass of the scheduler loop. This means that the worst case jitter on timer expiry is determined by the longest period between context switches of any running L-threads.

In a synthetic test with many threads sleeping and resuming then the measured jitter is typically orders of magnitude lower than the same measurement made for nanosleep().

Note 10:

Spin locks are not provided because they are problematical in a cooperative environment, see Locks and spinlocks for a more detailed discussion on how to avoid spin locks.

Thread local storage

Of the three L-thread local storage options the simplest and most efficient is storing a single application data pointer in the L-thread struct.

The PER_LTHREAD macros involve a run time computation to obtain the address of the variable being saved/retrieved and also require that the accesses are de-referenced via a pointer. This means that code that has used RTE_PER_LCORE macros being ported to L-threads might need some slight adjustment (see Thread local storage for hints about porting code that makes use of thread local storage).

The get/set specific APIs are consistent with their pthread counterparts both in use and in performance.

Memory allocation and NUMA awareness

All memory allocation is from DPDK huge pages, and is NUMA aware. Each scheduler maintains its own caches of objects: lthreads, their stacks, TLS, mutexes and condition variables. These caches are implemented as unbounded lock free MPSC queues. When objects are created they are always allocated from the caches on the local core (current EAL thread).

If an L-thread has been affinitized to a different scheduler, then it can always safely free resources to the caches from which they originated (because the caches are MPSC queues).

If the L-thread has been affinitized to a different NUMA node then the memory resources associated with it may incur longer access latency.

The commonly used pattern of setting affinity on entry to a thread after it has started, means that memory allocation for both the stack and TLS will have been made from caches on the NUMA node on which the threads creator is running. This has the side effect that access latency will be sub-optimal after affinitizing.

This side effect can be mitigated to some extent (although not completely) by specifying the destination CPU as a parameter of lthread_create() this causes the L-thread’s stack and TLS to be allocated when it is first scheduled on the destination scheduler, if the destination is a on another NUMA node it results in a more optimal memory allocation.

Note that the lthread struct itself remains allocated from memory on the creating node, this is unavoidable because an L-thread is known everywhere by the address of this struct.

Object cache sizing

The per lcore object caches pre-allocate objects in bulk whenever a request to allocate an object finds a cache empty. By default 100 objects are pre-allocated, this is defined by LTHREAD_PREALLOC in the public API header file lthread_api.h. This means that the caches constantly grow to meet system demand.

In the present implementation there is no mechanism to reduce the cache sizes if system demand reduces. Thus the caches will remain at their maximum extent indefinitely.

A consequence of the bulk pre-allocation of objects is that every 100 (default value) additional new object create operations results in a call to rte_malloc(). For creation of objects such as L-threads, which trigger the allocation of even more objects (i.e. their stacks and TLS) then this can cause outliers in scheduling performance.

If this is a problem the simplest mitigation strategy is to dimension the system, by setting the bulk object pre-allocation size to some large number that you do not expect to be exceeded. This means the caches will be populated once only, the very first time a thread is created.

Ready queue backlog

One of the more subtle performance considerations is managing the ready queue backlog. The fewer threads that are waiting in the ready queue then the faster any particular thread will get serviced.

In a naive L-thread application with N L-threads simply looping and yielding, this backlog will always be equal to the number of L-threads, thus the cost of a yield to a particular L-thread will be N times the context switch time.

This side effect can be mitigated by arranging for threads to be suspended and wait to be resumed, rather than polling for work by constantly yielding. Blocking on a mutex or condition variable or even more obviously having a thread sleep if it has a low frequency workload are all mechanisms by which a thread can be excluded from the ready queue until it really does need to be run. This can have a significant positive impact on performance.

Initialization, shutdown and dependencies

The L-thread subsystem depends on DPDK for huge page allocation and depends on the rte_timer subsystem. The DPDK EAL initialization and rte_timer_subsystem_init() MUST be completed before the L-thread sub system can be used.

Thereafter initialization of the L-thread subsystem is largely transparent to the application. Constructor functions ensure that global variables are properly initialized. Other than global variables each scheduler is initialized independently the first time that an L-thread is created by a particular EAL thread.

If the schedulers are to be run as isolated and independent schedulers, with no intention that L-threads running on different schedulers will migrate between schedulers or synchronize with L-threads running on other schedulers, then initialization consists simply of creating an L-thread, and then running the L-thread scheduler.

If there will be interaction between L-threads running on different schedulers, then it is important that the starting of schedulers on different EAL threads is synchronized.

To achieve this an additional initialization step is necessary, this is simply to set the number of schedulers by calling the API function lthread_num_schedulers_set(n), where n is the number of EAL threads that will run L-thread schedulers. Setting the number of schedulers to a number greater than 0 will cause all schedulers to wait until the others have started before beginning to schedule L-threads.

The L-thread scheduler is started by calling the function lthread_run() and should be called from the EAL thread and thus become the main loop of the EAL thread.

The function lthread_run(), will not return until all threads running on the scheduler have exited, and the scheduler has been explicitly stopped by calling lthread_scheduler_shutdown(lcore) or lthread_scheduler_shutdown_all().

All these function do is tell the scheduler that it can exit when there are no longer any running L-threads, neither function forces any running L-thread to terminate. Any desired application shutdown behavior must be designed and built into the application to ensure that L-threads complete in a timely manner.

Important Note: It is assumed when the scheduler exits that the application is terminating for good, the scheduler does not free resources before exiting and running the scheduler a subsequent time will result in undefined behavior.

Porting legacy code to run on L-threads

Legacy code originally written for a pthread environment may be ported to L-threads if the considerations about differences in scheduling policy, and constraints discussed in the previous sections can be accommodated.

This section looks in more detail at some of the issues that may have to be resolved when porting code.

pthread API compatibility

The first step is to establish exactly which pthread APIs the legacy application uses, and to understand the requirements of those APIs. If there are corresponding L-lthread APIs, and where the default pthread functionality is used by the application then, notwithstanding the other issues discussed here, it should be feasible to run the application with L-threads. If the legacy code modifies the default behavior using attributes then if may be necessary to make some adjustments to eliminate those requirements.

Blocking system API calls

It is important to understand what other system services the application may be using, bearing in mind that in a cooperatively scheduled environment a thread cannot block without stalling the scheduler and with it all other cooperative threads. Any kind of blocking system call, for example file or socket IO, is a potential problem, a good tool to analyze the application for this purpose is the strace utility.

There are many strategies to resolve these kind of issues, each with it merits. Possible solutions include:

  • Adopting a polled mode of the system API concerned (if available).
  • Arranging for another core to perform the function and synchronizing with that core via constructs that will not block the L-thread.
  • Affinitizing the thread to another scheduler devoted (as a matter of policy) to handling threads wishing to make blocking calls, and then back again when finished.
Locks and spinlocks

Locks and spinlocks are another source of blocking behavior that for the same reasons as system calls will need to be addressed.

If the application design ensures that the contending L-threads will always run on the same scheduler then it its probably safe to remove locks and spin locks completely.

The only exception to the above rule is if for some reason the code performs any kind of context switch whilst holding the lock (e.g. yield, sleep, or block on a different lock, or on a condition variable). This will need to determined before deciding to eliminate a lock.

If a lock cannot be eliminated then an L-thread mutex can be substituted for either kind of lock.

An L-thread blocking on an L-thread mutex will be suspended and will cause another ready L-thread to be resumed, thus not blocking the scheduler. When default behavior is required, it can be used as a direct replacement for a pthread mutex lock.

Spin locks are typically used when lock contention is likely to be rare and where the period during which the lock may be held is relatively short. When the contending L-threads are running on the same scheduler then an L-thread blocking on a spin lock will enter an infinite loop stopping the scheduler completely (see Infinite loops below).

If the application design ensures that contending L-threads will always run on different schedulers then it might be reasonable to leave a short spin lock that rarely experiences contention in place.

If after all considerations it appears that a spin lock can neither be eliminated completely, replaced with an L-thread mutex, or left in place as is, then an alternative is to loop on a flag, with a call to lthread_yield() inside the loop (n.b. if the contending L-threads might ever run on different schedulers the flag will need to be manipulated atomically).

Spinning and yielding is the least preferred solution since it introduces ready queue backlog (see also Ready queue backlog).

Sleeps and delays

Yet another kind of blocking behavior (albeit momentary) are delay functions like sleep(), usleep(), nanosleep() etc. All will have the consequence of stalling the L-thread scheduler and unless the delay is very short (e.g. a very short nanosleep) calls to these functions will need to be eliminated.

The simplest mitigation strategy is to use the L-thread sleep API functions, of which two variants exist, lthread_sleep() and lthread_sleep_clks(). These functions start an rte_timer against the L-thread, suspend the L-thread and cause another ready L-thread to be resumed. The suspended L-thread is resumed when the rte_timer matures.

Infinite loops

Some applications have threads with loops that contain no inherent rescheduling opportunity, and rely solely on the OS time slicing to share the CPU. In a cooperative environment this will stop everything dead. These kind of loops are not hard to identify, in a debug session you will find the debugger is always stopping in the same loop.

The simplest solution to this kind of problem is to insert an explicit lthread_yield() or lthread_sleep() into the loop. Another solution might be to include the function performed by the loop into the execution path of some other loop that does in fact yield, if this is possible.

Thread local storage

If the application uses thread local storage, the use case should be studied carefully.

In a legacy pthread application either or both the __thread prefix, or the pthread set/get specific APIs may have been used to define storage local to a pthread.

In some applications it may be a reasonable assumption that the data could or in fact most likely should be placed in L-thread local storage.

If the application (like many DPDK applications) has assumed a certain relationship between a pthread and the CPU to which it is affinitized, there is a risk that thread local storage may have been used to save some data items that are correctly logically associated with the CPU, and others items which relate to application context for the thread. Only a good understanding of the application will reveal such cases.

If the application requires an that an L-thread is to be able to move between schedulers then care should be taken to separate these kinds of data, into per lcore, and per L-thread storage. In this way a migrating thread will bring with it the local data it needs, and pick up the new logical core specific values from pthread local storage at its new home.

Pthread shim

A convenient way to get something working with legacy code can be to use a shim that adapts pthread API calls to the corresponding L-thread ones. This approach will not mitigate any of the porting considerations mentioned in the previous sections, but it will reduce the amount of code churn that would otherwise been involved. It is a reasonable approach to evaluate L-threads, before investing effort in porting to the native L-thread APIs.

Overview

The L-thread subsystem includes an example pthread shim. This is a partial implementation but does contain the API stubs needed to get basic applications running. There is a simple “hello world” application that demonstrates the use of the pthread shim.

A subtlety of working with a shim is that the application will still need to make use of the genuine pthread library functions, at the very least in order to create the EAL threads in which the L-thread schedulers will run. This is the case with DPDK initialization, and exit.

To deal with the initialization and shutdown scenarios, the shim is capable of switching on or off its adaptor functionality, an application can control this behavior by the calling the function pt_override_set(). The default state is disabled.

The pthread shim uses the dynamic linker loader and saves the loaded addresses of the genuine pthread API functions in an internal table, when the shim functionality is enabled it performs the adaptor function, when disabled it invokes the genuine pthread function.

The function pthread_exit() has additional special handling. The standard system header file pthread.h declares pthread_exit() with __attribute__((noreturn)) this is an optimization that is possible because the pthread is terminating and this enables the compiler to omit the normal handling of stack and protection of registers since the function is not expected to return, and in fact the thread is being destroyed. These optimizations are applied in both the callee and the caller of the pthread_exit() function.

In our cooperative scheduling environment this behavior is inadmissible. The pthread is the L-thread scheduler thread, and, although an L-thread is terminating, there must be a return to the scheduler in order that the system can continue to run. Further, returning from a function with attribute noreturn is invalid and may result in undefined behavior.

The solution is to redefine the pthread_exit function with a macro, causing it to be mapped to a stub function in the shim that does not have the noreturn attribute. This macro is defined in the file pthread_shim.h. The stub function is otherwise no different than any of the other stub functions in the shim, and will switch between the real pthread_exit() function or the lthread_exit() function as required. The only difference is that the mapping to the stub by macro substitution.

A consequence of this is that the file pthread_shim.h must be included in legacy code wishing to make use of the shim. It also means that dynamic linkage of a pre-compiled binary that did not include pthread_shim.h is not be supported.

Given the requirements for porting legacy code outlined in Porting legacy code to run on L-threads most applications will require at least some minimal adjustment and recompilation to run on L-threads so pre-compiled binaries are unlikely to be met in practice.

In summary the shim approach adds some overhead but can be a useful tool to help establish the feasibility of a code reuse project. It is also a fairly straightforward task to extend the shim if necessary.

Note: Bearing in mind the preceding discussions about the impact of making blocking calls then switching the shim in and out on the fly to invoke any pthread API this might block is something that should typically be avoided.

Building and running the pthread shim

The shim example application is located in the sample application in the performance-thread folder

To build and run the pthread shim example

  1. Go to the example applications folder

    export RTE_SDK=/path/to/rte_sdk
    cd ${RTE_SDK}/examples/performance-thread/pthread_shim
    
  2. Set the target (a default target is used if not specified). For example:

    export RTE_TARGET=x86_64-native-linuxapp-gcc
    

    See the DPDK Getting Started Guide for possible RTE_TARGET values.

  3. Build the application:

    make
    
  4. To run the pthread_shim example

    lthread-pthread-shim -c core_mask -n number_of_channels
    
L-thread Diagnostics

When debugging you must take account of the fact that the L-threads are run in a single pthread. The current scheduler is defined by RTE_PER_LCORE(this_sched), and the current lthread is stored at RTE_PER_LCORE(this_sched)->current_lthread. Thus on a breakpoint in a GDB session the current lthread can be obtained by displaying the pthread local variable per_lcore_this_sched->current_lthread.

Another useful diagnostic feature is the possibility to trace significant events in the life of an L-thread, this feature is enabled by changing the value of LTHREAD_DIAG from 0 to 1 in the file lthread_diag_api.h.

Tracing of events can be individually masked, and the mask may be programmed at run time. An unmasked event results in a callback that provides information about the event. The default callback simply prints trace information. The default mask is 0 (all events off) the mask can be modified by calling the function lthread_diagniostic_set_mask().

It is possible register a user callback function to implement more sophisticated diagnostic functions. Object creation events (lthread, mutex, and condition variable) accept, and store in the created object, a user supplied reference value returned by the callback function.

The lthread reference value is passed back in all subsequent event callbacks, the mutex and APIs are provided to retrieve the reference value from mutexes and condition variables. This enables a user to monitor, count, or filter for specific events, on specific objects, for example to monitor for a specific thread signaling a specific condition variable, or to monitor on all timer events, the possibilities and combinations are endless.

The callback function can be set by calling the function lthread_diagnostic_enable() supplying a callback function pointer and an event mask.

Setting LTHREAD_DIAG also enables counting of statistics about cache and queue usage, and these statistics can be displayed by calling the function lthread_diag_stats_display(). This function also performs a consistency check on the caches and queues. The function should only be called from the master EAL thread after all slave threads have stopped and returned to the C main program, otherwise the consistency check will fail.

IPsec Security Gateway Sample Application

The IPsec Security Gateway application is an example of a “real world” application using DPDK cryptodev framework.

Overview

The application demonstrates the implementation of a Security Gateway (not IPsec compliant, see the Constraints section below) using DPDK based on RFC4301, RFC4303, RFC3602 and RFC2404.

Internet Key Exchange (IKE) is not implemented, so only manual setting of Security Policies and Security Associations is supported.

The Security Policies (SP) are implemented as ACL rules, the Security Associations (SA) are stored in a table and the routing is implemented using LPM.

The application classifies the ports as Protected and Unprotected. Thus, traffic received on an Unprotected or Protected port is consider Inbound or Outbound respectively.

The application also supports complete IPSec protocol offload to hardware (Look aside crypto accelarator or using ethernet device). It also support inline ipsec processing by the supported ethernet device during transmission. These modes can be selected during the SA creation configuration.

In case of complete protocol offload, the processing of headers(ESP and outer IP header) is done by the hardware and the application does not need to add/remove them during outbound/inbound processing.

The Path for IPsec Inbound traffic is:

  • Read packets from the port.
  • Classify packets between IPv4 and ESP.
  • Perform Inbound SA lookup for ESP packets based on their SPI.
  • Perform Verification/Decryption (Not needed in case of inline ipsec).
  • Remove ESP and outer IP header (Not needed in case of protocol offload).
  • Inbound SP check using ACL of decrypted packets and any other IPv4 packets.
  • Routing.
  • Write packet to port.

The Path for the IPsec Outbound traffic is:

  • Read packets from the port.
  • Perform Outbound SP check using ACL of all IPv4 traffic.
  • Perform Outbound SA lookup for packets that need IPsec protection.
  • Add ESP and outer IP header (Not needed in case protocol offload).
  • Perform Encryption/Digest (Not needed in case of inline ipsec).
  • Routing.
  • Write packet to port.

Constraints

  • No IPv6 options headers.
  • No AH mode.
  • Supported algorithms: AES-CBC, AES-CTR, AES-GCM, HMAC-SHA1 and NULL.
  • Each SA must be handle by a unique lcore (1 RX queue per port).
  • No chained mbufs.

Compiling the Application

To compile the sample application see Compiling the Sample Applications.

The application is located in the rpsec-secgw sub-directory.

  1. [Optional] Build the application for debugging: This option adds some extra flags, disables compiler optimizations and is verbose:

    make DEBUG=1
    

Running the Application

The application has a number of command line options:

./build/ipsec-secgw [EAL options] --
                     -p PORTMASK -P -u PORTMASK -j FRAMESIZE
                     --config (port,queue,lcore)[,(port,queue,lcore]
                     --single-sa SAIDX
                     -f CONFIG_FILE_PATH

Where:

  • -p PORTMASK: Hexadecimal bitmask of ports to configure.
  • -P: optional. Sets all ports to promiscuous mode so that packets are accepted regardless of the packet’s Ethernet MAC destination address. Without this option, only packets with the Ethernet MAC destination address set to the Ethernet address of the port are accepted (default is enabled).
  • -u PORTMASK: hexadecimal bitmask of unprotected ports
  • -j FRAMESIZE: optional. Enables jumbo frames with the maximum size specified as FRAMESIZE. If an invalid value is provided as FRAMESIZE then the default value 9000 is used.
  • --config (port,queue,lcore)[,(port,queue,lcore)]: determines which queues from which ports are mapped to which cores.
  • --single-sa SAIDX: use a single SA for outbound traffic, bypassing the SP on both Inbound and Outbound. This option is meant for debugging/performance purposes.
  • -f CONFIG_FILE_PATH: the full path of text-based file containing all configuration items for running the application (See Configuration file syntax section below). -f CONFIG_FILE_PATH must be specified. ONLY the UNIX format configuration file is accepted.

The mapping of lcores to port/queues is similar to other l3fwd applications.

For example, given the following command line:

./build/ipsec-secgw -l 20,21 -n 4 --socket-mem 0,2048       \
       --vdev "crypto_null" -- -p 0xf -P -u 0x3      \
       --config="(0,0,20),(1,0,20),(2,0,21),(3,0,21)"       \
       -f /path/to/config_file                              \

where each options means:

  • The -l option enables cores 20 and 21.

  • The -n option sets memory 4 channels.

  • The --socket-mem to use 2GB on socket 1.

  • The --vdev "crypto_null" option creates virtual NULL cryptodev PMD.

  • The -p option enables ports (detected) 0, 1, 2 and 3.

  • The -P option enables promiscuous mode.

  • The -u option sets ports 1 and 2 as unprotected, leaving 2 and 3 as protected.

  • The --config option enables one queue per port with the following mapping:

    Port Queue lcore Description
    0 0 20 Map queue 0 from port 0 to lcore 20.
    1 0 20 Map queue 0 from port 1 to lcore 20.
    2 0 21 Map queue 0 from port 2 to lcore 21.
    3 0 21 Map queue 0 from port 3 to lcore 21.
  • The -f /path/to/config_file option enables the application read and parse the configuration file specified, and configures the application with a given set of SP, SA and Routing entries accordingly. The syntax of the configuration file will be explained below in more detail. Please note the parser only accepts UNIX format text file. Other formats such as DOS/MAC format will cause a parse error.

Refer to the DPDK Getting Started Guide for general information on running applications and the Environment Abstraction Layer (EAL) options.

The application would do a best effort to “map” crypto devices to cores, with hardware devices having priority. Basically, hardware devices if present would be assigned to a core before software ones. This means that if the application is using a single core and both hardware and software crypto devices are detected, hardware devices will be used.

A way to achieve the case where you want to force the use of virtual crypto devices is to whitelist the Ethernet devices needed and therefore implicitly blacklisting all hardware crypto devices.

For example, something like the following command line:

./build/ipsec-secgw -l 20,21 -n 4 --socket-mem 0,2048 \
        -w 81:00.0 -w 81:00.1 -w 81:00.2 -w 81:00.3 \
        --vdev "crypto_aesni_mb" --vdev "crypto_null" \
        -- \
        -p 0xf -P -u 0x3 --config="(0,0,20),(1,0,20),(2,0,21),(3,0,21)" \
        -f sample.cfg

Configurations

The following sections provide the syntax of configurations to initialize your SP, SA and Routing tables. Configurations shall be specified in the configuration file to be passed to the application. The file is then parsed by the application. The successful parsing will result in the appropriate rules being applied to the tables accordingly.

Configuration File Syntax

As mention in the overview, the Security Policies are ACL rules. The application parsers the rules specified in the configuration file and passes them to the ACL table, and replicates them per socket in use.

Following are the configuration file syntax.

General rule syntax

The parse treats one line in the configuration file as one configuration item (unless the line concatenation symbol exists). Every configuration item shall follow the syntax of either SP, SA, or Routing rules specified below.

The configuration parser supports the following special symbols:

  • Comment symbol #. Any character from this symbol to the end of line is treated as comment and will not be parsed.
  • Line concatenation symbol \. This symbol shall be placed in the end of the line to be concatenated to the line below. Multiple lines’ concatenation is supported.
SP rule syntax

The SP rule syntax is shown as follows:

sp <ip_ver> <dir> esp <action> <priority> <src_ip> <dst_ip>
<proto> <sport> <dport>

where each options means:

<ip_ver>

  • IP protocol version
  • Optional: No
  • Available options:
    • ipv4: IP protocol version 4
    • ipv6: IP protocol version 6

<dir>

  • The traffic direction
  • Optional: No
  • Available options:
    • in: inbound traffic
    • out: outbound traffic

<action>

  • IPsec action
  • Optional: No
  • Available options:
    • protect <SA_idx>: the specified traffic is protected by SA rule with id SA_idx
    • bypass: the specified traffic traffic is bypassed
    • discard: the specified traffic is discarded

<priority>

  • Rule priority
  • Optional: Yes, default priority 0 will be used
  • Syntax: pri <id>

<src_ip>

  • The source IP address and mask
  • Optional: Yes, default address 0.0.0.0 and mask of 0 will be used
  • Syntax:
    • src X.X.X.X/Y for IPv4
    • src XXXX:XXXX:XXXX:XXXX:XXXX:XXXX:XXXX:XXXX/Y for IPv6

<dst_ip>

  • The destination IP address and mask
  • Optional: Yes, default address 0.0.0.0 and mask of 0 will be used
  • Syntax:
    • dst X.X.X.X/Y for IPv4
    • dst XXXX:XXXX:XXXX:XXXX:XXXX:XXXX:XXXX:XXXX/Y for IPv6

<proto>

  • The protocol start and end range
  • Optional: yes, default range of 0 to 0 will be used
  • Syntax: proto X:Y

<sport>

  • The source port start and end range
  • Optional: yes, default range of 0 to 0 will be used
  • Syntax: sport X:Y

<dport>

  • The destination port start and end range
  • Optional: yes, default range of 0 to 0 will be used
  • Syntax: dport X:Y

Example SP rules:

sp ipv4 out esp protect 105 pri 1 dst 192.168.115.0/24 sport 0:65535 \
dport 0:65535

sp ipv6 in esp bypass pri 1 dst 0000:0000:0000:0000:5555:5555:\
0000:0000/96 sport 0:65535 dport 0:65535
SA rule syntax

The successfully parsed SA rules will be stored in an array table.

The SA rule syntax is shown as follows:

sa <dir> <spi> <cipher_algo> <cipher_key> <auth_algo> <auth_key>
<mode> <src_ip> <dst_ip> <action_type> <port_id>

where each options means:

<dir>

  • The traffic direction
  • Optional: No
  • Available options:
    • in: inbound traffic
    • out: outbound traffic

<spi>

  • The SPI number
  • Optional: No
  • Syntax: unsigned integer number

<cipher_algo>

  • Cipher algorithm
  • Optional: Yes, unless <aead_algo> is not used
  • Available options:
    • null: NULL algorithm
    • aes-128-cbc: AES-CBC 128-bit algorithm
    • aes-128-ctr: AES-CTR 128-bit algorithm
  • Syntax: cipher_algo <your algorithm>

<cipher_key>

  • Cipher key, NOT available when ‘null’ algorithm is used

  • Optional: Yes, unless <aead_algo> is not used. Must be followed by <cipher_algo> option

  • Syntax: Hexadecimal bytes (0x0-0xFF) concatenate by colon symbol ‘:’. The number of bytes should be as same as the specified cipher algorithm key size.

    For example: cipher_key A1:B2:C3:D4:A1:B2:C3:D4:A1:B2:C3:D4: A1:B2:C3:D4

<auth_algo>

  • Authentication algorithm

  • Optional: Yes, unless <aead_algo> is not used

  • Available options:

    • null: NULL algorithm
    • sha1-hmac: HMAC SHA1 algorithm

<auth_key>

  • Authentication key, NOT available when ‘null’ or ‘aes-128-gcm’ algorithm is used.

  • Optional: Yes, unless <aead_algo> is not used. Must be followed by <auth_algo> option

  • Syntax: Hexadecimal bytes (0x0-0xFF) concatenate by colon symbol ‘:’. The number of bytes should be as same as the specified authentication algorithm key size.

    For example: auth_key A1:B2:C3:D4:A1:B2:C3:D4:A1:B2:C3:D4:A1:B2:C3:D4: A1:B2:C3:D4

<aead_algo>

  • AEAD algorithm
  • Optional: Yes, unless <cipher_algo> and <auth_algo> are not used
  • Available options:
    • aes-128-gcm: AES-GCM 128-bit algorithm
  • Syntax: cipher_algo <your algorithm>

<aead_key>

  • Cipher key, NOT available when ‘null’ algorithm is used

  • Optional: Yes, unless <cipher_algo> and <auth_algo> are not used. Must be followed by <aead_algo> option

  • Syntax: Hexadecimal bytes (0x0-0xFF) concatenate by colon symbol ‘:’. The number of bytes should be as same as the specified AEAD algorithm key size.

    For example: aead_key A1:B2:C3:D4:A1:B2:C3:D4:A1:B2:C3:D4: A1:B2:C3:D4

<mode>

  • The operation mode
  • Optional: No
  • Available options:
    • ipv4-tunnel: Tunnel mode for IPv4 packets
    • ipv6-tunnel: Tunnel mode for IPv6 packets
    • transport: transport mode
  • Syntax: mode XXX

<src_ip>

  • The source IP address. This option is not available when transport mode is used
  • Optional: Yes, default address 0.0.0.0 will be used
  • Syntax:
    • src X.X.X.X for IPv4
    • src XXXX:XXXX:XXXX:XXXX:XXXX:XXXX:XXXX:XXXX for IPv6

<dst_ip>

  • The destination IP address. This option is not available when transport mode is used
  • Optional: Yes, default address 0.0.0.0 will be used
  • Syntax:
    • dst X.X.X.X for IPv4
    • dst XXXX:XXXX:XXXX:XXXX:XXXX:XXXX:XXXX:XXXX for IPv6

<type>

  • Action type to specify the security action. This option specify the SA to be performed with look aside protocol offload to HW accelerator or protocol offload on ethernet device or inline crypto processing on the ethernet device during transmission.
  • Optional: Yes, default type no-offload
  • Available options:
    • lookaside-protocol-offload: look aside protocol offload to HW accelerator
    • inline-protocol-offload: inline protocol offload on ethernet device
    • inline-crypto-offload: inline crypto processing on ethernet device
    • no-offload: no offloading to hardware

<port_id>

  • Port/device ID of the ethernet/crypto accelerator for which the SA is configured. This option is used when type is NOT no-offload
  • Optional: No, if type is not no-offload
  • Syntax:
    • port_id X X is a valid device number in decimal

Example SA rules:

sa out 5 cipher_algo null auth_algo null mode ipv4-tunnel \
src 172.16.1.5 dst 172.16.2.5

sa out 25 cipher_algo aes-128-cbc \
cipher_key c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3 \
auth_algo sha1-hmac \
auth_key c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3 \
mode ipv6-tunnel \
src 1111:1111:1111:1111:1111:1111:1111:5555 \
dst 2222:2222:2222:2222:2222:2222:2222:5555

sa in 105 aead_algo aes-128-gcm \
aead_key de:ad:be:ef:de:ad:be:ef:de:ad:be:ef:de:ad:be:ef:de:ad:be:ef \
mode ipv4-tunnel src 172.16.2.5 dst 172.16.1.5

sa out 5 cipher_algo aes-128-cbc cipher_key 0:0:0:0:0:0:0:0:0:0:0:0:0:0:0:0 \
auth_algo sha1-hmac auth_key 0:0:0:0:0:0:0:0:0:0:0:0:0:0:0:0:0:0:0:0 \
mode ipv4-tunnel src 172.16.1.5 dst 172.16.2.5 \
type lookaside-protocol-offload port_id 4
Routing rule syntax

The Routing rule syntax is shown as follows:

rt <ip_ver> <src_ip> <dst_ip> <port>

where each options means:

<ip_ver>

  • IP protocol version
  • Optional: No
  • Available options:
    • ipv4: IP protocol version 4
    • ipv6: IP protocol version 6

<src_ip>

  • The source IP address and mask
  • Optional: Yes, default address 0.0.0.0 and mask of 0 will be used
  • Syntax:
    • src X.X.X.X/Y for IPv4
    • src XXXX:XXXX:XXXX:XXXX:XXXX:XXXX:XXXX:XXXX/Y for IPv6

<dst_ip>

  • The destination IP address and mask
  • Optional: Yes, default address 0.0.0.0 and mask of 0 will be used
  • Syntax:
    • dst X.X.X.X/Y for IPv4
    • dst XXXX:XXXX:XXXX:XXXX:XXXX:XXXX:XXXX:XXXX/Y for IPv6

<port>

  • The traffic output port id
  • Optional: yes, default output port 0 will be used
  • Syntax: port X

Example SP rules:

rt ipv4 dst 172.16.1.5/32 port 0

rt ipv6 dst 1111:1111:1111:1111:1111:1111:1111:5555/116 port 0

Figures

Fig. 5.1 Packet Flow

Fig. 14.1 Kernel NIC Application Packet Flow

Fig. 17.1 Performance Benchmark Setup (Basic Environment)

Fig. 17.2 Performance Benchmark Setup (Virtualized Environment)

Fig. 18.1 Performance Benchmark Setup (Basic Environment)

Fig. 18.2 Performance Benchmark Setup (Virtualized Environment)

Fig. 16.1 Encryption flow Through the L2 Forwarding with Crypto Application

Fig. 22.1 A typical IPv4 ACL rule

Fig. 22.2 Rules example

Fig. 25.1 Load Balancer Application Architecture

Fig. 28.1 Example Data Flow in a Symmetric Multi-process Application

Fig. 28.2 Example Data Flow in a Client-Server Symmetric Multi-process Application

Fig. 28.3 Master-slave Process Workflow

Fig. 28.4 Slave Process Recovery Process Flow

Fig. 30.1 QoS Scheduler Application Architecture

Fig. 31.1 Pipeline Overview

Fig. 31.2 Ring-based Processing Pipeline Performance Setup

Fig. 31.3 Threads and Pipelines

Fig. 34.1 Packet Flow Through the VMDQ and DCB Sample Application

Fig. 39.1 Test Pipeline Application

Fig. 41.1 Performance Benchmarking Setup (Basic Environment)

Fig. 41.2 Distributor Sample Application Layout

Fig. 42.1 Highlevel Solution

Fig. 42.2 VM request to scale frequency Fig. 43.1 Overlay Networking. Fig. 43.2 TEP termination Framework Overview

Fig. 44.1 PTP Synchronization Protocol

Fig. 26.1 Using EFD as a Flow-Level Load Balancer

Tables

Table 29.1 Output Traffic Marking

Table 30.1 Entity Types

Table 39.1 Table Types

Programmer’s Guide

Introduction

This document provides software architecture information, development environment information and optimization guidelines.

For programming examples and for instructions on compiling and running each sample application, see the DPDK Sample Applications User Guide for details.

For general information on compiling and running applications, see the DPDK Getting Started Guide.

Documentation Roadmap

The following is a list of DPDK documents in the suggested reading order:

  • Release Notes : Provides release-specific information, including supported features, limitations, fixed issues, known issues and so on. Also, provides the answers to frequently asked questions in FAQ format.

  • Getting Started Guide : Describes how to install and configure the DPDK software; designed to get users up and running quickly with the software.

  • FreeBSD* Getting Started Guide : A document describing the use of the DPDK with FreeBSD* has been added in DPDK Release 1.6.0. Refer to this guide for installation and configuration instructions to get started using the DPDK with FreeBSD*.

  • Programmer’s Guide (this document): Describes:

    • The software architecture and how to use it (through examples), specifically in a Linux* application (linuxapp) environment
    • The content of the DPDK, the build system (including the commands that can be used in the root DPDK Makefile to build the development kit and an application) and guidelines for porting an application
    • Optimizations used in the software and those that should be considered for new development

    A glossary of terms is also provided.

  • API Reference : Provides detailed information about DPDK functions, data structures and other programming constructs.

  • Sample Applications User Guide: Describes a set of sample applications. Each chapter describes a sample application that showcases specific functionality and provides instructions on how to compile, run and use the sample application.

Part 1: Architecture Overview

Overview

This section gives a global overview of the architecture of Data Plane Development Kit (DPDK).

The main goal of the DPDK is to provide a simple, complete framework for fast packet processing in data plane applications. Users may use the code to understand some of the techniques employed, to build upon for prototyping or to add their own protocol stacks. Alternative ecosystem options that use the DPDK are available.

The framework creates a set of libraries for specific environments through the creation of an Environment Abstraction Layer (EAL), which may be specific to a mode of the Intel® architecture (32-bit or 64-bit), Linux* user space compilers or a specific platform. These environments are created through the use of make files and configuration files. Once the EAL library is created, the user may link with the library to create their own applications. Other libraries, outside of EAL, including the Hash, Longest Prefix Match (LPM) and rings libraries are also provided. Sample applications are provided to help show the user how to use various features of the DPDK.

The DPDK implements a run to completion model for packet processing, where all resources must be allocated prior to calling Data Plane applications, running as execution units on logical processing cores. The model does not support a scheduler and all devices are accessed by polling. The primary reason for not using interrupts is the performance overhead imposed by interrupt processing.

In addition to the run-to-completion model, a pipeline model may also be used by passing packets or messages between cores via the rings. This allows work to be performed in stages and may allow more efficient use of code on cores.

Development Environment

The DPDK project installation requires Linux and the associated toolchain, such as one or more compilers, assembler, make utility, editor and various libraries to create the DPDK components and libraries.

Once these libraries are created for the specific environment and architecture, they may then be used to create the user’s data plane application.

When creating applications for the Linux user space, the glibc library is used. For DPDK applications, two environmental variables (RTE_SDK and RTE_TARGET) must be configured before compiling the applications. The following are examples of how the variables can be set:

export RTE_SDK=/home/user/DPDK
export RTE_TARGET=x86_64-native-linuxapp-gcc

See the DPDK Getting Started Guide for information on setting up the development environment.

Environment Abstraction Layer

The Environment Abstraction Layer (EAL) provides a generic interface that hides the environment specifics from the applications and libraries. The services provided by the EAL are:

  • DPDK loading and launching
  • Support for multi-process and multi-thread execution types
  • Core affinity/assignment procedures
  • System memory allocation/de-allocation
  • Atomic/lock operations
  • Time reference
  • PCI bus access
  • Trace and debug functions
  • CPU feature identification
  • Interrupt handling
  • Alarm operations
  • Memory management (malloc)

The EAL is fully described in Environment Abstraction Layer.

Core Components

The core components are a set of libraries that provide all the elements needed for high-performance packet processing applications.

_images/architecture-overview.svg

Core Components Architecture

Ring Manager (librte_ring)

The ring structure provides a lockless multi-producer, multi-consumer FIFO API in a finite size table. It has some advantages over lockless queues; easier to implement, adapted to bulk operations and faster. A ring is used by the Memory Pool Manager (librte_mempool) and may be used as a general communication mechanism between cores and/or execution blocks connected together on a logical core.

This ring buffer and its usage are fully described in Ring Library.

Memory Pool Manager (librte_mempool)

The Memory Pool Manager is responsible for allocating pools of objects in memory. A pool is identified by name and uses a ring to store free objects. It provides some other optional services, such as a per-core object cache and an alignment helper to ensure that objects are padded to spread them equally on all RAM channels.

This memory pool allocator is described in Mempool Library.

Network Packet Buffer Management (librte_mbuf)

The mbuf library provides the facility to create and destroy buffers that may be used by the DPDK application to store message buffers. The message buffers are created at startup time and stored in a mempool, using the DPDK mempool library.

This library provides an API to allocate/free mbufs, manipulate control message buffers (ctrlmbuf) which are generic message buffers, and packet buffers (pktmbuf) which are used to carry network packets.

Network Packet Buffer Management is described in Mbuf Library.

Timer Manager (librte_timer)

This library provides a timer service to DPDK execution units, providing the ability to execute a function asynchronously. It can be periodic function calls, or just a one-shot call. It uses the timer interface provided by the Environment Abstraction Layer (EAL) to get a precise time reference and can be initiated on a per-core basis as required.

The library documentation is available in Timer Library.

Ethernet* Poll Mode Driver Architecture

The DPDK includes Poll Mode Drivers (PMDs) for 1 GbE, 10 GbE and 40GbE, and para virtualized virtio Ethernet controllers which are designed to work without asynchronous, interrupt-based signaling mechanisms.

See Poll Mode Driver.

Packet Forwarding Algorithm Support

The DPDK includes Hash (librte_hash) and Longest Prefix Match (LPM,librte_lpm) libraries to support the corresponding packet forwarding algorithms.

See Hash Library and LPM Library for more information.

librte_net

The librte_net library is a collection of IP protocol definitions and convenience macros. It is based on code from the FreeBSD* IP stack and contains protocol numbers (for use in IP headers), IP-related macros, IPv4/IPv6 header structures and TCP, UDP and SCTP header structures.

Environment Abstraction Layer

The Environment Abstraction Layer (EAL) is responsible for gaining access to low-level resources such as hardware and memory space. It provides a generic interface that hides the environment specifics from the applications and libraries. It is the responsibility of the initialization routine to decide how to allocate these resources (that is, memory space, PCI devices, timers, consoles, and so on).

Typical services expected from the EAL are:

  • DPDK Loading and Launching: The DPDK and its application are linked as a single application and must be loaded by some means.
  • Core Affinity/Assignment Procedures: The EAL provides mechanisms for assigning execution units to specific cores as well as creating execution instances.
  • System Memory Reservation: The EAL facilitates the reservation of different memory zones, for example, physical memory areas for device interactions.
  • PCI Address Abstraction: The EAL provides an interface to access PCI address space.
  • Trace and Debug Functions: Logs, dump_stack, panic and so on.
  • Utility Functions: Spinlocks and atomic counters that are not provided in libc.
  • CPU Feature Identification: Determine at runtime if a particular feature, for example, Intel® AVX is supported. Determine if the current CPU supports the feature set that the binary was compiled for.
  • Interrupt Handling: Interfaces to register/unregister callbacks to specific interrupt sources.
  • Alarm Functions: Interfaces to set/remove callbacks to be run at a specific time.

EAL in a Linux-userland Execution Environment

In a Linux user space environment, the DPDK application runs as a user-space application using the pthread library. PCI information about devices and address space is discovered through the /sys kernel interface and through kernel modules such as uio_pci_generic, or igb_uio. Refer to the UIO: User-space drivers documentation in the Linux kernel. This memory is mmap’d in the application.

The EAL performs physical memory allocation using mmap() in hugetlbfs (using huge page sizes to increase performance). This memory is exposed to DPDK service layers such as the Mempool Library.

At this point, the DPDK services layer will be initialized, then through pthread setaffinity calls, each execution unit will be assigned to a specific logical core to run as a user-level thread.

The time reference is provided by the CPU Time-Stamp Counter (TSC) or by the HPET kernel API through a mmap() call.

Initialization and Core Launching

Part of the initialization is done by the start function of glibc. A check is also performed at initialization time to ensure that the micro architecture type chosen in the config file is supported by the CPU. Then, the main() function is called. The core initialization and launch is done in rte_eal_init() (see the API documentation). It consist of calls to the pthread library (more specifically, pthread_self(), pthread_create(), and pthread_setaffinity_np()).

_images/linuxapp_launch.svg

EAL Initialization in a Linux Application Environment

Note

Initialization of objects, such as memory zones, rings, memory pools, lpm tables and hash tables, should be done as part of the overall application initialization on the master lcore. The creation and initialization functions for these objects are not multi-thread safe. However, once initialized, the objects themselves can safely be used in multiple threads simultaneously.

Multi-process Support

The Linuxapp EAL allows a multi-process as well as a multi-threaded (pthread) deployment model. See chapter Multi-process Support for more details.

Memory Mapping Discovery and Memory Reservation

The allocation of large contiguous physical memory is done using the hugetlbfs kernel filesystem. The EAL provides an API to reserve named memory zones in this contiguous memory. The physical address of the reserved memory for that memory zone is also returned to the user by the memory zone reservation API.

Note

Memory reservations done using the APIs provided by rte_malloc are also backed by pages from the hugetlbfs filesystem.

PCI Access

The EAL uses the /sys/bus/pci utilities provided by the kernel to scan the content on the PCI bus. To access PCI memory, a kernel module called uio_pci_generic provides a /dev/uioX device file and resource files in /sys that can be mmap’d to obtain access to PCI address space from the application. The DPDK-specific igb_uio module can also be used for this. Both drivers use the uio kernel feature (userland driver).

Per-lcore and Shared Variables

Note

lcore refers to a logical execution unit of the processor, sometimes called a hardware thread.

Shared variables are the default behavior. Per-lcore variables are implemented using Thread Local Storage (TLS) to provide per-thread local storage.

Logs

A logging API is provided by EAL. By default, in a Linux application, logs are sent to syslog and also to the console. However, the log function can be overridden by the user to use a different logging mechanism.

Trace and Debug Functions

There are some debug functions to dump the stack in glibc. The rte_panic() function can voluntarily provoke a SIG_ABORT, which can trigger the generation of a core file, readable by gdb.

CPU Feature Identification

The EAL can query the CPU at runtime (using the rte_cpu_get_features() function) to determine which CPU features are available.

User Space Interrupt Event
  • User Space Interrupt and Alarm Handling in Host Thread

The EAL creates a host thread to poll the UIO device file descriptors to detect the interrupts. Callbacks can be registered or unregistered by the EAL functions for a specific interrupt event and are called in the host thread asynchronously. The EAL also allows timed callbacks to be used in the same way as for NIC interrupts.

Note

In DPDK PMD, the only interrupts handled by the dedicated host thread are those for link status change (link up and link down notification) and for sudden device removal.

  • RX Interrupt Event

The receive and transmit routines provided by each PMD don’t limit themselves to execute in polling thread mode. To ease the idle polling with tiny throughput, it’s useful to pause the polling and wait until the wake-up event happens. The RX interrupt is the first choice to be such kind of wake-up event, but probably won’t be the only one.

EAL provides the event APIs for this event-driven thread mode. Taking linuxapp as an example, the implementation relies on epoll. Each thread can monitor an epoll instance in which all the wake-up events’ file descriptors are added. The event file descriptors are created and mapped to the interrupt vectors according to the UIO/VFIO spec. From bsdapp’s perspective, kqueue is the alternative way, but not implemented yet.

EAL initializes the mapping between event file descriptors and interrupt vectors, while each device initializes the mapping between interrupt vectors and queues. In this way, EAL actually is unaware of the interrupt cause on the specific vector. The eth_dev driver takes responsibility to program the latter mapping.

Note

Per queue RX interrupt event is only allowed in VFIO which supports multiple MSI-X vector. In UIO, the RX interrupt together with other interrupt causes shares the same vector. In this case, when RX interrupt and LSC(link status change) interrupt are both enabled(intr_conf.lsc == 1 && intr_conf.rxq == 1), only the former is capable.

The RX interrupt are controlled/enabled/disabled by ethdev APIs - ‘rte_eth_dev_rx_intr_*’. They return failure if the PMD hasn’t support them yet. The intr_conf.rxq flag is used to turn on the capability of RX interrupt per device.

  • Device Removal Event

This event is triggered by a device being removed at a bus level. Its underlying resources may have been made unavailable (i.e. PCI mappings unmapped). The PMD must make sure that on such occurrence, the application can still safely use its callbacks.

This event can be subscribed to in the same way one would subscribe to a link status change event. The execution context is thus the same, i.e. it is the dedicated interrupt host thread.

Considering this, it is likely that an application would want to close a device having emitted a Device Removal Event. In such case, calling rte_eth_dev_close() can trigger it to unregister its own Device Removal Event callback. Care must be taken not to close the device from the interrupt handler context. It is necessary to reschedule such closing operation.

Blacklisting

The EAL PCI device blacklist functionality can be used to mark certain NIC ports as blacklisted, so they are ignored by the DPDK. The ports to be blacklisted are identified using the PCIe* description (Domain:Bus:Device.Function).

Misc Functions

Locks and atomic operations are per-architecture (i686 and x86_64).

Memory Segments and Memory Zones (memzone)

The mapping of physical memory is provided by this feature in the EAL. As physical memory can have gaps, the memory is described in a table of descriptors, and each descriptor (called rte_memseg ) describes a contiguous portion of memory.

On top of this, the memzone allocator’s role is to reserve contiguous portions of physical memory. These zones are identified by a unique name when the memory is reserved.

The rte_memzone descriptors are also located in the configuration structure. This structure is accessed using rte_eal_get_configuration(). The lookup (by name) of a memory zone returns a descriptor containing the physical address of the memory zone.

Memory zones can be reserved with specific start address alignment by supplying the align parameter (by default, they are aligned to cache line size). The alignment value should be a power of two and not less than the cache line size (64 bytes). Memory zones can also be reserved from either 2 MB or 1 GB hugepages, provided that both are available on the system.

Multiple pthread

DPDK usually pins one pthread per core to avoid the overhead of task switching. This allows for significant performance gains, but lacks flexibility and is not always efficient.

Power management helps to improve the CPU efficiency by limiting the CPU runtime frequency. However, alternately it is possible to utilize the idle cycles available to take advantage of the full capability of the CPU.

By taking advantage of cgroup, the CPU utilization quota can be simply assigned. This gives another way to improve the CPU efficiency, however, there is a prerequisite; DPDK must handle the context switching between multiple pthreads per core.

For further flexibility, it is useful to set pthread affinity not only to a CPU but to a CPU set.

EAL pthread and lcore Affinity

The term “lcore” refers to an EAL thread, which is really a Linux/FreeBSD pthread. “EAL pthreads” are created and managed by EAL and execute the tasks issued by remote_launch. In each EAL pthread, there is a TLS (Thread Local Storage) called _lcore_id for unique identification. As EAL pthreads usually bind 1:1 to the physical CPU, the _lcore_id is typically equal to the CPU ID.

When using multiple pthreads, however, the binding is no longer always 1:1 between an EAL pthread and a specified physical CPU. The EAL pthread may have affinity to a CPU set, and as such the _lcore_id will not be the same as the CPU ID. For this reason, there is an EAL long option ‘–lcores’ defined to assign the CPU affinity of lcores. For a specified lcore ID or ID group, the option allows setting the CPU set for that EAL pthread.

The format pattern:
–lcores=’<lcore_set>[@cpu_set][,<lcore_set>[@cpu_set],...]’

‘lcore_set’ and ‘cpu_set’ can be a single number, range or a group.

A number is a “digit([0-9]+)”; a range is “<number>-<number>”; a group is “(<number|range>[,<number|range>,...])”.

If a ‘@cpu_set’ value is not supplied, the value of ‘cpu_set’ will default to the value of ‘lcore_set’.

For example, "--lcores='1,2@(5-7),(3-5)@(0,2),(0,6),7-8'" which means start 9 EAL thread;
    lcore 0 runs on cpuset 0x41 (cpu 0,6);
    lcore 1 runs on cpuset 0x2 (cpu 1);
    lcore 2 runs on cpuset 0xe0 (cpu 5,6,7);
    lcore 3,4,5 runs on cpuset 0x5 (cpu 0,2);
    lcore 6 runs on cpuset 0x41 (cpu 0,6);
    lcore 7 runs on cpuset 0x80 (cpu 7);
    lcore 8 runs on cpuset 0x100 (cpu 8).

Using this option, for each given lcore ID, the associated CPUs can be assigned. It’s also compatible with the pattern of corelist(‘-l’) option.

non-EAL pthread support

It is possible to use the DPDK execution context with any user pthread (aka. Non-EAL pthreads). In a non-EAL pthread, the _lcore_id is always LCORE_ID_ANY which identifies that it is not an EAL thread with a valid, unique, _lcore_id. Some libraries will use an alternative unique ID (e.g. TID), some will not be impacted at all, and some will work but with limitations (e.g. timer and mempool libraries).

All these impacts are mentioned in Known Issues section.

Public Thread API

There are two public APIs rte_thread_set_affinity() and rte_thread_get_affinity() introduced for threads. When they’re used in any pthread context, the Thread Local Storage(TLS) will be set/get.

Those TLS include _cpuset and _socket_id:

  • _cpuset stores the CPUs bitmap to which the pthread is affinitized.
  • _socket_id stores the NUMA node of the CPU set. If the CPUs in CPU set belong to different NUMA node, the _socket_id will be set to SOCKET_ID_ANY.
Known Issues
  • rte_mempool

    The rte_mempool uses a per-lcore cache inside the mempool. For non-EAL pthreads, rte_lcore_id() will not return a valid number. So for now, when rte_mempool is used with non-EAL pthreads, the put/get operations will bypass the default mempool cache and there is a performance penalty because of this bypass. Only user-owned external caches can be used in a non-EAL context in conjunction with rte_mempool_generic_put() and rte_mempool_generic_get() that accept an explicit cache parameter.

  • rte_ring

    rte_ring supports multi-producer enqueue and multi-consumer dequeue. However, it is non-preemptive, this has a knock on effect of making rte_mempool non-preemptable.

    Note

    The “non-preemptive” constraint means:

    • a pthread doing multi-producers enqueues on a given ring must not be preempted by another pthread doing a multi-producer enqueue on the same ring.
    • a pthread doing multi-consumers dequeues on a given ring must not be preempted by another pthread doing a multi-consumer dequeue on the same ring.

    Bypassing this constraint may cause the 2nd pthread to spin until the 1st one is scheduled again. Moreover, if the 1st pthread is preempted by a context that has an higher priority, it may even cause a dead lock.

    This does not mean it cannot be used, simply, there is a need to narrow down the situation when it is used by multi-pthread on the same core.

    1. It CAN be used for any single-producer or single-consumer situation.
    2. It MAY be used by multi-producer/consumer pthread whose scheduling policy are all SCHED_OTHER(cfs). User SHOULD be aware of the performance penalty before using it.
    3. It MUST not be used by multi-producer/consumer pthreads, whose scheduling policies are SCHED_FIFO or SCHED_RR.
  • rte_timer

    Running rte_timer_manager() on a non-EAL pthread is not allowed. However, resetting/stopping the timer from a non-EAL pthread is allowed.

  • rte_log

    In non-EAL pthreads, there is no per thread loglevel and logtype, global loglevels are used.

  • misc

    The debug statistics of rte_ring, rte_mempool and rte_timer are not supported in a non-EAL pthread.

cgroup control

The following is a simple example of cgroup control usage, there are two pthreads(t0 and t1) doing packet I/O on the same core ($CPU). We expect only 50% of CPU spend on packet IO.

mkdir /sys/fs/cgroup/cpu/pkt_io
mkdir /sys/fs/cgroup/cpuset/pkt_io

echo $cpu > /sys/fs/cgroup/cpuset/cpuset.cpus

echo $t0 > /sys/fs/cgroup/cpu/pkt_io/tasks
echo $t0 > /sys/fs/cgroup/cpuset/pkt_io/tasks

echo $t1 > /sys/fs/cgroup/cpu/pkt_io/tasks
echo $t1 > /sys/fs/cgroup/cpuset/pkt_io/tasks

cd /sys/fs/cgroup/cpu/pkt_io
echo 100000 > pkt_io/cpu.cfs_period_us
echo  50000 > pkt_io/cpu.cfs_quota_us

Malloc

The EAL provides a malloc API to allocate any-sized memory.

The objective of this API is to provide malloc-like functions to allow allocation from hugepage memory and to facilitate application porting. The DPDK API Reference manual describes the available functions.

Typically, these kinds of allocations should not be done in data plane processing because they are slower than pool-based allocation and make use of locks within the allocation and free paths. However, they can be used in configuration code.

Refer to the rte_malloc() function description in the DPDK API Reference manual for more information.

Cookies

When CONFIG_RTE_MALLOC_DEBUG is enabled, the allocated memory contains overwrite protection fields to help identify buffer overflows.

Alignment and NUMA Constraints

The rte_malloc() takes an align argument that can be used to request a memory area that is aligned on a multiple of this value (which must be a power of two).

On systems with NUMA support, a call to the rte_malloc() function will return memory that has been allocated on the NUMA socket of the core which made the call. A set of APIs is also provided, to allow memory to be explicitly allocated on a NUMA socket directly, or by allocated on the NUMA socket where another core is located, in the case where the memory is to be used by a logical core other than on the one doing the memory allocation.

Use Cases

This API is meant to be used by an application that requires malloc-like functions at initialization time.

For allocating/freeing data at runtime, in the fast-path of an application, the memory pool library should be used instead.

Internal Implementation
Data Structures

There are two data structure types used internally in the malloc library:

  • struct malloc_heap - used to track free space on a per-socket basis
  • struct malloc_elem - the basic element of allocation and free-space tracking inside the library.
Structure: malloc_heap

The malloc_heap structure is used to manage free space on a per-socket basis. Internally, there is one heap structure per NUMA node, which allows us to allocate memory to a thread based on the NUMA node on which this thread runs. While this does not guarantee that the memory will be used on that NUMA node, it is no worse than a scheme where the memory is always allocated on a fixed or random node.

The key fields of the heap structure and their function are described below (see also diagram above):

  • lock - the lock field is needed to synchronize access to the heap. Given that the free space in the heap is tracked using a linked list, we need a lock to prevent two threads manipulating the list at the same time.
  • free_head - this points to the first element in the list of free nodes for this malloc heap.

Note

The malloc_heap structure does not keep track of in-use blocks of memory, since these are never touched except when they are to be freed again - at which point the pointer to the block is an input to the free() function.

_images/malloc_heap.svg

Example of a malloc heap and malloc elements within the malloc library

Structure: malloc_elem

The malloc_elem structure is used as a generic header structure for various blocks of memory. It is used in three different ways - all shown in the diagram above:

  1. As a header on a block of free or allocated memory - normal case
  2. As a padding header inside a block of memory
  3. As an end-of-memseg marker

The most important fields in the structure and how they are used are described below.

Note

If the usage of a particular field in one of the above three usages is not described, the field can be assumed to have an undefined value in that situation, for example, for padding headers only the “state” and “pad” fields have valid values.

  • heap - this pointer is a reference back to the heap structure from which this block was allocated. It is used for normal memory blocks when they are being freed, to add the newly-freed block to the heap’s free-list.
  • prev - this pointer points to the header element/block in the memseg immediately behind the current one. When freeing a block, this pointer is used to reference the previous block to check if that block is also free. If so, then the two free blocks are merged to form a single larger block.
  • next_free - this pointer is used to chain the free-list of unallocated memory blocks together. It is only used in normal memory blocks; on malloc() to find a suitable free block to allocate and on free() to add the newly freed element to the free-list.
  • state - This field can have one of three values: FREE, BUSY or PAD. The former two are to indicate the allocation state of a normal memory block and the latter is to indicate that the element structure is a dummy structure at the end of the start-of-block padding, i.e. where the start of the data within a block is not at the start of the block itself, due to alignment constraints. In that case, the pad header is used to locate the actual malloc element header for the block. For the end-of-memseg structure, this is always a BUSY value, which ensures that no element, on being freed, searches beyond the end of the memseg for other blocks to merge with into a larger free area.
  • pad - this holds the length of the padding present at the start of the block. In the case of a normal block header, it is added to the address of the end of the header to give the address of the start of the data area, i.e. the value passed back to the application on a malloc. Within a dummy header inside the padding, this same value is stored, and is subtracted from the address of the dummy header to yield the address of the actual block header.
  • size - the size of the data block, including the header itself. For end-of-memseg structures, this size is given as zero, though it is never actually checked. For normal blocks which are being freed, this size value is used in place of a “next” pointer to identify the location of the next block of memory that in the case of being FREE, the two free blocks can be merged into one.
Memory Allocation

On EAL initialization, all memsegs are setup as part of the malloc heap. This setup involves placing a dummy structure at the end with BUSY state, which may contain a sentinel value if CONFIG_RTE_MALLOC_DEBUG is enabled, and a proper element header with FREE at the start for each memseg. The FREE element is then added to the free_list for the malloc heap.

When an application makes a call to a malloc-like function, the malloc function will first index the lcore_config structure for the calling thread, and determine the NUMA node of that thread. The NUMA node is used to index the array of malloc_heap structures which is passed as a parameter to the heap_alloc() function, along with the requested size, type, alignment and boundary parameters.

The heap_alloc() function will scan the free_list of the heap, and attempt to find a free block suitable for storing data of the requested size, with the requested alignment and boundary constraints.

When a suitable free element has been identified, the pointer to be returned to the user is calculated. The cache-line of memory immediately preceding this pointer is filled with a struct malloc_elem header. Because of alignment and boundary constraints, there could be free space at the start and/or end of the element, resulting in the following behavior:

  1. Check for trailing space. If the trailing space is big enough, i.e. > 128 bytes, then the free element is split. If it is not, then we just ignore it (wasted space).
  2. Check for space at the start of the element. If the space at the start is small, i.e. <=128 bytes, then a pad header is used, and the remaining space is wasted. If, however, the remaining space is greater, then the free element is split.

The advantage of allocating the memory from the end of the existing element is that no adjustment of the free list needs to take place - the existing element on the free list just has its size pointer adjusted, and the following element has its “prev” pointer redirected to the newly created element.

Freeing Memory

To free an area of memory, the pointer to the start of the data area is passed to the free function. The size of the malloc_elem structure is subtracted from this pointer to get the element header for the block. If this header is of type PAD then the pad length is further subtracted from the pointer to get the proper element header for the entire block.

From this element header, we get pointers to the heap from which the block was allocated and to where it must be freed, as well as the pointer to the previous element, and via the size field, we can calculate the pointer to the next element. These next and previous elements are then checked to see if they are also FREE, and if so, they are merged with the current element. This means that we can never have two FREE memory blocks adjacent to one another, as they are always merged into a single block.

Service Cores

DPDK has a concept known as service cores, which enables a dynamic way of performing work on DPDK lcores. Service core support is built into the EAL, and an API is provided to optionally allow applications to control how the service cores are used at runtime.

The service cores concept is built up out of services (components of DPDK that require CPU cycles to operate) and service cores (DPDK lcores, tasked with running services). The power of the service core concept is that the mapping between service cores and services can be configured to abstract away the difference between platforms and environments.

For example, the Eventdev has hardware and software PMDs. Of these the software PMD requires an lcore to perform the scheduling operations, while the hardware PMD does not. With service cores, the application would not directly notice that the scheduling is done in software.

For detailed information about the service core API, please refer to the docs.

Service Core Initialization

There are two methods to having service cores in a DPDK application, either by using the service coremask, or by dynamically adding cores using the API. The simpler of the two is to pass the -s coremask argument to EAL, which will take any cores available in the main DPDK coremask, an if the bits are also set in the service coremask the cores become service-cores instead of DPDK application lcores.

Enabling Services on Cores

Each registered service can be individually mapped to a service core, or set of service cores. Enabling a service on a particular core means that the lcore in question will run the service. Disabling that core on the service stops the lcore in question from running the service.

Using this method, it is possible to assign specific workloads to each service core, and map N workloads to M number of service cores. Each service lcore loops over the services that are enabled for that core, and invokes the function to run the service.

Service Core Statistics

The service core library is capable of collecting runtime statistics like number of calls to a specific service, and number of cycles used by the service. The cycle count collection is dynamically configurable, allowing any application to profile the services running on the system at any time.

Ring Library

The ring allows the management of queues. Instead of having a linked list of infinite size, the rte_ring has the following properties:

  • FIFO
  • Maximum size is fixed, the pointers are stored in a table
  • Lockless implementation
  • Multi-consumer or single-consumer dequeue
  • Multi-producer or single-producer enqueue
  • Bulk dequeue - Dequeues the specified count of objects if successful; otherwise fails
  • Bulk enqueue - Enqueues the specified count of objects if successful; otherwise fails
  • Burst dequeue - Dequeue the maximum available objects if the specified count cannot be fulfilled
  • Burst enqueue - Enqueue the maximum available objects if the specified count cannot be fulfilled

The advantages of this data structure over a linked list queue are as follows:

  • Faster; only requires a single Compare-And-Swap instruction of sizeof(void *) instead of several double-Compare-And-Swap instructions.
  • Simpler than a full lockless queue.
  • Adapted to bulk enqueue/dequeue operations. As pointers are stored in a table, a dequeue of several objects will not produce as many cache misses as in a linked queue. Also, a bulk dequeue of many objects does not cost more than a dequeue of a simple object.

The disadvantages:

  • Size is fixed
  • Having many rings costs more in terms of memory than a linked list queue. An empty ring contains at least N pointers.

A simplified representation of a Ring is shown in with consumer and producer head and tail pointers to objects stored in the data structure.

_images/ring1.svg

Ring Structure

References for Ring Implementation in FreeBSD*

The following code was added in FreeBSD 8.0, and is used in some network device drivers (at least in Intel drivers):

Lockless Ring Buffer in Linux*

The following is a link describing the Linux Lockless Ring Buffer Design.

Additional Features

Name

A ring is identified by a unique name. It is not possible to create two rings with the same name (rte_ring_create() returns NULL if this is attempted).

Use Cases

Use cases for the Ring library include:

  • Communication between applications in the DPDK
  • Used by memory pool allocator

Anatomy of a Ring Buffer

This section explains how a ring buffer operates. The ring structure is composed of two head and tail couples; one is used by producers and one is used by the consumers. The figures of the following sections refer to them as prod_head, prod_tail, cons_head and cons_tail.

Each figure represents a simplified state of the ring, which is a circular buffer. The content of the function local variables is represented on the top of the figure, and the content of ring structure is represented on the bottom of the figure.

Single Producer Enqueue

This section explains what occurs when a producer adds an object to the ring. In this example, only the producer head and tail (prod_head and prod_tail) are modified, and there is only one producer.

The initial state is to have a prod_head and prod_tail pointing at the same location.

Enqueue First Step

First, ring->prod_head and ring->cons_tail are copied in local variables. The prod_next local variable points to the next element of the table, or several elements after in case of bulk enqueue.

If there is not enough room in the ring (this is detected by checking cons_tail), it returns an error.

_images/ring-enqueue1.svg

Enqueue first step

Enqueue Second Step

The second step is to modify ring->prod_head in ring structure to point to the same location as prod_next.

A pointer to the added object is copied in the ring (obj4).

_images/ring-enqueue2.svg

Enqueue second step

Enqueue Last Step

Once the object is added in the ring, ring->prod_tail in the ring structure is modified to point to the same location as ring->prod_head. The enqueue operation is finished.

_images/ring-enqueue3.svg

Enqueue last step

Single Consumer Dequeue

This section explains what occurs when a consumer dequeues an object from the ring. In this example, only the consumer head and tail (cons_head and cons_tail) are modified and there is only one consumer.

The initial state is to have a cons_head and cons_tail pointing at the same location.

Dequeue First Step

First, ring->cons_head and ring->prod_tail are copied in local variables. The cons_next local variable points to the next element of the table, or several elements after in the case of bulk dequeue.

If there are not enough objects in the ring (this is detected by checking prod_tail), it returns an error.

_images/ring-dequeue1.svg

Dequeue last step

Dequeue Second Step

The second step is to modify ring->cons_head in the ring structure to point to the same location as cons_next.

The pointer to the dequeued object (obj1) is copied in the pointer given by the user.

_images/ring-dequeue2.svg

Dequeue second step

Dequeue Last Step

Finally, ring->cons_tail in the ring structure is modified to point to the same location as ring->cons_head. The dequeue operation is finished.

_images/ring-dequeue3.svg

Dequeue last step

Multiple Producers Enqueue

This section explains what occurs when two producers concurrently add an object to the ring. In this example, only the producer head and tail (prod_head and prod_tail) are modified.

The initial state is to have a prod_head and prod_tail pointing at the same location.

Multiple Producers Enqueue First Step

On both cores, ring->prod_head and ring->cons_tail are copied in local variables. The prod_next local variable points to the next element of the table, or several elements after in the case of bulk enqueue.

If there is not enough room in the ring (this is detected by checking cons_tail), it returns an error.

_images/ring-mp-enqueue1.svg

Multiple producer enqueue first step

Multiple Producers Enqueue Second Step

The second step is to modify ring->prod_head in the ring structure to point to the same location as prod_next. This operation is done using a Compare And Swap (CAS) instruction, which does the following operations atomically:

  • If ring->prod_head is different to local variable prod_head, the CAS operation fails, and the code restarts at first step.
  • Otherwise, ring->prod_head is set to local prod_next, the CAS operation is successful, and processing continues.

In the figure, the operation succeeded on core 1, and step one restarted on core 2.

_images/ring-mp-enqueue2.svg

Multiple producer enqueue second step

Multiple Producers Enqueue Third Step

The CAS operation is retried on core 2 with success.

The core 1 updates one element of the ring(obj4), and the core 2 updates another one (obj5).

_images/ring-mp-enqueue3.svg

Multiple producer enqueue third step

Multiple Producers Enqueue Fourth Step

Each core now wants to update ring->prod_tail. A core can only update it if ring->prod_tail is equal to the prod_head local variable. This is only true on core 1. The operation is finished on core 1.

_images/ring-mp-enqueue4.svg

Multiple producer enqueue fourth step

Multiple Producers Enqueue Last Step

Once ring->prod_tail is updated by core 1, core 2 is allowed to update it too. The operation is also finished on core 2.

_images/ring-mp-enqueue5.svg

Multiple producer enqueue last step

Modulo 32-bit Indexes

In the preceding figures, the prod_head, prod_tail, cons_head and cons_tail indexes are represented by arrows. In the actual implementation, these values are not between 0 and size(ring)-1 as would be assumed. The indexes are between 0 and 2^32 -1, and we mask their value when we access the pointer table (the ring itself). 32-bit modulo also implies that operations on indexes (such as, add/subtract) will automatically do 2^32 modulo if the result overflows the 32-bit number range.

The following are two examples that help to explain how indexes are used in a ring.

Note

To simplify the explanation, operations with modulo 16-bit are used instead of modulo 32-bit. In addition, the four indexes are defined as unsigned 16-bit integers, as opposed to unsigned 32-bit integers in the more realistic case.

_images/ring-modulo1.svg

Modulo 32-bit indexes - Example 1

This ring contains 11000 entries.

_images/ring-modulo2.svg

Modulo 32-bit indexes - Example 2

This ring contains 12536 entries.

Note

For ease of understanding, we use modulo 65536 operations in the above examples. In real execution cases, this is redundant for low efficiency, but is done automatically when the result overflows.

The code always maintains a distance between producer and consumer between 0 and size(ring)-1. Thanks to this property, we can do subtractions between 2 index values in a modulo-32bit base: that’s why the overflow of the indexes is not a problem.

At any time, entries and free_entries are between 0 and size(ring)-1, even if only the first term of subtraction has overflowed:

uint32_t entries = (prod_tail - cons_head);
uint32_t free_entries = (mask + cons_tail -prod_head);

Mempool Library

A memory pool is an allocator of a fixed-sized object. In the DPDK, it is identified by name and uses a mempool handler to store free objects. The default mempool handler is ring based. It provides some other optional services such as a per-core object cache and an alignment helper to ensure that objects are padded to spread them equally on all DRAM or DDR3 channels.

This library is used by the Mbuf Library.

Cookies

In debug mode (CONFIG_RTE_LIBRTE_MEMPOOL_DEBUG is enabled), cookies are added at the beginning and end of allocated blocks. The allocated objects then contain overwrite protection fields to help debugging buffer overflows.

Stats

In debug mode (CONFIG_RTE_LIBRTE_MEMPOOL_DEBUG is enabled), statistics about get from/put in the pool are stored in the mempool structure. Statistics are per-lcore to avoid concurrent access to statistics counters.

Memory Alignment Constraints

Depending on hardware memory configuration, performance can be greatly improved by adding a specific padding between objects. The objective is to ensure that the beginning of each object starts on a different channel and rank in memory so that all channels are equally loaded.

This is particularly true for packet buffers when doing L3 forwarding or flow classification. Only the first 64 bytes are accessed, so performance can be increased by spreading the start addresses of objects among the different channels.

The number of ranks on any DIMM is the number of independent sets of DRAMs that can be accessed for the full data bit-width of the DIMM. The ranks cannot be accessed simultaneously since they share the same data path. The physical layout of the DRAM chips on the DIMM itself does not necessarily relate to the number of ranks.

When running an application, the EAL command line options provide the ability to add the number of memory channels and ranks.

Note

The command line must always have the number of memory channels specified for the processor.

Examples of alignment for different DIMM architectures are shown in Fig. 6.1 and Fig. 6.2.

_images/memory-management.svg

Two Channels and Quad-ranked DIMM Example

In this case, the assumption is that a packet is 16 blocks of 64 bytes, which is not true.

The Intel® 5520 chipset has three channels, so in most cases, no padding is required between objects (except for objects whose size are n x 3 x 64 bytes blocks).

_images/memory-management2.svg

Three Channels and Two Dual-ranked DIMM Example

When creating a new pool, the user can specify to use this feature or not.

Local Cache

In terms of CPU usage, the cost of multiple cores accessing a memory pool’s ring of free buffers may be high since each access requires a compare-and-set (CAS) operation. To avoid having too many access requests to the memory pool’s ring, the memory pool allocator can maintain a per-core cache and do bulk requests to the memory pool’s ring, via the cache with many fewer locks on the actual memory pool structure. In this way, each core has full access to its own cache (with locks) of free objects and only when the cache fills does the core need to shuffle some of the free objects back to the pools ring or obtain more objects when the cache is empty.

While this may mean a number of buffers may sit idle on some core’s cache, the speed at which a core can access its own cache for a specific memory pool without locks provides performance gains.

The cache is composed of a small, per-core table of pointers and its length (used as a stack). This internal cache can be enabled or disabled at creation of the pool.

The maximum size of the cache is static and is defined at compilation time (CONFIG_RTE_MEMPOOL_CACHE_MAX_SIZE).

Fig. 6.3 shows a cache in operation.

_images/mempool.svg

A mempool in Memory with its Associated Ring

Alternatively to the internal default per-lcore local cache, an application can create and manage external caches through the rte_mempool_cache_create(), rte_mempool_cache_free() and rte_mempool_cache_flush() calls. These user-owned caches can be explicitly passed to rte_mempool_generic_put() and rte_mempool_generic_get(). The rte_mempool_default_cache() call returns the default internal cache if any. In contrast to the default caches, user-owned caches can be used by non-EAL threads too.

Mempool Handlers

This allows external memory subsystems, such as external hardware memory management systems and software based memory allocators, to be used with DPDK.

There are two aspects to a mempool handler.

  • Adding the code for your new mempool operations (ops). This is achieved by adding a new mempool ops code, and using the MEMPOOL_REGISTER_OPS macro.
  • Using the new API to call rte_mempool_create_empty() and rte_mempool_set_ops_byname() to create a new mempool and specifying which ops to use.

Several different mempool handlers may be used in the same application. A new mempool can be created by using the rte_mempool_create_empty() function, then using rte_mempool_set_ops_byname() to point the mempool to the relevant mempool handler callback (ops) structure.

Legacy applications may continue to use the old rte_mempool_create() API call, which uses a ring based mempool handler by default. These applications will need to be modified to use a new mempool handler.

For applications that use rte_pktmbuf_create(), there is a config setting (RTE_MBUF_DEFAULT_MEMPOOL_OPS) that allows the application to make use of an alternative mempool handler.

Use Cases

All allocations that require a high level of performance should use a pool-based memory allocator. Below are some examples:

Mbuf Library

The mbuf library provides the ability to allocate and free buffers (mbufs) that may be used by the DPDK application to store message buffers. The message buffers are stored in a mempool, using the Mempool Library.

A rte_mbuf struct can carry network packet buffers or generic control buffers (indicated by the CTRL_MBUF_FLAG). This can be extended to other types. The rte_mbuf header structure is kept as small as possible and currently uses just two cache lines, with the most frequently used fields being on the first of the two cache lines.

Design of Packet Buffers

For the storage of the packet data (including protocol headers), two approaches were considered:

  1. Embed metadata within a single memory buffer the structure followed by a fixed size area for the packet data.
  2. Use separate memory buffers for the metadata structure and for the packet data.

The advantage of the first method is that it only needs one operation to allocate/free the whole memory representation of a packet. On the other hand, the second method is more flexible and allows the complete separation of the allocation of metadata structures from the allocation of packet data buffers.

The first method was chosen for the DPDK. The metadata contains control information such as message type, length, offset to the start of the data and a pointer for additional mbuf structures allowing buffer chaining.

Message buffers that are used to carry network packets can handle buffer chaining where multiple buffers are required to hold the complete packet. This is the case for jumbo frames that are composed of many mbufs linked together through their next field.

For a newly allocated mbuf, the area at which the data begins in the message buffer is RTE_PKTMBUF_HEADROOM bytes after the beginning of the buffer, which is cache aligned. Message buffers may be used to carry control information, packets, events, and so on between different entities in the system. Message buffers may also use their buffer pointers to point to other message buffer data sections or other structures.

Fig. 7.1 and Fig. 7.2 show some of these scenarios.

_images/mbuf1.svg

An mbuf with One Segment

_images/mbuf2.svg

An mbuf with Three Segments

The Buffer Manager implements a fairly standard set of buffer access functions to manipulate network packets.

Buffers Stored in Memory Pools

The Buffer Manager uses the Mempool Library to allocate buffers. Therefore, it ensures that the packet header is interleaved optimally across the channels and ranks for L3 processing. An mbuf contains a field indicating the pool that it originated from. When calling rte_ctrlmbuf_free(m) or rte_pktmbuf_free(m), the mbuf returns to its original pool.

Constructors

Packet and control mbuf constructors are provided by the API. The rte_pktmbuf_init() and rte_ctrlmbuf_init() functions initialize some fields in the mbuf structure that are not modified by the user once created (mbuf type, origin pool, buffer start address, and so on). This function is given as a callback function to the rte_mempool_create() function at pool creation time.

Allocating and Freeing mbufs

Allocating a new mbuf requires the user to specify the mempool from which the mbuf should be taken. For any newly-allocated mbuf, it contains one segment, with a length of 0. The offset to data is initialized to have some bytes of headroom in the buffer (RTE_PKTMBUF_HEADROOM).

Freeing a mbuf means returning it into its original mempool. The content of an mbuf is not modified when it is stored in a pool (as a free mbuf). Fields initialized by the constructor do not need to be re-initialized at mbuf allocation.

When freeing a packet mbuf that contains several segments, all of them are freed and returned to their original mempool.

Manipulating mbufs

This library provides some functions for manipulating the data in a packet mbuf. For instance:

  • Get data length
  • Get a pointer to the start of data
  • Prepend data before data
  • Append data after data
  • Remove data at the beginning of the buffer (rte_pktmbuf_adj())
  • Remove data at the end of the buffer (rte_pktmbuf_trim()) Refer to the DPDK API Reference for details.

Meta Information

Some information is retrieved by the network driver and stored in an mbuf to make processing easier. For instance, the VLAN, the RSS hash result (see Poll Mode Driver) and a flag indicating that the checksum was computed by hardware.

An mbuf also contains the input port (where it comes from), and the number of segment mbufs in the chain.

For chained buffers, only the first mbuf of the chain stores this meta information.

For instance, this is the case on RX side for the IEEE1588 packet timestamp mechanism, the VLAN tagging and the IP checksum computation.

On TX side, it is also possible for an application to delegate some processing to the hardware if it supports it. For instance, the PKT_TX_IP_CKSUM flag allows to offload the computation of the IPv4 checksum.

The following examples explain how to configure different TX offloads on a vxlan-encapsulated tcp packet: out_eth/out_ip/out_udp/vxlan/in_eth/in_ip/in_tcp/payload

  • calculate checksum of out_ip:

    mb->l2_len = len(out_eth)
    mb->l3_len = len(out_ip)
    mb->ol_flags |= PKT_TX_IPV4 | PKT_TX_IP_CSUM
    set out_ip checksum to 0 in the packet
    

    This is supported on hardware advertising DEV_TX_OFFLOAD_IPV4_CKSUM.

  • calculate checksum of out_ip and out_udp:

    mb->l2_len = len(out_eth)
    mb->l3_len = len(out_ip)
    mb->ol_flags |= PKT_TX_IPV4 | PKT_TX_IP_CSUM | PKT_TX_UDP_CKSUM
    set out_ip checksum to 0 in the packet
    set out_udp checksum to pseudo header using rte_ipv4_phdr_cksum()
    

    This is supported on hardware advertising DEV_TX_OFFLOAD_IPV4_CKSUM and DEV_TX_OFFLOAD_UDP_CKSUM.

  • calculate checksum of in_ip:

    mb->l2_len = len(out_eth + out_ip + out_udp + vxlan + in_eth)
    mb->l3_len = len(in_ip)
    mb->ol_flags |= PKT_TX_IPV4 | PKT_TX_IP_CSUM
    set in_ip checksum to 0 in the packet
    

    This is similar to case 1), but l2_len is different. It is supported on hardware advertising DEV_TX_OFFLOAD_IPV4_CKSUM. Note that it can only work if outer L4 checksum is 0.

  • calculate checksum of in_ip and in_tcp:

    mb->l2_len = len(out_eth + out_ip + out_udp + vxlan + in_eth)
    mb->l3_len = len(in_ip)
    mb->ol_flags |= PKT_TX_IPV4 | PKT_TX_IP_CSUM | PKT_TX_TCP_CKSUM
    set in_ip checksum to 0 in the packet
    set in_tcp checksum to pseudo header using rte_ipv4_phdr_cksum()
    

    This is similar to case 2), but l2_len is different. It is supported on hardware advertising DEV_TX_OFFLOAD_IPV4_CKSUM and DEV_TX_OFFLOAD_TCP_CKSUM. Note that it can only work if outer L4 checksum is 0.

  • segment inner TCP:

    mb->l2_len = len(out_eth + out_ip + out_udp + vxlan + in_eth)
    mb->l3_len = len(in_ip)
    mb->l4_len = len(in_tcp)
    mb->ol_flags |= PKT_TX_IPV4 | PKT_TX_IP_CKSUM | PKT_TX_TCP_CKSUM |
      PKT_TX_TCP_SEG;
    set in_ip checksum to 0 in the packet
    set in_tcp checksum to pseudo header without including the IP
      payload length using rte_ipv4_phdr_cksum()
    

    This is supported on hardware advertising DEV_TX_OFFLOAD_TCP_TSO. Note that it can only work if outer L4 checksum is 0.

  • calculate checksum of out_ip, in_ip, in_tcp:

    mb->outer_l2_len = len(out_eth)
    mb->outer_l3_len = len(out_ip)
    mb->l2_len = len(out_udp + vxlan + in_eth)
    mb->l3_len = len(in_ip)
    mb->ol_flags |= PKT_TX_OUTER_IPV4 | PKT_TX_OUTER_IP_CKSUM  | \
      PKT_TX_IP_CKSUM |  PKT_TX_TCP_CKSUM;
    set out_ip checksum to 0 in the packet
    set in_ip checksum to 0 in the packet
    set in_tcp checksum to pseudo header using rte_ipv4_phdr_cksum()
    

    This is supported on hardware advertising DEV_TX_OFFLOAD_IPV4_CKSUM, DEV_TX_OFFLOAD_UDP_CKSUM and DEV_TX_OFFLOAD_OUTER_IPV4_CKSUM.

The list of flags and their precise meaning is described in the mbuf API documentation (rte_mbuf.h). Also refer to the testpmd source code (specifically the csumonly.c file) for details.

Direct and Indirect Buffers

A direct buffer is a buffer that is completely separate and self-contained. An indirect buffer behaves like a direct buffer but for the fact that the buffer pointer and data offset in it refer to data in another direct buffer. This is useful in situations where packets need to be duplicated or fragmented, since indirect buffers provide the means to reuse the same packet data across multiple buffers.

A buffer becomes indirect when it is “attached” to a direct buffer using the rte_pktmbuf_attach() function. Each buffer has a reference counter field and whenever an indirect buffer is attached to the direct buffer, the reference counter on the direct buffer is incremented. Similarly, whenever the indirect buffer is detached, the reference counter on the direct buffer is decremented. If the resulting reference counter is equal to 0, the direct buffer is freed since it is no longer in use.

There are a few things to remember when dealing with indirect buffers. First of all, an indirect buffer is never attached to another indirect buffer. Attempting to attach buffer A to indirect buffer B that is attached to C, makes rte_pktmbuf_attach() automatically attach A to C, effectively cloning B. Secondly, for a buffer to become indirect, its reference counter must be equal to 1, that is, it must not be already referenced by another indirect buffer. Finally, it is not possible to reattach an indirect buffer to the direct buffer (unless it is detached first).

While the attach/detach operations can be invoked directly using the recommended rte_pktmbuf_attach() and rte_pktmbuf_detach() functions, it is suggested to use the higher-level rte_pktmbuf_clone() function, which takes care of the correct initialization of an indirect buffer and can clone buffers with multiple segments.

Since indirect buffers are not supposed to actually hold any data, the memory pool for indirect buffers should be configured to indicate the reduced memory consumption. Examples of the initialization of a memory pool for indirect buffers (as well as use case examples for indirect buffers) can be found in several of the sample applications, for example, the IPv4 Multicast sample application.

Debug

In debug mode (CONFIG_RTE_MBUF_DEBUG is enabled), the functions of the mbuf library perform sanity checks before any operation (such as, buffer corruption, bad type, and so on).

Use Cases

All networking application should use mbufs to transport network packets.

Poll Mode Driver

The DPDK includes 1 Gigabit, 10 Gigabit and 40 Gigabit and para virtualized virtio Poll Mode Drivers.

A Poll Mode Driver (PMD) consists of APIs, provided through the BSD driver running in user space, to configure the devices and their respective queues. In addition, a PMD accesses the RX and TX descriptors directly without any interrupts (with the exception of Link Status Change interrupts) to quickly receive, process and deliver packets in the user’s application. This section describes the requirements of the PMDs, their global design principles and proposes a high-level architecture and a generic external API for the Ethernet PMDs.

Requirements and Assumptions

The DPDK environment for packet processing applications allows for two models, run-to-completion and pipe-line:

  • In the run-to-completion model, a specific port’s RX descriptor ring is polled for packets through an API. Packets are then processed on the same core and placed on a port’s TX descriptor ring through an API for transmission.
  • In the pipe-line model, one core polls one or more port’s RX descriptor ring through an API. Packets are received and passed to another core via a ring. The other core continues to process the packet which then may be placed on a port’s TX descriptor ring through an API for transmission.

In a synchronous run-to-completion model, each logical core assigned to the DPDK executes a packet processing loop that includes the following steps:

  • Retrieve input packets through the PMD receive API
  • Process each received packet one at a time, up to its forwarding
  • Send pending output packets through the PMD transmit API

Conversely, in an asynchronous pipe-line model, some logical cores may be dedicated to the retrieval of received packets and other logical cores to the processing of previously received packets. Received packets are exchanged between logical cores through rings. The loop for packet retrieval includes the following steps:

  • Retrieve input packets through the PMD receive API
  • Provide received packets to processing lcores through packet queues

The loop for packet processing includes the following steps:

  • Retrieve the received packet from the packet queue
  • Process the received packet, up to its retransmission if forwarded

To avoid any unnecessary interrupt processing overhead, the execution environment must not use any asynchronous notification mechanisms. Whenever needed and appropriate, asynchronous communication should be introduced as much as possible through the use of rings.

Avoiding lock contention is a key issue in a multi-core environment. To address this issue, PMDs are designed to work with per-core private resources as much as possible. For example, a PMD maintains a separate transmit queue per-core, per-port, if the PMD is not DEV_TX_OFFLOAD_MT_LOCKFREE capable. In the same way, every receive queue of a port is assigned to and polled by a single logical core (lcore).

To comply with Non-Uniform Memory Access (NUMA), memory management is designed to assign to each logical core a private buffer pool in local memory to minimize remote memory access. The configuration of packet buffer pools should take into account the underlying physical memory architecture in terms of DIMMS, channels and ranks. The application must ensure that appropriate parameters are given at memory pool creation time. See Mempool Library.

Design Principles

The API and architecture of the Ethernet* PMDs are designed with the following guidelines in mind.

PMDs must help global policy-oriented decisions to be enforced at the upper application level. Conversely, NIC PMD functions should not impede the benefits expected by upper-level global policies, or worse prevent such policies from being applied.

For instance, both the receive and transmit functions of a PMD have a maximum number of packets/descriptors to poll. This allows a run-to-completion processing stack to statically fix or to dynamically adapt its overall behavior through different global loop policies, such as:

  • Receive, process immediately and transmit packets one at a time in a piecemeal fashion.
  • Receive as many packets as possible, then process all received packets, transmitting them immediately.
  • Receive a given maximum number of packets, process the received packets, accumulate them and finally send all accumulated packets to transmit.

To achieve optimal performance, overall software design choices and pure software optimization techniques must be considered and balanced against available low-level hardware-based optimization features (CPU cache properties, bus speed, NIC PCI bandwidth, and so on). The case of packet transmission is an example of this software/hardware tradeoff issue when optimizing burst-oriented network packet processing engines. In the initial case, the PMD could export only an rte_eth_tx_one function to transmit one packet at a time on a given queue. On top of that, one can easily build an rte_eth_tx_burst function that loops invoking the rte_eth_tx_one function to transmit several packets at a time. However, an rte_eth_tx_burst function is effectively implemented by the PMD to minimize the driver-level transmit cost per packet through the following optimizations:

  • Share among multiple packets the un-amortized cost of invoking the rte_eth_tx_one function.
  • Enable the rte_eth_tx_burst function to take advantage of burst-oriented hardware features (prefetch data in cache, use of NIC head/tail registers) to minimize the number of CPU cycles per packet, for example by avoiding unnecessary read memory accesses to ring transmit descriptors, or by systematically using arrays of pointers that exactly fit cache line boundaries and sizes.
  • Apply burst-oriented software optimization techniques to remove operations that would otherwise be unavoidable, such as ring index wrap back management.

Burst-oriented functions are also introduced via the API for services that are intensively used by the PMD. This applies in particular to buffer allocators used to populate NIC rings, which provide functions to allocate/free several buffers at a time. For example, an mbuf_multiple_alloc function returning an array of pointers to rte_mbuf buffers which speeds up the receive poll function of the PMD when replenishing multiple descriptors of the receive ring.

Logical Cores, Memory and NIC Queues Relationships

The DPDK supports NUMA allowing for better performance when a processor’s logical cores and interfaces utilize its local memory. Therefore, mbuf allocation associated with local PCIe* interfaces should be allocated from memory pools created in the local memory. The buffers should, if possible, remain on the local processor to obtain the best performance results and RX and TX buffer descriptors should be populated with mbufs allocated from a mempool allocated from local memory.

The run-to-completion model also performs better if packet or data manipulation is in local memory instead of a remote processors memory. This is also true for the pipe-line model provided all logical cores used are located on the same processor.

Multiple logical cores should never share receive or transmit queues for interfaces since this would require global locks and hinder performance.

If the PMD is DEV_TX_OFFLOAD_MT_LOCKFREE capable, multiple threads can invoke rte_eth_tx_burst() concurrently on the same tx queue without SW lock. This PMD feature found in some NICs and useful in the following use cases:

  • Remove explicit spinlock in some applications where lcores are not mapped to Tx queues with 1:1 relation.
  • In the eventdev use case, avoid dedicating a separate TX core for transmitting and thus enables more scaling as all workers can send the packets.

See Hardware Offload for DEV_TX_OFFLOAD_MT_LOCKFREE capability probing details.

Device Identification and Configuration

Device Identification

Each NIC port is uniquely designated by its (bus/bridge, device, function) PCI identifiers assigned by the PCI probing/enumeration function executed at DPDK initialization. Based on their PCI identifier, NIC ports are assigned two other identifiers:

  • A port index used to designate the NIC port in all functions exported by the PMD API.
  • A port name used to designate the port in console messages, for administration or debugging purposes. For ease of use, the port name includes the port index.
Device Configuration

The configuration of each NIC port includes the following operations:

  • Allocate PCI resources
  • Reset the hardware (issue a Global Reset) to a well-known default state
  • Set up the PHY and the link
  • Initialize statistics counters

The PMD API must also export functions to start/stop the all-multicast feature of a port and functions to set/unset the port in promiscuous mode.

Some hardware offload features must be individually configured at port initialization through specific configuration parameters. This is the case for the Receive Side Scaling (RSS) and Data Center Bridging (DCB) features for example.

On-the-Fly Configuration

All device features that can be started or stopped “on the fly” (that is, without stopping the device) do not require the PMD API to export dedicated functions for this purpose.

All that is required is the mapping address of the device PCI registers to implement the configuration of these features in specific functions outside of the drivers.

For this purpose, the PMD API exports a function that provides all the information associated with a device that can be used to set up a given device feature outside of the driver. This includes the PCI vendor identifier, the PCI device identifier, the mapping address of the PCI device registers, and the name of the driver.

The main advantage of this approach is that it gives complete freedom on the choice of the API used to configure, to start, and to stop such features.

As an example, refer to the configuration of the IEEE1588 feature for the Intel® 82576 Gigabit Ethernet Controller and the Intel® 82599 10 Gigabit Ethernet Controller controllers in the testpmd application.

Other features such as the L3/L4 5-Tuple packet filtering feature of a port can be configured in the same way. Ethernet* flow control (pause frame) can be configured on the individual port. Refer to the testpmd source code for details. Also, L4 (UDP/TCP/ SCTP) checksum offload by the NIC can be enabled for an individual packet as long as the packet mbuf is set up correctly. See Hardware Offload for details.

Configuration of Transmit Queues

Each transmit queue is independently configured with the following information:

  • The number of descriptors of the transmit ring
  • The socket identifier used to identify the appropriate DMA memory zone from which to allocate the transmit ring in NUMA architectures
  • The values of the Prefetch, Host and Write-Back threshold registers of the transmit queue
  • The minimum transmit packets to free threshold (tx_free_thresh). When the number of descriptors used to transmit packets exceeds this threshold, the network adaptor should be checked to see if it has written back descriptors. A value of 0 can be passed during the TX queue configuration to indicate the default value should be used. The default value for tx_free_thresh is 32. This ensures that the PMD does not search for completed descriptors until at least 32 have been processed by the NIC for this queue.
  • The minimum RS bit threshold. The minimum number of transmit descriptors to use before setting the Report Status (RS) bit in the transmit descriptor. Note that this parameter may only be valid for Intel 10 GbE network adapters. The RS bit is set on the last descriptor used to transmit a packet if the number of descriptors used since the last RS bit setting, up to the first descriptor used to transmit the packet, exceeds the transmit RS bit threshold (tx_rs_thresh). In short, this parameter controls which transmit descriptors are written back to host memory by the network adapter. A value of 0 can be passed during the TX queue configuration to indicate that the default value should be used. The default value for tx_rs_thresh is 32. This ensures that at least 32 descriptors are used before the network adapter writes back the most recently used descriptor. This saves upstream PCIe* bandwidth resulting from TX descriptor write-backs. It is important to note that the TX Write-back threshold (TX wthresh) should be set to 0 when tx_rs_thresh is greater than 1. Refer to the Intel® 82599 10 Gigabit Ethernet Controller Datasheet for more details.

The following constraints must be satisfied for tx_free_thresh and tx_rs_thresh:

  • tx_rs_thresh must be greater than 0.
  • tx_rs_thresh must be less than the size of the ring minus 2.
  • tx_rs_thresh must be less than or equal to tx_free_thresh.
  • tx_free_thresh must be greater than 0.
  • tx_free_thresh must be less than the size of the ring minus 3.
  • For optimal performance, TX wthresh should be set to 0 when tx_rs_thresh is greater than 1.

One descriptor in the TX ring is used as a sentinel to avoid a hardware race condition, hence the maximum threshold constraints.

Note

When configuring for DCB operation, at port initialization, both the number of transmit queues and the number of receive queues must be set to 128.

Free Tx mbuf on Demand

Many of the drivers do not release the mbuf back to the mempool, or local cache, immediately after the packet has been transmitted. Instead, they leave the mbuf in their Tx ring and either perform a bulk release when the tx_rs_thresh has been crossed or free the mbuf when a slot in the Tx ring is needed.

An application can request the driver to release used mbufs with the rte_eth_tx_done_cleanup() API. This API requests the driver to release mbufs that are no longer in use, independent of whether or not the tx_rs_thresh has been crossed. There are two scenarios when an application may want the mbuf released immediately:

  • When a given packet needs to be sent to multiple destination interfaces (either for Layer 2 flooding or Layer 3 multi-cast). One option is to make a copy of the packet or a copy of the header portion that needs to be manipulated. A second option is to transmit the packet and then poll the rte_eth_tx_done_cleanup() API until the reference count on the packet is decremented. Then the same packet can be transmitted to the next destination interface. The application is still responsible for managing any packet manipulations needed between the different destination interfaces, but a packet copy can be avoided. This API is independent of whether the packet was transmitted or dropped, only that the mbuf is no longer in use by the interface.
  • Some applications are designed to make multiple runs, like a packet generator. For performance reasons and consistency between runs, the application may want to reset back to an initial state between each run, where all mbufs are returned to the mempool. In this case, it can call the rte_eth_tx_done_cleanup() API for each destination interface it has been using to request it to release of all its used mbufs.

To determine if a driver supports this API, check for the Free Tx mbuf on demand feature in the Network Interface Controller Drivers document.

Hardware Offload

Depending on driver capabilities advertised by rte_eth_dev_info_get(), the PMD may support hardware offloading feature like checksumming, TCP segmentation, VLAN insertion or lockfree multithreaded TX burst on the same TX queue.

The support of these offload features implies the addition of dedicated status bit(s) and value field(s) into the rte_mbuf data structure, along with their appropriate handling by the receive/transmit functions exported by each PMD. The list of flags and their precise meaning is described in the mbuf API documentation and in the in Mbuf Library, section “Meta Information”.

Per-Port and Per-Queue Offloads

In the DPDK offload API, offloads are divided into per-port and per-queue offloads. The different offloads capabilities can be queried using rte_eth_dev_info_get(). Supported offloads can be either per-port or per-queue.

Offloads are enabled using the existing DEV_TX_OFFLOAD_* or DEV_RX_OFFLOAD_* flags. Per-port offload configuration is set using rte_eth_dev_configure. Per-queue offload configuration is set using rte_eth_rx_queue_setup and rte_eth_tx_queue_setup. To enable per-port offload, the offload should be set on both device configuration and queue setup. In case of a mixed configuration the queue setup shall return with an error. To enable per-queue offload, the offload can be set only on the queue setup. Offloads which are not enabled are disabled by default.

For an application to use the Tx offloads API it should set the ETH_TXQ_FLAGS_IGNORE flag in the txq_flags field located in rte_eth_txconf struct. In such cases it is not required to set other flags in txq_flags. For an application to use the Rx offloads API it should set the ignore_offload_bitfield bit in the rte_eth_rxmode struct. In such cases it is not required to set other bitfield offloads in the rxmode struct.

Poll Mode Driver API

Generalities

By default, all functions exported by a PMD are lock-free functions that are assumed not to be invoked in parallel on different logical cores to work on the same target object. For instance, a PMD receive function cannot be invoked in parallel on two logical cores to poll the same RX queue of the same port. Of course, this function can be invoked in parallel by different logical cores on different RX queues. It is the responsibility of the upper-level application to enforce this rule.

If needed, parallel accesses by multiple logical cores to shared queues can be explicitly protected by dedicated inline lock-aware functions built on top of their corresponding lock-free functions of the PMD API.

Generic Packet Representation

A packet is represented by an rte_mbuf structure, which is a generic metadata structure containing all necessary housekeeping information. This includes fields and status bits corresponding to offload hardware features, such as checksum computation of IP headers or VLAN tags.

The rte_mbuf data structure includes specific fields to represent, in a generic way, the offload features provided by network controllers. For an input packet, most fields of the rte_mbuf structure are filled in by the PMD receive function with the information contained in the receive descriptor. Conversely, for output packets, most fields of rte_mbuf structures are used by the PMD transmit function to initialize transmit descriptors.

The mbuf structure is fully described in the Mbuf Library chapter.

Ethernet Device API

The Ethernet device API exported by the Ethernet PMDs is described in the DPDK API Reference.

Extended Statistics API

The extended statistics API allows a PMD to expose all statistics that are available to it, including statistics that are unique to the device. Each statistic has three properties name, id and value:

  • name: A human readable string formatted by the scheme detailed below.
  • id: An integer that represents only that statistic.
  • value: A unsigned 64-bit integer that is the value of the statistic.

Note that extended statistic identifiers are driver-specific, and hence might not be the same for different ports. The API consists of various rte_eth_xstats_*() functions, and allows an application to be flexible in how it retrieves statistics.

Scheme for Human Readable Names

A naming scheme exists for the strings exposed to clients of the API. This is to allow scraping of the API for statistics of interest. The naming scheme uses strings split by a single underscore _. The scheme is as follows:

  • direction
  • detail 1
  • detail 2
  • detail n
  • unit

Examples of common statistics xstats strings, formatted to comply to the scheme proposed above:

  • rx_bytes
  • rx_crc_errors
  • tx_multicast_packets

The scheme, although quite simple, allows flexibility in presenting and reading information from the statistic strings. The following example illustrates the naming scheme:rx_packets. In this example, the string is split into two components. The first component rx indicates that the statistic is associated with the receive side of the NIC. The second component packets indicates that the unit of measure is packets.

A more complicated example: tx_size_128_to_255_packets. In this example, tx indicates transmission, size is the first detail, 128 etc are more details, and packets indicates that this is a packet counter.

Some additions in the metadata scheme are as follows:

  • If the first part does not match rx or tx, the statistic does not have an affinity with either receive of transmit.
  • If the first letter of the second part is q and this q is followed by a number, this statistic is part of a specific queue.

An example where queue numbers are used is as follows: tx_q7_bytes which indicates this statistic applies to queue number 7, and represents the number of transmitted bytes on that queue.

API Design

The xstats API uses the name, id, and value to allow performant lookup of specific statistics. Performant lookup means two things;

  • No string comparisons with the name of the statistic in fast-path
  • Allow requesting of only the statistics of interest

The API ensures these requirements are met by mapping the name of the statistic to a unique id, which is used as a key for lookup in the fast-path. The API allows applications to request an array of id values, so that the PMD only performs the required calculations. Expected usage is that the application scans the name of each statistic, and caches the id if it has an interest in that statistic. On the fast-path, the integer can be used to retrieve the actual value of the statistic that the id represents.

API Functions

The API is built out of a small number of functions, which can be used to retrieve the number of statistics and the names, IDs and values of those statistics.

  • rte_eth_xstats_get_names_by_id(): returns the names of the statistics. When given a NULL parameter the function returns the number of statistics that are available.
  • rte_eth_xstats_get_id_by_name(): Searches for the statistic ID that matches xstat_name. If found, the id integer is set.
  • rte_eth_xstats_get_by_id(): Fills in an array of uint64_t values with matching the provided ids array. If the ids array is NULL, it returns all statistics that are available.
Application Usage

Imagine an application that wants to view the dropped packet count. If no packets are dropped, the application does not read any other metrics for performance reasons. If packets are dropped, the application has a particular set of statistics that it requests. This “set” of statistics allows the app to decide what next steps to perform. The following code-snippets show how the xstats API can be used to achieve this goal.

First step is to get all statistics names and list them:

struct rte_eth_xstat_name *xstats_names;
uint64_t *values;
int len, i;

/* Get number of stats */
len = rte_eth_xstats_get_names_by_id(port_id, NULL, NULL, 0);
if (len < 0) {
    printf("Cannot get xstats count\n");
    goto err;
}

xstats_names = malloc(sizeof(struct rte_eth_xstat_name) * len);
if (xstats_names == NULL) {
    printf("Cannot allocate memory for xstat names\n");
    goto err;
}

/* Retrieve xstats names, passing NULL for IDs to return all statistics */
if (len != rte_eth_xstats_get_names_by_id(port_id, xstats_names, NULL, len)) {
    printf("Cannot get xstat names\n");
    goto err;
}

values = malloc(sizeof(values) * len);
if (values == NULL) {
    printf("Cannot allocate memory for xstats\n");
    goto err;
}

/* Getting xstats values */
if (len != rte_eth_xstats_get_by_id(port_id, NULL, values, len)) {
    printf("Cannot get xstat values\n");
    goto err;
}

/* Print all xstats names and values */
for (i = 0; i < len; i++) {
    printf("%s: %"PRIu64"\n", xstats_names[i].name, values[i]);
}

The application has access to the names of all of the statistics that the PMD exposes. The application can decide which statistics are of interest, cache the ids of those statistics by looking up the name as follows:

uint64_t id;
uint64_t value;
const char *xstat_name = "rx_errors";

if(!rte_eth_xstats_get_id_by_name(port_id, xstat_name, &id)) {
    rte_eth_xstats_get_by_id(port_id, &id, &value, 1);
    printf("%s: %"PRIu64"\n", xstat_name, value);
}
else {
    printf("Cannot find xstats with a given name\n");
    goto err;
}

The API provides flexibility to the application so that it can look up multiple statistics using an array containing multiple id numbers. This reduces the function call overhead of retrieving statistics, and makes lookup of multiple statistics simpler for the application.

#define APP_NUM_STATS 4
/* application cached these ids previously; see above */
uint64_t ids_array[APP_NUM_STATS] = {3,4,7,21};
uint64_t value_array[APP_NUM_STATS];

/* Getting multiple xstats values from array of IDs */
rte_eth_xstats_get_by_id(port_id, ids_array, value_array, APP_NUM_STATS);

uint32_t i;
for(i = 0; i < APP_NUM_STATS; i++) {
    printf("%d: %"PRIu64"\n", ids_array[i], value_array[i]);
}

This array lookup API for xstats allows the application create multiple “groups” of statistics, and look up the values of those IDs using a single API call. As an end result, the application is able to achieve its goal of monitoring a single statistic (“rx_errors” in this case), and if that shows packets being dropped, it can easily retrieve a “set” of statistics using the IDs array parameter to rte_eth_xstats_get_by_id function.

NIC Reset API
int rte_eth_dev_reset(uint16_t port_id);

Sometimes a port has to be reset passively. For example when a PF is reset, all its VFs should also be reset by the application to make them consistent with the PF. A DPDK application also can call this function to trigger a port reset. Normally, a DPDK application would invokes this function when an RTE_ETH_EVENT_INTR_RESET event is detected.

It is the duty of the PMD to trigger RTE_ETH_EVENT_INTR_RESET events and the application should register a callback function to handle these events. When a PMD needs to trigger a reset, it can trigger an RTE_ETH_EVENT_INTR_RESET event. On receiving an RTE_ETH_EVENT_INTR_RESET event, applications can handle it as follows: Stop working queues, stop calling Rx and Tx functions, and then call rte_eth_dev_reset(). For thread safety all these operations should be called from the same thread.

For example when PF is reset, the PF sends a message to notify VFs of this event and also trigger an interrupt to VFs. Then in the interrupt service routine the VFs detects this notification message and calls _rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_RESET, NULL, NULL). This means that a PF reset triggers an RTE_ETH_EVENT_INTR_RESET event within VFs. The function _rte_eth_dev_callback_process() will call the registered callback function. The callback function can trigger the application to handle all operations the VF reset requires including stopping Rx/Tx queues and calling rte_eth_dev_reset().

The rte_eth_dev_reset() itself is a generic function which only does some hardware reset operations through calling dev_unint() and dev_init(), and itself does not handle synchronization, which is handled by application.

The PMD itself should not call rte_eth_dev_reset(). The PMD can trigger the application to handle reset event. It is duty of application to handle all synchronization before it calls rte_eth_dev_reset().

Generic flow API (rte_flow)

Overview

This API provides a generic means to configure hardware to match specific ingress or egress traffic, alter its fate and query related counters according to any number of user-defined rules.

It is named rte_flow after the prefix used for all its symbols, and is defined in rte_flow.h.

  • Matching can be performed on packet data (protocol headers, payload) and properties (e.g. associated physical port, virtual device function ID).
  • Possible operations include dropping traffic, diverting it to specific queues, to virtual/physical device functions or ports, performing tunnel offloads, adding marks and so on.

It is slightly higher-level than the legacy filtering framework which it encompasses and supersedes (including all functions and filter types) in order to expose a single interface with an unambiguous behavior that is common to all poll-mode drivers (PMDs).

Several methods to migrate existing applications are described in API migration.

Flow rule

Description

A flow rule is the combination of attributes with a matching pattern and a list of actions. Flow rules form the basis of this API.

Flow rules can have several distinct actions (such as counting, encapsulating, decapsulating before redirecting packets to a particular queue, etc.), instead of relying on several rules to achieve this and having applications deal with hardware implementation details regarding their order.

Support for different priority levels on a rule basis is provided, for example in order to force a more specific rule to come before a more generic one for packets matched by both. However hardware support for more than a single priority level cannot be guaranteed. When supported, the number of available priority levels is usually low, which is why they can also be implemented in software by PMDs (e.g. missing priority levels may be emulated by reordering rules).

In order to remain as hardware-agnostic as possible, by default all rules are considered to have the same priority, which means that the order between overlapping rules (when a packet is matched by several filters) is undefined.

PMDs may refuse to create overlapping rules at a given priority level when they can be detected (e.g. if a pattern matches an existing filter).

Thus predictable results for a given priority level can only be achieved with non-overlapping rules, using perfect matching on all protocol layers.

Flow rules can also be grouped, the flow rule priority is specific to the group they belong to. All flow rules in a given group are thus processed either before or after another group.

Support for multiple actions per rule may be implemented internally on top of non-default hardware priorities, as a result both features may not be simultaneously available to applications.

Considering that allowed pattern/actions combinations cannot be known in advance and would result in an impractically large number of capabilities to expose, a method is provided to validate a given rule from the current device configuration state.

This enables applications to check if the rule types they need is supported at initialization time, before starting their data path. This method can be used anytime, its only requirement being that the resources needed by a rule should exist (e.g. a target RX queue should be configured first).

Each defined rule is associated with an opaque handle managed by the PMD, applications are responsible for keeping it. These can be used for queries and rules management, such as retrieving counters or other data and destroying them.

To avoid resource leaks on the PMD side, handles must be explicitly destroyed by the application before releasing associated resources such as queues and ports.

The following sections cover:

  • Attributes (represented by struct rte_flow_attr): properties of a flow rule such as its direction (ingress or egress) and priority.
  • Pattern item (represented by struct rte_flow_item): part of a matching pattern that either matches specific packet data or traffic properties. It can also describe properties of the pattern itself, such as inverted matching.
  • Matching pattern: traffic properties to look for, a combination of any number of items.
  • Actions (represented by struct rte_flow_action): operations to perform whenever a packet is matched by a pattern.
Attributes
Attribute: Group

Flow rules can be grouped by assigning them a common group number. Lower values have higher priority. Group 0 has the highest priority.

Although optional, applications are encouraged to group similar rules as much as possible to fully take advantage of hardware capabilities (e.g. optimized matching) and work around limitations (e.g. a single pattern type possibly allowed in a given group).

Note that support for more than a single group is not guaranteed.

Attribute: Priority

A priority level can be assigned to a flow rule. Like groups, lower values denote higher priority, with 0 as the maximum.

A rule with priority 0 in group 8 is always matched after a rule with priority 8 in group 0.

Group and priority levels are arbitrary and up to the application, they do not need to be contiguous nor start from 0, however the maximum number varies between devices and may be affected by existing flow rules.

If a packet is matched by several rules of a given group for a given priority level, the outcome is undefined. It can take any path, may be duplicated or even cause unrecoverable errors.

Note that support for more than a single priority level is not guaranteed.

Attribute: Traffic direction

Flow rules can apply to inbound and/or outbound traffic (ingress/egress).

Several pattern items and actions are valid and can be used in both directions. At least one direction must be specified.

Specifying both directions at once for a given rule is not recommended but may be valid in a few cases (e.g. shared counters).

Pattern item

Pattern items fall in two categories:

  • Matching protocol headers and packet data (ANY, RAW, ETH, VLAN, IPV4, IPV6, ICMP, UDP, TCP, SCTP, VXLAN, MPLS, GRE, ESP and so on), usually associated with a specification structure.
  • Matching meta-data or affecting pattern processing (END, VOID, INVERT, PF, VF, PORT and so on), often without a specification structure.

Item specification structures are used to match specific values among protocol fields (or item properties). Documentation describes for each item whether they are associated with one and their type name if so.

Up to three structures of the same type can be set for a given item:

  • spec: values to match (e.g. a given IPv4 address).
  • last: upper bound for an inclusive range with corresponding fields in spec.
  • mask: bit-mask applied to both spec and last whose purpose is to distinguish the values to take into account and/or partially mask them out (e.g. in order to match an IPv4 address prefix).

Usage restrictions and expected behavior:

  • Setting either mask or last without spec is an error.
  • Field values in last which are either 0 or equal to the corresponding values in spec are ignored; they do not generate a range. Nonzero values lower than those in spec are not supported.
  • Setting spec and optionally last without mask causes the PMD to use the default mask defined for that item (defined as rte_flow_item_{name}_mask constants).
  • Not setting any of them (assuming item type allows it) is equivalent to providing an empty (zeroed) mask for broad (nonspecific) matching.
  • mask is a simple bit-mask applied before interpreting the contents of spec and last, which may yield unexpected results if not used carefully. For example, if for an IPv4 address field, spec provides 10.1.2.3, last provides 10.3.4.5 and mask provides 255.255.0.0, the effective range becomes 10.1.0.0 to 10.3.255.255.

Example of an item specification matching an Ethernet header:

Ethernet item
Field Subfield Value
spec src 00:01:02:03:04
dst 00:2a:66:00:01
type 0x22aa
last unspecified
mask src 00:ff:ff:ff:00
dst 00:00:00:00:ff
type 0x0000

Non-masked bits stand for any value (shown as ? below), Ethernet headers with the following properties are thus matched:

  • src: ??:01:02:03:??
  • dst: ??:??:??:??:01
  • type: 0x????
Matching pattern

A pattern is formed by stacking items starting from the lowest protocol layer to match. This stacking restriction does not apply to meta items which can be placed anywhere in the stack without affecting the meaning of the resulting pattern.

Patterns are terminated by END items.

Examples:

TCPv4 as L4
Index Item
0 Ethernet
1 IPv4
2 TCP
3 END

TCPv6 in VXLAN
Index Item
0 Ethernet
1 IPv4
2 UDP
3 VXLAN
4 Ethernet
5 IPv6
6 TCP
7 END

TCPv4 as L4 with meta items
Index Item
0 VOID
1 Ethernet
2 VOID
3 IPv4
4 TCP
5 VOID
6 VOID
7 END

The above example shows how meta items do not affect packet data matching items, as long as those remain stacked properly. The resulting matching pattern is identical to “TCPv4 as L4”.

UDPv6 anywhere
Index Item
0 IPv6
1 UDP
2 END

If supported by the PMD, omitting one or several protocol layers at the bottom of the stack as in the above example (missing an Ethernet specification) enables looking up anywhere in packets.

It is unspecified whether the payload of supported encapsulations (e.g. VXLAN payload) is matched by such a pattern, which may apply to inner, outer or both packets.

Invalid, missing L3
Index Item
0 Ethernet
1 UDP
2 END

The above pattern is invalid due to a missing L3 specification between L2 (Ethernet) and L4 (UDP). Doing so is only allowed at the bottom and at the top of the stack.

Meta item types

They match meta-data or affect pattern processing instead of matching packet data directly, most of them do not need a specification structure. This particularity allows them to be specified anywhere in the stack without causing any side effect.

Item: END

End marker for item lists. Prevents further processing of items, thereby ending the pattern.

  • Its numeric value is 0 for convenience.
  • PMD support is mandatory.
  • spec, last and mask are ignored.
END
Field Value
spec ignored
last ignored
mask ignored
Item: VOID

Used as a placeholder for convenience. It is ignored and simply discarded by PMDs.

  • PMD support is mandatory.
  • spec, last and mask are ignored.
VOID
Field Value
spec ignored
last ignored
mask ignored

One usage example for this type is generating rules that share a common prefix quickly without reallocating memory, only by updating item types:

TCP, UDP or ICMP as L4
Index Item
0 Ethernet
1 IPv4
2 UDP VOID VOID
3 VOID TCP VOID
4 VOID VOID ICMP
5 END
Item: INVERT

Inverted matching, i.e. process packets that do not match the pattern.

  • spec, last and mask are ignored.
INVERT
Field Value
spec ignored
last ignored
mask ignored

Usage example, matching non-TCPv4 packets only:

Anything but TCPv4
Index Item
0 INVERT
1 Ethernet
2 IPv4
3 TCP
4 END
Item: PF

Matches packets addressed to the physical function of the device.

If the underlying device function differs from the one that would normally receive the matched traffic, specifying this item prevents it from reaching that device unless the flow rule contains a Action: PF. Packets are not duplicated between device instances by default.

  • Likely to return an error or never match any traffic if applied to a VF device.
  • Can be combined with any number of Item: VF to match both PF and VF traffic.
  • spec, last and mask must not be set.
PF
Field Value
spec unset
last unset
mask unset
Item: VF

Matches packets addressed to a virtual function ID of the device.

If the underlying device function differs from the one that would normally receive the matched traffic, specifying this item prevents it from reaching that device unless the flow rule contains a Action: VF. Packets are not duplicated between device instances by default.

  • Likely to return an error or never match any traffic if this causes a VF device to match traffic addressed to a different VF.
  • Can be specified multiple times to match traffic addressed to several VF IDs.
  • Can be combined with a PF item to match both PF and VF traffic.
  • Default mask matches any VF ID.
VF
Field Subfield Value
spec id destination VF ID
last id upper range value
mask id zeroed to match any VF ID
Item: PORT

Matches packets coming from the specified physical port of the underlying device.

The first PORT item overrides the physical port normally associated with the specified DPDK input port (port_id). This item can be provided several times to match additional physical ports.

Note that physical ports are not necessarily tied to DPDK input ports (port_id) when those are not under DPDK control. Possible values are specific to each device, they are not necessarily indexed from zero and may not be contiguous.

As a device property, the list of allowed values as well as the value associated with a port_id should be retrieved by other means.

  • Default mask matches any port index.
PORT
Field Subfield Value
spec index physical port index
last index upper range value
mask index zeroed to match any port index
Data matching item types

Most of these are basically protocol header definitions with associated bit-masks. They must be specified (stacked) from lowest to highest protocol layer to form a matching pattern.

The following list is not exhaustive, new protocols will be added in the future.

Item: ANY

Matches any protocol in place of the current layer, a single ANY may also stand for several protocol layers.

This is usually specified as the first pattern item when looking for a protocol anywhere in a packet.

  • Default mask stands for any number of layers.
ANY
Field Subfield Value
spec num number of layers covered
last num upper range value
mask num zeroed to cover any number of layers

Example for VXLAN TCP payload matching regardless of outer L3 (IPv4 or IPv6) and L4 (UDP) both matched by the first ANY specification, and inner L3 (IPv4 or IPv6) matched by the second ANY specification:

TCP in VXLAN with wildcards
Index Item Field Subfield Value
0 Ethernet
1 ANY spec num 2
2 VXLAN
3 Ethernet
4 ANY spec num 1
5 TCP
6 END
Item: RAW

Matches a byte string of a given length at a given offset.

Offset is either absolute (using the start of the packet) or relative to the end of the previous matched item in the stack, in which case negative values are allowed.

If search is enabled, offset is used as the starting point. The search area can be delimited by setting limit to a nonzero value, which is the maximum number of bytes after offset where the pattern may start.

Matching a zero-length pattern is allowed, doing so resets the relative offset for subsequent items.

  • This type does not support ranges (last field).
  • Default mask matches all fields exactly.
RAW
Field Subfield Value
spec relative look for pattern after the previous item
search search pattern from offset (see also limit)
reserved reserved, must be set to zero
offset absolute or relative offset for pattern
limit search area limit for start of pattern
length pattern length
pattern byte string to look for
last if specified, either all 0 or with the same values as spec
mask bit-mask applied to spec values with usual behavior

Example pattern looking for several strings at various offsets of a UDP payload, using combined RAW items:

UDP payload matching
Index Item Field Subfield Value
0 Ethernet
1 IPv4
2 UDP
3 RAW spec relative 1
search 1
offset 10
limit 0
length 3
pattern “foo”
4 RAW spec relative 1
search 0
offset 20
limit 0
length 3
pattern “bar”
5 RAW spec relative 1
search 0
offset -29
limit 0
length 3
pattern “baz”
6 END

This translates to:

  • Locate “foo” at least 10 bytes deep inside UDP payload.
  • Locate “bar” after “foo” plus 20 bytes.
  • Locate “baz” after “bar” minus 29 bytes.

Such a packet may be represented as follows (not to scale):

0                     >= 10 B           == 20 B
|                  |<--------->|     |<--------->|
|                  |           |     |           |
|-----|------|-----|-----|-----|-----|-----------|-----|------|
| ETH | IPv4 | UDP | ... | baz | foo | ......... | bar | .... |
|-----|------|-----|-----|-----|-----|-----------|-----|------|
                         |                             |
                         |<--------------------------->|
                                     == 29 B

Note that matching subsequent pattern items would resume after “baz”, not “bar” since matching is always performed after the previous item of the stack.

Item: ETH

Matches an Ethernet header.

  • dst: destination MAC.
  • src: source MAC.
  • type: EtherType.
  • Default mask matches destination and source addresses only.
Item: VLAN

Matches an 802.1Q/ad VLAN tag.

  • tpid: tag protocol identifier.
  • tci: tag control information.
  • Default mask matches TCI only.
Item: IPV4

Matches an IPv4 header.

Note: IPv4 options are handled by dedicated pattern items.

  • hdr: IPv4 header definition (rte_ip.h).
  • Default mask matches source and destination addresses only.
Item: IPV6

Matches an IPv6 header.

Note: IPv6 options are handled by dedicated pattern items.

  • hdr: IPv6 header definition (rte_ip.h).
  • Default mask matches source and destination addresses only.
Item: ICMP

Matches an ICMP header.

  • hdr: ICMP header definition (rte_icmp.h).
  • Default mask matches ICMP type and code only.
Item: UDP

Matches a UDP header.

  • hdr: UDP header definition (rte_udp.h).
  • Default mask matches source and destination ports only.
Item: TCP

Matches a TCP header.

  • hdr: TCP header definition (rte_tcp.h).
  • Default mask matches source and destination ports only.
Item: SCTP

Matches a SCTP header.

  • hdr: SCTP header definition (rte_sctp.h).
  • Default mask matches source and destination ports only.
Item: VXLAN

Matches a VXLAN header (RFC 7348).

  • flags: normally 0x08 (I flag).
  • rsvd0: reserved, normally 0x000000.
  • vni: VXLAN network identifier.
  • rsvd1: reserved, normally 0x00.
  • Default mask matches VNI only.
Item: E_TAG

Matches an IEEE 802.1BR E-Tag header.

  • tpid: tag protocol identifier (0x893F)
  • epcp_edei_in_ecid_b: E-Tag control information (E-TCI), E-PCP (3b), E-DEI (1b), ingress E-CID base (12b).
  • rsvd_grp_ecid_b: reserved (2b), GRP (2b), E-CID base (12b).
  • in_ecid_e: ingress E-CID ext.
  • ecid_e: E-CID ext.
  • Default mask simultaneously matches GRP and E-CID base.
Item: NVGRE

Matches a NVGRE header (RFC 7637).

  • c_k_s_rsvd0_ver: checksum (1b), undefined (1b), key bit (1b), sequence number (1b), reserved 0 (9b), version (3b). This field must have value 0x2000 according to RFC 7637.
  • protocol: protocol type (0x6558).
  • tni: virtual subnet ID.
  • flow_id: flow ID.
  • Default mask matches TNI only.
Item: MPLS

Matches a MPLS header.

  • label_tc_s_ttl: label, TC, Bottom of Stack and TTL.
  • Default mask matches label only.
Item: GRE

Matches a GRE header.

  • c_rsvd0_ver: checksum, reserved 0 and version.
  • protocol: protocol type.
  • Default mask matches protocol only.
Item: FUZZY

Fuzzy pattern match, expect faster than default.

This is for device that support fuzzy match option. Usually a fuzzy match is fast but the cost is accuracy. i.e. Signature Match only match pattern’s hash value, but it is possible two different patterns have the same hash value.

Matching accuracy level can be configured by threshold. Driver can divide the range of threshold and map to different accuracy levels that device support.

Threshold 0 means perfect match (no fuzziness), while threshold 0xffffffff means fuzziest match.

FUZZY
Field Subfield Value
spec threshold 0 as perfect match, 0xffffffff as fuzziest match
last threshold upper range value
mask threshold bit-mask apply to “spec” and “last”

Usage example, fuzzy match a TCPv4 packets:

Fuzzy matching
Index Item
0 FUZZY
1 Ethernet
2 IPv4
3 TCP
4 END
Item: GTP, GTPC, GTPU

Matches a GTPv1 header.

Note: GTP, GTPC and GTPU use the same structure. GTPC and GTPU item are defined for a user-friendly API when creating GTP-C and GTP-U flow rules.

  • v_pt_rsv_flags: version (3b), protocol type (1b), reserved (1b), extension header flag (1b), sequence number flag (1b), N-PDU number flag (1b).
  • msg_type: message type.
  • msg_len: message length.
  • teid: tunnel endpoint identifier.
  • Default mask matches teid only.
Item: ESP

Matches an ESP header.

  • hdr: ESP header definition (rte_esp.h).
  • Default mask matches SPI only.
Actions

Each possible action is represented by a type. Some have associated configuration structures. Several actions combined in a list can be affected to a flow rule. That list is not ordered.

They fall in three categories:

  • Terminating actions (such as QUEUE, DROP, RSS, PF, VF) that prevent processing matched packets by subsequent flow rules, unless overridden with PASSTHRU.
  • Non-terminating actions (PASSTHRU, DUP) that leave matched packets up for additional processing by subsequent flow rules.
  • Other non-terminating meta actions that do not affect the fate of packets (END, VOID, MARK, FLAG, COUNT, SECURITY).

When several actions are combined in a flow rule, they should all have different types (e.g. dropping a packet twice is not possible).

Only the last action of a given type is taken into account. PMDs still perform error checking on the entire list.

Like matching patterns, action lists are terminated by END items.

Note that PASSTHRU is the only action able to override a terminating rule.

Example of action that redirects packets to queue index 10:

Queue action
Field Value
index 10

Action lists examples, their order is not significant, applications must consider all actions to be performed simultaneously:

Count and drop
Index Action
0 COUNT
1 DROP
2 END

Mark, count and redirect
Index Action Field Value
0 MARK mark 0x2a
1 COUNT
2 QUEUE queue 10
3 END

Redirect to queue 5
Index Action Field Value
0 DROP
1 QUEUE queue 5
2 END

In the above example, considering both actions are performed simultaneously, the end result is that only QUEUE has any effect.

Redirect to queue 3
Index Action Field Value
0 QUEUE queue 5
1 VOID
2 QUEUE queue 3
3 END

As previously described, only the last action of a given type found in the list is taken into account. The above example also shows that VOID is ignored.

Action types

Common action types are described in this section. Like pattern item types, this list is not exhaustive as new actions will be added in the future.

Action: END

End marker for action lists. Prevents further processing of actions, thereby ending the list.

  • Its numeric value is 0 for convenience.
  • PMD support is mandatory.
  • No configurable properties.
END
Field
no properties
Action: VOID

Used as a placeholder for convenience. It is ignored and simply discarded by PMDs.

  • PMD support is mandatory.
  • No configurable properties.
VOID
Field
no properties
Action: PASSTHRU

Leaves packets up for additional processing by subsequent flow rules. This is the default when a rule does not contain a terminating action, but can be specified to force a rule to become non-terminating.

  • No configurable properties.
PASSTHRU
Field
no properties

Example to copy a packet to a queue and continue processing by subsequent flow rules:

Copy to queue 8
Index Action Field Value
0 PASSTHRU
1 QUEUE queue 8
2 END
Action: MARK

Attaches an integer value to packets and sets PKT_RX_FDIR and PKT_RX_FDIR_ID mbuf flags.

This value is arbitrary and application-defined. Maximum allowed value depends on the underlying implementation. It is returned in the hash.fdir.hi mbuf field.

MARK
Field Value
id integer value to return with packets
Action: FLAG

Flags packets. Similar to Action: MARK without a specific value; only sets the PKT_RX_FDIR mbuf flag.

  • No configurable properties.
FLAG
Field
no properties
Action: QUEUE

Assigns packets to a given queue index.

  • Terminating by default.
QUEUE
Field Value
index queue index to use
Action: DROP

Drop packets.

  • No configurable properties.
  • Terminating by default.
  • PASSTHRU overrides this action if both are specified.
DROP
Field
no properties
Action: COUNT

Enables counters for this rule.

These counters can be retrieved and reset through rte_flow_query(), see struct rte_flow_query_count.

  • Counters can be retrieved with rte_flow_query().
  • No configurable properties.
COUNT
Field
no properties

Query structure to retrieve and reset flow rule counters:

COUNT query
Field I/O Value
reset in reset counter after query
hits_set out hits field is set
bytes_set out bytes field is set
hits out number of hits for this rule
bytes out number of bytes through this rule
Action: DUP

Duplicates packets to a given queue index.

This is normally combined with QUEUE, however when used alone, it is actually similar to QUEUE + PASSTHRU.

  • Non-terminating by default.
DUP
Field Value
index queue index to duplicate packet to
Action: RSS

Similar to QUEUE, except RSS is additionally performed on packets to spread them among several queues according to the provided parameters.

Note: RSS hash result is stored in the hash.rss mbuf field which overlaps hash.fdir.lo. Since Action: MARK sets the hash.fdir.hi field only, both can be requested simultaneously.

  • Terminating by default.
RSS
Field Value
rss_conf RSS parameters
num number of entries in queue[]
queue[] queue indices to use
Action: PF

Redirects packets to the physical function (PF) of the current device.

  • No configurable properties.
  • Terminating by default.
PF
Field
no properties
Action: VF

Redirects packets to a virtual function (VF) of the current device.

Packets matched by a VF pattern item can be redirected to their original VF ID instead of the specified one. This parameter may not be available and is not guaranteed to work properly if the VF part is matched by a prior flow rule or if packets are not addressed to a VF in the first place.

  • Terminating by default.
VF
Field Value
original use original VF ID if possible
vf VF ID to redirect packets to
Action: METER

Applies a stage of metering and policing.

The metering and policing (MTR) object has to be first created using the rte_mtr_create() API function. The ID of the MTR object is specified as action parameter. More than one flow can use the same MTR object through the meter action. The MTR object can be further updated or queried using the rte_mtr* API.

  • Non-terminating by default.
METER
Field Value
mtr_id MTR object ID
Action: SECURITY

Perform the security action on flows matched by the pattern items according to the configuration of the security session.

This action modifies the payload of matched flows. For INLINE_CRYPTO, the security protocol headers and IV are fully provided by the application as specified in the flow pattern. The payload of matching packets is encrypted on egress, and decrypted and authenticated on ingress. For INLINE_PROTOCOL, the security protocol is fully offloaded to HW, providing full encapsulation and decapsulation of packets in security protocols. The flow pattern specifies both the outer security header fields and the inner packet fields. The security session specified in the action must match the pattern parameters.

The security session specified in the action must be created on the same port as the flow action that is being specified.

The ingress/egress flow attribute should match that specified in the security session if the security session supports the definition of the direction.

Multiple flows can be configured to use the same security session.

  • Non-terminating by default.
SECURITY
Field Value
security_session security session to apply

The following is an example of configuring IPsec inline using the INLINE_CRYPTO security session:

The encryption algorithm, keys and salt are part of the opaque rte_security_session. The SA is identified according to the IP and ESP fields in the pattern items.

IPsec inline crypto flow pattern items.
Index Item
0 Ethernet
1 IPv4
2 ESP
3 END
IPsec inline flow actions.
Index Action
0 SECURITY
1 END
Negative types

All specified pattern items (enum rte_flow_item_type) and actions (enum rte_flow_action_type) use positive identifiers.

The negative space is reserved for dynamic types generated by PMDs during run-time. PMDs may encounter them as a result but must not accept negative identifiers they are not aware of.

A method to generate them remains to be defined.

Planned types

Pattern item types will be added as new protocols are implemented.

Variable headers support through dedicated pattern items, for example in order to match specific IPv4 options and IPv6 extension headers would be stacked after IPv4/IPv6 items.

Other action types are planned but are not defined yet. These include the ability to alter packet data in several ways, such as performing encapsulation/decapsulation of tunnel headers.

Rules management

A rather simple API with few functions is provided to fully manage flow rules.

Each created flow rule is associated with an opaque, PMD-specific handle pointer. The application is responsible for keeping it until the rule is destroyed.

Flows rules are represented by struct rte_flow objects.

Validation

Given that expressing a definite set of device capabilities is not practical, a dedicated function is provided to check if a flow rule is supported and can be created.

int
rte_flow_validate(uint16_t port_id,
                  const struct rte_flow_attr *attr,
                  const struct rte_flow_item pattern[],
                  const struct rte_flow_action actions[],
                  struct rte_flow_error *error);

The flow rule is validated for correctness and whether it could be accepted by the device given sufficient resources. The rule is checked against the current device mode and queue configuration. The flow rule may also optionally be validated against existing flow rules and device resources. This function has no effect on the target device.

The returned value is guaranteed to remain valid only as long as no successful calls to rte_flow_create() or rte_flow_destroy() are made in the meantime and no device parameter affecting flow rules in any way are modified, due to possible collisions or resource limitations (although in such cases EINVAL should not be returned).

Arguments:

  • port_id: port identifier of Ethernet device.
  • attr: flow rule attributes.
  • pattern: pattern specification (list terminated by the END pattern item).
  • actions: associated actions (list terminated by the END action).
  • error: perform verbose error reporting if not NULL. PMDs initialize this structure in case of error only.

Return values:

  • 0 if flow rule is valid and can be created. A negative errno value otherwise (rte_errno is also set), the following errors are defined.
  • -ENOSYS: underlying device does not support this functionality.
  • -EINVAL: unknown or invalid rule specification.
  • -ENOTSUP: valid but unsupported rule specification (e.g. partial bit-masks are unsupported).
  • EEXIST: collision with an existing rule. Only returned if device supports flow rule collision checking and there was a flow rule collision. Not receiving this return code is no guarantee that creating the rule will not fail due to a collision.
  • ENOMEM: not enough memory to execute the function, or if the device supports resource validation, resource limitation on the device.
  • -EBUSY: action cannot be performed due to busy device resources, may succeed if the affected queues or even the entire port are in a stopped state (see rte_eth_dev_rx_queue_stop() and rte_eth_dev_stop()).
Creation

Creating a flow rule is similar to validating one, except the rule is actually created and a handle returned.

struct rte_flow *
rte_flow_create(uint16_t port_id,
                const struct rte_flow_attr *attr,
                const struct rte_flow_item pattern[],
                const struct rte_flow_action *actions[],
                struct rte_flow_error *error);

Arguments:

  • port_id: port identifier of Ethernet device.
  • attr: flow rule attributes.
  • pattern: pattern specification (list terminated by the END pattern item).
  • actions: associated actions (list terminated by the END action).
  • error: perform verbose error reporting if not NULL. PMDs initialize this structure in case of error only.

Return values:

A valid handle in case of success, NULL otherwise and rte_errno is set to the positive version of one of the error codes defined for rte_flow_validate().

Destruction

Flow rules destruction is not automatic, and a queue or a port should not be released if any are still attached to them. Applications must take care of performing this step before releasing resources.

int
rte_flow_destroy(uint16_t port_id,
                 struct rte_flow *flow,
                 struct rte_flow_error *error);

Failure to destroy a flow rule handle may occur when other flow rules depend on it, and destroying it would result in an inconsistent state.

This function is only guaranteed to succeed if handles are destroyed in reverse order of their creation.

Arguments:

  • port_id: port identifier of Ethernet device.
  • flow: flow rule handle to destroy.
  • error: perform verbose error reporting if not NULL. PMDs initialize this structure in case of error only.

Return values:

  • 0 on success, a negative errno value otherwise and rte_errno is set.
Flush

Convenience function to destroy all flow rule handles associated with a port. They are released as with successive calls to rte_flow_destroy().

int
rte_flow_flush(uint16_t port_id,
               struct rte_flow_error *error);

In the unlikely event of failure, handles are still considered destroyed and no longer valid but the port must be assumed to be in an inconsistent state.

Arguments:

  • port_id: port identifier of Ethernet device.
  • error: perform verbose error reporting if not NULL. PMDs initialize this structure in case of error only.

Return values:

  • 0 on success, a negative errno value otherwise and rte_errno is set.
Query

Query an existing flow rule.

This function allows retrieving flow-specific data such as counters. Data is gathered by special actions which must be present in the flow rule definition.

int
rte_flow_query(uint16_t port_id,
               struct rte_flow *flow,
               enum rte_flow_action_type action,
               void *data,
               struct rte_flow_error *error);

Arguments:

  • port_id: port identifier of Ethernet device.
  • flow: flow rule handle to query.
  • action: action type to query.
  • data: pointer to storage for the associated query data type.
  • error: perform verbose error reporting if not NULL. PMDs initialize this structure in case of error only.

Return values:

  • 0 on success, a negative errno value otherwise and rte_errno is set.

Isolated mode

The general expectation for ingress traffic is that flow rules process it first; the remaining unmatched or pass-through traffic usually ends up in a queue (with or without RSS, locally or in some sub-device instance) depending on the global configuration settings of a port.

While fine from a compatibility standpoint, this approach makes drivers more complex as they have to check for possible side effects outside of this API when creating or destroying flow rules. It results in a more limited set of available rule types due to the way device resources are assigned (e.g. no support for the RSS action even on capable hardware).

Given that nonspecific traffic can be handled by flow rules as well, isolated mode is a means for applications to tell a driver that ingress on the underlying port must be injected from the defined flow rules only; that no default traffic is expected outside those rules.

This has the following benefits:

  • Applications get finer-grained control over the kind of traffic they want to receive (no traffic by default).
  • More importantly they control at what point nonspecific traffic is handled relative to other flow rules, by adjusting priority levels.
  • Drivers can assign more hardware resources to flow rules and expand the set of supported rule types.

Because toggling isolated mode may cause profound changes to the ingress processing path of a driver, it may not be possible to leave it once entered. Likewise, existing flow rules or global configuration settings may prevent a driver from entering isolated mode.

Applications relying on this mode are therefore encouraged to toggle it as soon as possible after device initialization, ideally before the first call to rte_eth_dev_configure() to avoid possible failures due to conflicting settings.

Once effective, the following functionality has no effect on the underlying port and may return errors such as ENOTSUP (“not supported”):

  • Toggling promiscuous mode.
  • Toggling allmulticast mode.
  • Configuring MAC addresses.
  • Configuring multicast addresses.
  • Configuring VLAN filters.
  • Configuring Rx filters through the legacy API (e.g. FDIR).
  • Configuring global RSS settings.
int
rte_flow_isolate(uint16_t port_id, int set, struct rte_flow_error *error);

Arguments:

  • port_id: port identifier of Ethernet device.
  • set: nonzero to enter isolated mode, attempt to leave it otherwise.
  • error: perform verbose error reporting if not NULL. PMDs initialize this structure in case of error only.

Return values:

  • 0 on success, a negative errno value otherwise and rte_errno is set.

Verbose error reporting

The defined errno values may not be accurate enough for users or application developers who want to investigate issues related to flow rules management. A dedicated error object is defined for this purpose:

enum rte_flow_error_type {
    RTE_FLOW_ERROR_TYPE_NONE, /**< No error. */
    RTE_FLOW_ERROR_TYPE_UNSPECIFIED, /**< Cause unspecified. */
    RTE_FLOW_ERROR_TYPE_HANDLE, /**< Flow rule (handle). */
    RTE_FLOW_ERROR_TYPE_ATTR_GROUP, /**< Group field. */
    RTE_FLOW_ERROR_TYPE_ATTR_PRIORITY, /**< Priority field. */
    RTE_FLOW_ERROR_TYPE_ATTR_INGRESS, /**< Ingress field. */
    RTE_FLOW_ERROR_TYPE_ATTR_EGRESS, /**< Egress field. */
    RTE_FLOW_ERROR_TYPE_ATTR, /**< Attributes structure. */
    RTE_FLOW_ERROR_TYPE_ITEM_NUM, /**< Pattern length. */
    RTE_FLOW_ERROR_TYPE_ITEM, /**< Specific pattern item. */
    RTE_FLOW_ERROR_TYPE_ACTION_NUM, /**< Number of actions. */
    RTE_FLOW_ERROR_TYPE_ACTION, /**< Specific action. */
};

struct rte_flow_error {
    enum rte_flow_error_type type; /**< Cause field and error types. */
    const void *cause; /**< Object responsible for the error. */
    const char *message; /**< Human-readable error message. */
};

Error type RTE_FLOW_ERROR_TYPE_NONE stands for no error, in which case remaining fields can be ignored. Other error types describe the type of the object pointed by cause.

If non-NULL, cause points to the object responsible for the error. For a flow rule, this may be a pattern item or an individual action.

If non-NULL, message provides a human-readable error message.

This object is normally allocated by applications and set by PMDs in case of error, the message points to a constant string which does not need to be freed by the application, however its pointer can be considered valid only as long as its associated DPDK port remains configured. Closing the underlying device or unloading the PMD invalidates it.

Helpers

Error initializer
static inline int
rte_flow_error_set(struct rte_flow_error *error,
                   int code,
                   enum rte_flow_error_type type,
                   const void *cause,
                   const char *message);

This function initializes error (if non-NULL) with the provided parameters and sets rte_errno to code. A negative error code is then returned.

Caveats

  • DPDK does not keep track of flow rules definitions or flow rule objects automatically. Applications may keep track of the former and must keep track of the latter. PMDs may also do it for internal needs, however this must not be relied on by applications.
  • Flow rules are not maintained between successive port initializations. An application exiting without releasing them and restarting must re-create them from scratch.
  • API operations are synchronous and blocking (EAGAIN cannot be returned).
  • There is no provision for reentrancy/multi-thread safety, although nothing should prevent different devices from being configured at the same time. PMDs may protect their control path functions accordingly.
  • Stopping the data path (TX/RX) should not be necessary when managing flow rules. If this cannot be achieved naturally or with workarounds (such as temporarily replacing the burst function pointers), an appropriate error code must be returned (EBUSY).
  • PMDs, not applications, are responsible for maintaining flow rules configuration when stopping and restarting a port or performing other actions which may affect them. They can only be destroyed explicitly by applications.

For devices exposing multiple ports sharing global settings affected by flow rules:

  • All ports under DPDK control must behave consistently, PMDs are responsible for making sure that existing flow rules on a port are not affected by other ports.
  • Ports not under DPDK control (unaffected or handled by other applications) are user’s responsibility. They may affect existing flow rules and cause undefined behavior. PMDs aware of this may prevent flow rules creation altogether in such cases.

PMD interface

The PMD interface is defined in rte_flow_driver.h. It is not subject to API/ABI versioning constraints as it is not exposed to applications and may evolve independently.

It is currently implemented on top of the legacy filtering framework through filter type RTE_ETH_FILTER_GENERIC that accepts the single operation RTE_ETH_FILTER_GET to return PMD-specific rte_flow callbacks wrapped inside struct rte_flow_ops.

This overhead is temporarily necessary in order to keep compatibility with the legacy filtering framework, which should eventually disappear.

  • PMD callbacks implement exactly the interface described in Rules management, except for the port ID argument which has already been converted to a pointer to the underlying struct rte_eth_dev.
  • Public API functions do not process flow rules definitions at all before calling PMD functions (no basic error checking, no validation whatsoever). They only make sure these callbacks are non-NULL or return the ENOSYS (function not supported) error.

This interface additionally defines the following helper function:

  • rte_flow_ops_get(): get generic flow operations structure from a port.

More will be added over time.

Device compatibility

No known implementation supports all the described features.

Unsupported features or combinations are not expected to be fully emulated in software by PMDs for performance reasons. Partially supported features may be completed in software as long as hardware performs most of the work (such as queue redirection and packet recognition).

However PMDs are expected to do their best to satisfy application requests by working around hardware limitations as long as doing so does not affect the behavior of existing flow rules.

The following sections provide a few examples of such cases and describe how PMDs should handle them, they are based on limitations built into the previous APIs.

Global bit-masks

Each flow rule comes with its own, per-layer bit-masks, while hardware may support only a single, device-wide bit-mask for a given layer type, so that two IPv4 rules cannot use different bit-masks.

The expected behavior in this case is that PMDs automatically configure global bit-masks according to the needs of the first flow rule created.

Subsequent rules are allowed only if their bit-masks match those, the EEXIST error code should be returned otherwise.

Unsupported layer types

Many protocols can be simulated by crafting patterns with the Item: RAW type.

PMDs can rely on this capability to simulate support for protocols with headers not directly recognized by hardware.

ANY pattern item

This pattern item stands for anything, which can be difficult to translate to something hardware would understand, particularly if followed by more specific types.

Consider the following pattern:

Pattern with ANY as L3
Index Item
0 ETHER
1 ANY num 1
2 TCP
3 END

Knowing that TCP does not make sense with something other than IPv4 and IPv6 as L3, such a pattern may be translated to two flow rules instead:

ANY replaced with IPV4
Index Item
0 ETHER
1 IPV4 (zeroed mask)
2 TCP
3 END

ANY replaced with IPV6
Index Item
0 ETHER
1 IPV6 (zeroed mask)
2 TCP
3 END

Note that as soon as a ANY rule covers several layers, this approach may yield a large number of hidden flow rules. It is thus suggested to only support the most common scenarios (anything as L2 and/or L3).

Unsupported actions
Flow rules priority

While it would naturally make sense, flow rules cannot be assumed to be processed by hardware in the same order as their creation for several reasons:

  • They may be managed internally as a tree or a hash table instead of a list.
  • Removing a flow rule before adding another one can either put the new rule at the end of the list or reuse a freed entry.
  • Duplication may occur when packets are matched by several rules.

For overlapping rules (particularly in order to use Action: PASSTHRU) predictable behavior is only guaranteed by using different priority levels.

Priority levels are not necessarily implemented in hardware, or may be severely limited (e.g. a single priority bit).

For these reasons, priority levels may be implemented purely in software by PMDs.

  • For devices expecting flow rules to be added in the correct order, PMDs may destroy and re-create existing rules after adding a new one with a higher priority.
  • A configurable number of dummy or empty rules can be created at initialization time to save high priority slots for later.
  • In order to save priority levels, PMDs may evaluate whether rules are likely to collide and adjust their priority accordingly.

Future evolutions

  • A device profile selection function which could be used to force a permanent profile instead of relying on its automatic configuration based on existing flow rules.
  • A method to optimize rte_flow rules with specific pattern items and action types generated on the fly by PMDs. DPDK should assign negative numbers to these in order to not collide with the existing types. See Negative types.
  • Adding specific egress pattern items and actions as described in Attribute: Traffic direction.
  • Optional software fallback when PMDs are unable to handle requested flow rules so applications do not have to implement their own.

API migration

Exhaustive list of deprecated filter types (normally prefixed with RTE_ETH_FILTER_) found in rte_eth_ctrl.h and methods to convert them to rte_flow rules.

MACVLAN to ETHVF, PF

MACVLAN can be translated to a basic Item: ETH flow rule with a terminating Action: VF or Action: PF.

MACVLAN conversion
Pattern Actions
0 ETH spec any VF, PF
last N/A
mask any
1 END END
ETHERTYPE to ETHQUEUE, DROP

ETHERTYPE is basically an Item: ETH flow rule with a terminating Action: QUEUE or Action: DROP.

ETHERTYPE conversion
Pattern Actions
0 ETH spec any QUEUE, DROP
last N/A
mask any
1 END END
FLEXIBLE to RAWQUEUE

FLEXIBLE can be translated to one Item: RAW pattern with a terminating Action: QUEUE and a defined priority level.

FLEXIBLE conversion
Pattern Actions
0 RAW spec any QUEUE
last N/A
mask any
1 END END
SYN to TCPQUEUE

SYN is a Item: TCP rule with only the syn bit enabled and masked, and a terminating Action: QUEUE.

Priority level can be set to simulate the high priority bit.

SYN conversion
Pattern Actions
0 ETH spec unset QUEUE
last unset
mask unset
1 IPV4 spec unset END
mask unset
mask unset
2 TCP spec syn 1
mask syn 1
3 END
NTUPLE to IPV4, TCP, UDPQUEUE

NTUPLE is similar to specifying an empty L2, Item: IPV4 as L3 with Item: TCP or Item: UDP as L4 and a terminating Action: QUEUE.

A priority level can be specified as well.

NTUPLE conversion
Pattern Actions
0 ETH spec unset QUEUE
last unset
mask unset
1 IPV4 spec any END
last unset
mask any
2 TCP, UDP spec any
last unset
mask any
3 END
TUNNEL to ETH, IPV4, IPV6, VXLAN (or other) → QUEUE

TUNNEL matches common IPv4 and IPv6 L3/L4-based tunnel types.

In the following table, Item: ANY is used to cover the optional L4.

TUNNEL conversion
Pattern Actions
0 ETH spec any QUEUE
last unset
mask any
1 IPV4, IPV6 spec any END
last unset
mask any
2 ANY spec any
last unset
mask num 0
3 VXLAN, GENEVE, TEREDO, NVGRE, GRE, ... spec any
last unset
mask any
4 END
FDIR to most item types → QUEUE, DROP, PASSTHRU

FDIR is more complex than any other type, there are several methods to emulate its functionality. It is summarized for the most part in the table below.

A few features are intentionally not supported:

  • The ability to configure the matching input set and masks for the entire device, PMDs should take care of it automatically according to the requested flow rules.

    For example if a device supports only one bit-mask per protocol type, source/address IPv4 bit-masks can be made immutable by the first created rule. Subsequent IPv4 or TCPv4 rules can only be created if they are compatible.

    Note that only protocol bit-masks affected by existing flow rules are immutable, others can be changed later. They become mutable again after the related flow rules are destroyed.

  • Returning four or eight bytes of matched data when using flex bytes filtering. Although a specific action could implement it, it conflicts with the much more useful 32 bits tagging on devices that support it.

  • Side effects on RSS processing of the entire device. Flow rules that conflict with the current device configuration should not be allowed. Similarly, device configuration should not be allowed when it affects existing flow rules.

  • Device modes of operation. “none” is unsupported since filtering cannot be disabled as long as a flow rule is present.

  • “MAC VLAN” or “tunnel” perfect matching modes should be automatically set according to the created flow rules.

  • Signature mode of operation is not defined but could be handled through “FUZZY” item.

FDIR conversion
Pattern Actions
0 ETH, RAW spec any QUEUE, DROP, PASSTHRU
last N/A
mask any
1 IPV4, IPv6 spec any MARK
last N/A
mask any
2 TCP, UDP, SCTP spec any END
last N/A
mask any
3 VF, PF, FUZZY (optional) spec any
last N/A
mask any
4 END
HASH

There is no counterpart to this filter type because it translates to a global device setting instead of a pattern item. Device settings are automatically set according to the created flow rules.

L2_TUNNEL to VOIDVXLAN (or others)

All packets are matched. This type alters incoming packets to encapsulate them in a chosen tunnel type, optionally redirect them to a VF as well.

The destination pool for tag based forwarding can be emulated with other flow rules using Action: DUP.

L2_TUNNEL conversion
Pattern Actions
0 VOID spec N/A VXLAN, GENEVE, ...
last N/A
mask N/A
1 END VF (optional)
2 END

Traffic Metering and Policing API

Overview

This is the generic API for the Quality of Service (QoS) Traffic Metering and Policing (MTR) of Ethernet devices. This API is agnostic of the underlying HW, SW or mixed HW-SW implementation.

The main features are:

  • Part of DPDK rte_ethdev API
  • Capability query API
  • Metering algorithms: RFC 2697 Single Rate Three Color Marker (srTCM), RFC 2698 and RFC 4115 Two Rate Three Color Marker (trTCM)
  • Policer actions (per meter output color): recolor, drop
  • Statistics (per policer output color)

Configuration steps

The metering and policing stage typically sits on top of flow classification, which is why the MTR objects are enabled through a special “meter” action.

The MTR objects are created and updated in their own name space (rte_mtr) within the librte_ether library. Whether an MTR object is private to a flow or potentially shared by several flows has to be specified at its creation time.

Once successfully created, an MTR object is hooked into the RX processing path of the Ethernet device by linking it to one or several flows through the dedicated “meter” flow action. One or several “meter” actions can be registered for the same flow. An MTR object can only be destroyed if there are no flows using it.

Run-time processing

Traffic metering determines the color for the current packet (green, yellow, red) based on the previous history for this flow as maintained by the MTR object. The policer can do nothing, override the color the packet or drop the packet. Statistics counters are maintained for MTR object, as configured.

The processing done for each input packet hitting an MTR object is:

  • Traffic metering: The packet is assigned a color (the meter output color) based on the previous traffic history reflected in the current state of the MTR object, according to the specific traffic metering algorithm. The traffic metering algorithm can typically work in color aware mode, in which case the input packet already has an initial color (the input color), or in color blind mode, which is equivalent to considering all input packets initially colored as green.
  • Policing: There is a separate policer action configured for each meter output color, which can:
    • Drop the packet.
    • Keep the same packet color: the policer output color matches the meter output color (essentially a no-op action).
    • Recolor the packet: the policer output color is set to a different color than the meter output color. The policer output color is the output color of the packet, which is set in the packet meta-data (i.e. struct rte_mbuf::sched::color).
  • Statistics: The set of counters maintained for each MTR object is configurable and subject to the implementation support. This set includes the number of packets and bytes dropped or passed for each output color.

Traffic Management API

Overview

This is the generic API for the Quality of Service (QoS) Traffic Management of Ethernet devices, which includes the following main features: hierarchical scheduling, traffic shaping, congestion management, packet marking. This API is agnostic of the underlying HW, SW or mixed HW-SW implementation.

Main features:

  • Part of DPDK rte_ethdev API
  • Capability query API per port, per hierarchy level and per hierarchy node
  • Scheduling algorithms: Strict Priority (SP), Weighed Fair Queuing (WFQ)
  • Traffic shaping: single/dual rate, private (per node) and shared (by multiple nodes) shapers
  • Congestion management for hierarchy leaf nodes: algorithms of tail drop, head drop, WRED, private (per node) and shared (by multiple nodes) WRED contexts
  • Packet marking: IEEE 802.1q (VLAN DEI), IETF RFC 3168 (IPv4/IPv6 ECN for TCP and SCTP), IETF RFC 2597 (IPv4 / IPv6 DSCP)

Capability API

The aim of these APIs is to advertise the capability information (i.e critical parameter values) that the TM implementation (HW/SW) is able to support for the application. The APIs supports the information disclosure at the TM level, at any hierarchical level of the TM and at any node level of the specific hierarchical level. Such information helps towards rapid understanding of whether a specific implementation does meet the needs to the user application.

At the TM level, users can get high level idea with the help of various parameters such as maximum number of nodes, maximum number of hierarchical levels, maximum number of shapers, maximum number of private shapers, type of scheduling algorithm (Strict Priority, Weighted Fair Queueing , etc.), etc., supported by the implementation.

Likewise, users can query the capability of the TM at the hierarchical level to have more granular knowledge about the specific level. The various parameters such as maximum number of nodes at the level, maximum number of leaf/non-leaf nodes at the level, type of the shaper(dual rate, single rate) supported at the level if node is non-leaf type etc., are exposed as a result of hierarchical level capability query.

Finally, the node level capability API offers knowledge about the capability supported by the node at any specific level. The information whether the support is available for private shaper, dual rate shaper, maximum and minimum shaper rate, etc. is exposed by node level capability API.

Scheduling Algorithms

The fundamental scheduling algorithms that are supported are Strict Priority (SP) and Weighted Fair Queuing (WFQ). The SP and WFQ algorithms are supported at the level of each node of the scheduling hierarchy, regardless of the node level/position in the tree. The SP algorithm is used to schedule between sibling nodes with different priority, while WFQ is used to schedule between groups of siblings that have the same priority.

Algorithms such as Weighed Round Robin (WRR), byte-level WRR, Deficit WRR (DWRR), etc are considered approximations of the ideal WFQ and are therefore assimilated to WFQ, although an associated implementation-dependent accuracy, performance and resource usage trade-off might exist.

Traffic Shaping

The TM API provides support for single rate and dual rate shapers (rate limiters) for the hierarchy nodes, subject to the specific implementation support being available.

Each hierarchy node has zero or one private shaper (only one node using it) and/or zero, one or several shared shapers (multiple nodes use the same shaper instance). A private shaper is used to perform traffic shaping for a single node, while a shared shaper is used to perform traffic shaping for a group of nodes.

The configuration of private and shared shapers is done through the definition of shaper profiles. Any shaper profile (single rate or dual rate shaper) can be used by one or several shaper instances (either private or shared).

Single rate shapers use a single token bucket. Therefore, single rate shaper is configured by setting the rate of the committed bucket to zero, which effectively disables this bucket. The peak bucket is used to limit the rate and the burst size for the single rate shaper. Dual rate shapers use both the committed and the peak token buckets. The rate of the peak bucket has to be bigger than zero, as well as greater than or equal to the rate of the committed bucket.

Congestion Management

Congestion management is used to control the admission of packets into a packet queue or group of packet queues on congestion. The congestion management algorithms that are supported are: Tail Drop, Head Drop and Weighted Random Early Detection (WRED). They are made available for every leaf node in the hierarchy, subject to the specific implementation supporting them. On request of writing a new packet into the current queue while the queue is full, the Tail Drop algorithm drops the new packet while leaving the queue unmodified, as opposed to the Head Drop* algorithm, which drops the packet at the head of the queue (the oldest packet waiting in the queue) and admits the new packet at the tail of the queue.

The Random Early Detection (RED) algorithm works by proactively dropping more and more input packets as the queue occupancy builds up. When the queue is full or almost full, RED effectively works as Tail Drop. The Weighted RED (WRED) algorithm uses a separate set of RED thresholds for each packet color and uses separate set of RED thresholds for each packet color.

Each hierarchy leaf node with WRED enabled as its congestion management mode has zero or one private WRED context (only one leaf node using it) and/or zero, one or several shared WRED contexts (multiple leaf nodes use the same WRED context). A private WRED context is used to perform congestion management for a single leaf node, while a shared WRED context is used to perform congestion management for a group of leaf nodes.

The configuration of WRED private and shared contexts is done through the definition of WRED profiles. Any WRED profile can be used by one or several WRED contexts (either private or shared).

Packet Marking

The TM APIs have been provided to support various types of packet marking such as VLAN DEI packet marking (IEEE 802.1Q), IPv4/IPv6 ECN marking of TCP and SCTP packets (IETF RFC 3168) and IPv4/IPv6 DSCP packet marking (IETF RFC 2597). All VLAN frames of a given color get their DEI bit set if marking is enabled for this color. In case, when marking for a given color is not enabled, the DEI bit is left as is (either set or not).

All IPv4/IPv6 packets of a given color with ECN set to 2’b01 or 2’b10 carrying TCP or SCTP have their ECN set to 2’b11 if the marking feature is enabled for the current color, otherwise the ECN field is left as is.

All IPv4/IPv6 packets have their color marked into DSCP bits 3 and 4 as follows: green mapped to Low Drop Precedence (2’b01), yellow to Medium (2’b10) and red to High (2’b11). Marking needs to be explicitly enabled for each color; when not enabled for a given color, the DSCP field of all packets with that color is left as is.

Steps to Setup the Hierarchy

The TM hierarchical tree consists of leaf nodes and non-leaf nodes. Each leaf node sits on top of a scheduling queue of the current Ethernet port. Therefore, the leaf nodes have predefined IDs in the range of 0... (N-1), where N is the number of scheduling queues of the current Ethernet port. The non-leaf nodes have their IDs generated by the application outside of the above range, which is reserved for leaf nodes.

Each non-leaf node has multiple inputs (its children nodes) and single output (which is input to its parent node). It arbitrates its inputs using Strict Priority (SP) and Weighted Fair Queuing (WFQ) algorithms to schedule input packets to its output while observing its shaping (rate limiting) constraints.

The children nodes with different priorities are scheduled using the SP algorithm based on their priority, with 0 as the highest priority. Children with the same priority are scheduled using the WFQ algorithm according to their weights. The WFQ weight of a given child node is relative to the sum of the weights of all its sibling nodes that have the same priority, with 1 as the lowest weight. For each SP priority, the WFQ weight mode can be set as either byte-based or packet-based.

Initial Hierarchy Specification

The hierarchy is specified by incrementally adding nodes to build up the scheduling tree. The first node that is added to the hierarchy becomes the root node and all the nodes that are subsequently added have to be added as descendants of the root node. The parent of the root node has to be specified as RTE_TM_NODE_ID_NULL and there can only be one node with this parent ID (i.e. the root node). The unique ID that is assigned to each node when the node is created is further used to update the node configuration or to connect children nodes to it.

During this phase, some limited checks on the hierarchy specification can be conducted, usually limited in scope to the current node, its parent node and its sibling nodes. At this time, since the hierarchy is not fully defined, there is typically no real action performed by the underlying implementation.

Hierarchy Commit

The hierarchy commit API is called during the port initialization phase (before the Ethernet port is started) to freeze the start-up hierarchy. This function typically performs the following steps:

  • It validates the start-up hierarchy that was previously defined for the current port through successive node add API invocations.
  • Assuming successful validation, it performs all the necessary implementation specific operations to install the specified hierarchy on the current port, with immediate effect once the port is started.

This function fails when the currently configured hierarchy is not supported by the Ethernet port, in which case the user can abort or try out another hierarchy configuration (e.g. a hierarchy with less leaf nodes), which can be built from scratch or by modifying the existing hierarchy configuration. Note that this function can still fail due to other causes (e.g. not enough memory available in the system, etc.), even though the specified hierarchy is supported in principle by the current port.

Run-Time Hierarchy Updates

The TM API provides support for on-the-fly changes to the scheduling hierarchy, thus operations such as node add/delete, node suspend/resume, parent node update, etc., can be invoked after the Ethernet port has been started, subject to the specific implementation supporting them. The set of dynamic updates supported by the implementation is advertised through the port capability set.

Cryptography Device Library

The cryptodev library provides a Crypto device framework for management and provisioning of hardware and software Crypto poll mode drivers, defining generic APIs which support a number of different Crypto operations. The framework currently only supports cipher, authentication, chained cipher/authentication and AEAD symmetric Crypto operations.

Design Principles

The cryptodev library follows the same basic principles as those used in DPDKs Ethernet Device framework. The Crypto framework provides a generic Crypto device framework which supports both physical (hardware) and virtual (software) Crypto devices as well as a generic Crypto API which allows Crypto devices to be managed and configured and supports Crypto operations to be provisioned on Crypto poll mode driver.

Device Management

Device Creation

Physical Crypto devices are discovered during the PCI probe/enumeration of the EAL function which is executed at DPDK initialization, based on their PCI device identifier, each unique PCI BDF (bus/bridge, device, function). Specific physical Crypto devices, like other physical devices in DPDK can be white-listed or black-listed using the EAL command line options.

Virtual devices can be created by two mechanisms, either using the EAL command line options or from within the application using an EAL API directly.

From the command line using the –vdev EAL option

--vdev  'crypto_aesni_mb0,max_nb_queue_pairs=2,max_nb_sessions=1024,socket_id=0'

Our using the rte_vdev_init API within the application code.

rte_vdev_init("crypto_aesni_mb",
                  "max_nb_queue_pairs=2,max_nb_sessions=1024,socket_id=0")

All virtual Crypto devices support the following initialization parameters:

  • max_nb_queue_pairs - maximum number of queue pairs supported by the device.
  • max_nb_sessions - maximum number of sessions supported by the device
  • socket_id - socket on which to allocate the device resources on.
Device Identification

Each device, whether virtual or physical is uniquely designated by two identifiers:

  • A unique device index used to designate the Crypto device in all functions exported by the cryptodev API.
  • A device name used to designate the Crypto device in console messages, for administration or debugging purposes. For ease of use, the port name includes the port index.
Device Configuration

The configuration of each Crypto device includes the following operations:

  • Allocation of resources, including hardware resources if a physical device.
  • Resetting the device into a well-known default state.
  • Initialization of statistics counters.

The rte_cryptodev_configure API is used to configure a Crypto device.

int rte_cryptodev_configure(uint8_t dev_id,
                            struct rte_cryptodev_config *config)

The rte_cryptodev_config structure is used to pass the configuration parameters for socket selection and number of queue pairs.

struct rte_cryptodev_config {
    int socket_id;
    /**< Socket to allocate resources on */
    uint16_t nb_queue_pairs;
    /**< Number of queue pairs to configure on device */
};
Configuration of Queue Pairs

Each Crypto devices queue pair is individually configured through the rte_cryptodev_queue_pair_setup API. Each queue pairs resources may be allocated on a specified socket.

int rte_cryptodev_queue_pair_setup(uint8_t dev_id, uint16_t queue_pair_id,
            const struct rte_cryptodev_qp_conf *qp_conf,
            int socket_id)

struct rte_cryptodev_qp_conf {
    uint32_t nb_descriptors; /**< Number of descriptors per queue pair */
};
Logical Cores, Memory and Queues Pair Relationships

The Crypto device Library as the Poll Mode Driver library support NUMA for when a processor’s logical cores and interfaces utilize its local memory. Therefore Crypto operations, and in the case of symmetric Crypto operations, the session and the mbuf being operated on, should be allocated from memory pools created in the local memory. The buffers should, if possible, remain on the local processor to obtain the best performance results and buffer descriptors should be populated with mbufs allocated from a mempool allocated from local memory.

The run-to-completion model also performs better, especially in the case of virtual Crypto devices, if the Crypto operation and session and data buffer is in local memory instead of a remote processor’s memory. This is also true for the pipe-line model provided all logical cores used are located on the same processor.

Multiple logical cores should never share the same queue pair for enqueuing operations or dequeuing operations on the same Crypto device since this would require global locks and hinder performance. It is however possible to use a different logical core to dequeue an operation on a queue pair from the logical core which it was enqueued on. This means that a crypto burst enqueue/dequeue APIs are a logical place to transition from one logical core to another in a packet processing pipeline.

Device Features and Capabilities

Crypto devices define their functionality through two mechanisms, global device features and algorithm capabilities. Global devices features identify device wide level features which are applicable to the whole device such as the device having hardware acceleration or supporting symmetric Crypto operations,

The capabilities mechanism defines the individual algorithms/functions which the device supports, such as a specific symmetric Crypto cipher, authentication operation or Authenticated Encryption with Associated Data (AEAD) operation.

Device Features

Currently the following Crypto device features are defined:

  • Symmetric Crypto operations
  • Asymmetric Crypto operations
  • Chaining of symmetric Crypto operations
  • SSE accelerated SIMD vector operations
  • AVX accelerated SIMD vector operations
  • AVX2 accelerated SIMD vector operations
  • AESNI accelerated instructions
  • Hardware off-load processing
Device Operation Capabilities

Crypto capabilities which identify particular algorithm which the Crypto PMD supports are defined by the operation type, the operation transform, the transform identifier and then the particulars of the transform. For the full scope of the Crypto capability see the definition of the structure in the DPDK API Reference.

struct rte_cryptodev_capabilities;

Each Crypto poll mode driver defines its own private array of capabilities for the operations it supports. Below is an example of the capabilities for a PMD which supports the authentication algorithm SHA1_HMAC and the cipher algorithm AES_CBC.

static const struct rte_cryptodev_capabilities pmd_capabilities[] = {
    {    /* SHA1 HMAC */
        .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC,
        .sym = {
            .xform_type = RTE_CRYPTO_SYM_XFORM_AUTH,
            .auth = {
                .algo = RTE_CRYPTO_AUTH_SHA1_HMAC,
                .block_size = 64,
                .key_size = {
                    .min = 64,
                    .max = 64,
                    .increment = 0
                },
                .digest_size = {
                    .min = 12,
                    .max = 12,
                    .increment = 0
                },
                .aad_size = { 0 },
                .iv_size = { 0 }
            }
        }
    },
    {    /* AES CBC */
        .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC,
        .sym = {
            .xform_type = RTE_CRYPTO_SYM_XFORM_CIPHER,
            .cipher = {
                .algo = RTE_CRYPTO_CIPHER_AES_CBC,
                .block_size = 16,
                .key_size = {
                    .min = 16,
                    .max = 32,
                    .increment = 8
                },
                .iv_size = {
                    .min = 16,
                    .max = 16,
                    .increment = 0
                }
            }
        }
    }
}
Capabilities Discovery

Discovering the features and capabilities of a Crypto device poll mode driver is achieved through the rte_cryptodev_info_get function.

void rte_cryptodev_info_get(uint8_t dev_id,
                            struct rte_cryptodev_info *dev_info);

This allows the user to query a specific Crypto PMD and get all the device features and capabilities. The rte_cryptodev_info structure contains all the relevant information for the device.

struct rte_cryptodev_info {
    const char *driver_name;
    uint8_t driver_id;
    struct rte_pci_device *pci_dev;

    uint64_t feature_flags;

    const struct rte_cryptodev_capabilities *capabilities;

    unsigned max_nb_queue_pairs;

    struct {
        unsigned max_nb_sessions;
    } sym;
};

Operation Processing

Scheduling of Crypto operations on DPDK’s application data path is performed using a burst oriented asynchronous API set. A queue pair on a Crypto device accepts a burst of Crypto operations using enqueue burst API. On physical Crypto devices the enqueue burst API will place the operations to be processed on the devices hardware input queue, for virtual devices the processing of the Crypto operations is usually completed during the enqueue call to the Crypto device. The dequeue burst API will retrieve any processed operations available from the queue pair on the Crypto device, from physical devices this is usually directly from the devices processed queue, and for virtual device’s from a rte_ring where processed operations are place after being processed on the enqueue call.

Enqueue / Dequeue Burst APIs

The burst enqueue API uses a Crypto device identifier and a queue pair identifier to specify the Crypto device queue pair to schedule the processing on. The nb_ops parameter is the number of operations to process which are supplied in the ops array of rte_crypto_op structures. The enqueue function returns the number of operations it actually enqueued for processing, a return value equal to nb_ops means that all packets have been enqueued.

uint16_t rte_cryptodev_enqueue_burst(uint8_t dev_id, uint16_t qp_id,
                                     struct rte_crypto_op **ops, uint16_t nb_ops)

The dequeue API uses the same format as the enqueue API of processed but the nb_ops and ops parameters are now used to specify the max processed operations the user wishes to retrieve and the location in which to store them. The API call returns the actual number of processed operations returned, this can never be larger than nb_ops.

uint16_t rte_cryptodev_dequeue_burst(uint8_t dev_id, uint16_t qp_id,
                                     struct rte_crypto_op **ops, uint16_t nb_ops)
Operation Representation

An Crypto operation is represented by an rte_crypto_op structure, which is a generic metadata container for all necessary information required for the Crypto operation to be processed on a particular Crypto device poll mode driver.

_images/crypto_op.svg

The operation structure includes the operation type, the operation status and the session type (session-based/less), a reference to the operation specific data, which can vary in size and content depending on the operation being provisioned. It also contains the source mempool for the operation, if it allocated from a mempool.

If Crypto operations are allocated from a Crypto operation mempool, see next section, there is also the ability to allocate private memory with the operation for applications purposes.

Application software is responsible for specifying all the operation specific fields in the rte_crypto_op structure which are then used by the Crypto PMD to process the requested operation.

Operation Management and Allocation

The cryptodev library provides an API set for managing Crypto operations which utilize the Mempool Library to allocate operation buffers. Therefore, it ensures that the crytpo operation is interleaved optimally across the channels and ranks for optimal processing. A rte_crypto_op contains a field indicating the pool that it originated from. When calling rte_crypto_op_free(op), the operation returns to its original pool.

extern struct rte_mempool *
rte_crypto_op_pool_create(const char *name, enum rte_crypto_op_type type,
                          unsigned nb_elts, unsigned cache_size, uint16_t priv_size,
                          int socket_id);

During pool creation rte_crypto_op_init() is called as a constructor to initialize each Crypto operation which subsequently calls __rte_crypto_op_reset() to configure any operation type specific fields based on the type parameter.

rte_crypto_op_alloc() and rte_crypto_op_bulk_alloc() are used to allocate Crypto operations of a specific type from a given Crypto operation mempool. __rte_crypto_op_reset() is called on each operation before being returned to allocate to a user so the operation is always in a good known state before use by the application.

struct rte_crypto_op *rte_crypto_op_alloc(struct rte_mempool *mempool,
                                          enum rte_crypto_op_type type)

unsigned rte_crypto_op_bulk_alloc(struct rte_mempool *mempool,
                                  enum rte_crypto_op_type type,
                                  struct rte_crypto_op **ops, uint16_t nb_ops)

rte_crypto_op_free() is called by the application to return an operation to its allocating pool.

void rte_crypto_op_free(struct rte_crypto_op *op)

Symmetric Cryptography Support

The cryptodev library currently provides support for the following symmetric Crypto operations; cipher, authentication, including chaining of these operations, as well as also supporting AEAD operations.

Session and Session Management

Sessions are used in symmetric cryptographic processing to store the immutable data defined in a cryptographic transform which is used in the operation processing of a packet flow. Sessions are used to manage information such as expand cipher keys and HMAC IPADs and OPADs, which need to be calculated for a particular Crypto operation, but are immutable on a packet to packet basis for a flow. Crypto sessions cache this immutable data in a optimal way for the underlying PMD and this allows further acceleration of the offload of Crypto workloads.

_images/cryptodev_sym_sess.svg

The Crypto device framework provides APIs to allocate and initizalize sessions for crypto devices, where sessions are mempool objects. It is the application’s responsibility to create and manage the session mempools. This approach allows for different scenarios such as having a single session mempool for all crypto devices (where the mempool object size is big enough to hold the private session of any crypto device), as well as having multiple session mempools of different sizes for better memory usage.

An application can use rte_cryptodev_get_private_session_size() to get the private session size of given crypto device. This function would allow an application to calculate the max device session size of all crypto devices to create a single session mempool. If instead an application creates multiple session mempools, the Crypto device framework also provides rte_cryptodev_get_header_session_size to get the size of an uninitialized session.

Once the session mempools have been created, rte_cryptodev_sym_session_create() is used to allocate an uninitialized session from the given mempool. The session then must be initialized using rte_cryptodev_sym_session_init() for each of the required crypto devices. A symmetric transform chain is used to specify the operation and its parameters. See the section below for details on transforms.

When a session is no longer used, user must call rte_cryptodev_sym_session_clear() for each of the crypto devices that are using the session, to free all driver private session data. Once this is done, session should be freed using rte_cryptodev_sym_session_free which returns them to their mempool.

Transforms and Transform Chaining

Symmetric Crypto transforms (rte_crypto_sym_xform) are the mechanism used to specify the details of the Crypto operation. For chaining of symmetric operations such as cipher encrypt and authentication generate, the next pointer allows transform to be chained together. Crypto devices which support chaining must publish the chaining of symmetric Crypto operations feature flag.

Currently there are three transforms types cipher, authentication and AEAD. Also it is important to note that the order in which the transforms are passed indicates the order of the chaining.

struct rte_crypto_sym_xform {
    struct rte_crypto_sym_xform *next;
    /**< next xform in chain */
    enum rte_crypto_sym_xform_type type;
    /**< xform type */
    union {
        struct rte_crypto_auth_xform auth;
        /**< Authentication / hash xform */
        struct rte_crypto_cipher_xform cipher;
        /**< Cipher xform */
        struct rte_crypto_aead_xform aead;
        /**< AEAD xform */
    };
};

The API does not place a limit on the number of transforms that can be chained together but this will be limited by the underlying Crypto device poll mode driver which is processing the operation.

_images/crypto_xform_chain.svg
Symmetric Operations

The symmetric Crypto operation structure contains all the mutable data relating to performing symmetric cryptographic processing on a referenced mbuf data buffer. It is used for either cipher, authentication, AEAD and chained operations.

As a minimum the symmetric operation must have a source data buffer (m_src), a valid session (or transform chain if in session-less mode) and the minimum authentication/ cipher/ AEAD parameters required depending on the type of operation specified in the session or the transform chain.

struct rte_crypto_sym_op {
    struct rte_mbuf *m_src;
    struct rte_mbuf *m_dst;

    union {
        struct rte_cryptodev_sym_session *session;
        /**< Handle for the initialised session context */
        struct rte_crypto_sym_xform *xform;
        /**< Session-less API Crypto operation parameters */
    };

    union {
        struct {
            struct {
                uint32_t offset;
                uint32_t length;
            } data; /**< Data offsets and length for AEAD */

            struct {
                uint8_t *data;
                rte_iova_t phys_addr;
            } digest; /**< Digest parameters */

            struct {
                uint8_t *data;
                rte_iova_t phys_addr;
            } aad;
            /**< Additional authentication parameters */
        } aead;

        struct {
            struct {
                struct {
                    uint32_t offset;
                    uint32_t length;
                } data; /**< Data offsets and length for ciphering */
            } cipher;

            struct {
                struct {
                    uint32_t offset;
                    uint32_t length;
                } data;
                /**< Data offsets and length for authentication */

                struct {
                    uint8_t *data;
                    rte_iova_t phys_addr;
                } digest; /**< Digest parameters */
            } auth;
        };
    };
};

Sample code

There are various sample applications that show how to use the cryptodev library, such as the L2fwd with Crypto sample application (L2fwd-crypto) and the IPSec Security Gateway application (ipsec-secgw).

While these applications demonstrate how an application can be created to perform generic crypto operation, the required complexity hides the basic steps of how to use the cryptodev APIs.

The following sample code shows the basic steps to encrypt several buffers with AES-CBC (although performing other crypto operations is similar), using one of the crypto PMDs available in DPDK.

/*
 * Simple example to encrypt several buffers with AES-CBC using
 * the Cryptodev APIs.
 */

#define MAX_SESSIONS         1024
#define NUM_MBUFS            1024
#define POOL_CACHE_SIZE      128
#define BURST_SIZE           32
#define BUFFER_SIZE          1024
#define AES_CBC_IV_LENGTH    16
#define AES_CBC_KEY_LENGTH   16
#define IV_OFFSET            (sizeof(struct rte_crypto_op) + \
                             sizeof(struct rte_crypto_sym_op))

struct rte_mempool *mbuf_pool, *crypto_op_pool, *session_pool;
unsigned int session_size;
int ret;

/* Initialize EAL. */
ret = rte_eal_init(argc, argv);
if (ret < 0)
    rte_exit(EXIT_FAILURE, "Invalid EAL arguments\n");

uint8_t socket_id = rte_socket_id();

/* Create the mbuf pool. */
mbuf_pool = rte_pktmbuf_pool_create("mbuf_pool",
                                NUM_MBUFS,
                                POOL_CACHE_SIZE,
                                0,
                                RTE_MBUF_DEFAULT_BUF_SIZE,
                                socket_id);
if (mbuf_pool == NULL)
    rte_exit(EXIT_FAILURE, "Cannot create mbuf pool\n");

/*
 * The IV is always placed after the crypto operation,
 * so some private data is required to be reserved.
 */
unsigned int crypto_op_private_data = AES_CBC_IV_LENGTH;

/* Create crypto operation pool. */
crypto_op_pool = rte_crypto_op_pool_create("crypto_op_pool",
                                        RTE_CRYPTO_OP_TYPE_SYMMETRIC,
                                        NUM_MBUFS,
                                        POOL_CACHE_SIZE,
                                        crypto_op_private_data,
                                        socket_id);
if (crypto_op_pool == NULL)
    rte_exit(EXIT_FAILURE, "Cannot create crypto op pool\n");

/* Create the virtual crypto device. */
char args[128];
const char *crypto_name = "crypto_aesni_mb0";
snprintf(args, sizeof(args), "socket_id=%d", socket_id);
ret = rte_vdev_init(crypto_name, args);
if (ret != 0)
    rte_exit(EXIT_FAILURE, "Cannot create virtual device");

uint8_t cdev_id = rte_cryptodev_get_dev_id(crypto_name);

/* Get private session data size. */
session_size = rte_cryptodev_get_private_session_size(cdev_id);

/*
 * Create session mempool, with two objects per session,
 * one for the session header and another one for the
 * private session data for the crypto device.
 */
session_pool = rte_mempool_create("session_pool",
                                MAX_SESSIONS * 2,
                                session_size,
                                POOL_CACHE_SIZE,
                                0, NULL, NULL, NULL,
                                NULL, socket_id,
                                0);

/* Configure the crypto device. */
struct rte_cryptodev_config conf = {
    .nb_queue_pairs = 1,
    .socket_id = socket_id
};
struct rte_cryptodev_qp_conf qp_conf = {
    .nb_descriptors = 2048
};

if (rte_cryptodev_configure(cdev_id, &conf) < 0)
    rte_exit(EXIT_FAILURE, "Failed to configure cryptodev %u", cdev_id);

if (rte_cryptodev_queue_pair_setup(cdev_id, 0, &qp_conf,
                        socket_id, session_pool) < 0)
    rte_exit(EXIT_FAILURE, "Failed to setup queue pair\n");

if (rte_cryptodev_start(cdev_id) < 0)
    rte_exit(EXIT_FAILURE, "Failed to start device\n");

/* Create the crypto transform. */
uint8_t cipher_key[16] = {0};
struct rte_crypto_sym_xform cipher_xform = {
    .next = NULL,
    .type = RTE_CRYPTO_SYM_XFORM_CIPHER,
    .cipher = {
        .op = RTE_CRYPTO_CIPHER_OP_ENCRYPT,
        .algo = RTE_CRYPTO_CIPHER_AES_CBC,
        .key = {
            .data = cipher_key,
            .length = AES_CBC_KEY_LENGTH
        },
        .iv = {
            .offset = IV_OFFSET,
            .length = AES_CBC_IV_LENGTH
        }
    }
};

/* Create crypto session and initialize it for the crypto device. */
struct rte_cryptodev_sym_session *session;
session = rte_cryptodev_sym_session_create(session_pool);
if (session == NULL)
    rte_exit(EXIT_FAILURE, "Session could not be created\n");

if (rte_cryptodev_sym_session_init(cdev_id, session,
                &cipher_xform, session_pool) < 0)
    rte_exit(EXIT_FAILURE, "Session could not be initialized "
                "for the crypto device\n");

/* Get a burst of crypto operations. */
struct rte_crypto_op *crypto_ops[BURST_SIZE];
if (rte_crypto_op_bulk_alloc(crypto_op_pool,
                        RTE_CRYPTO_OP_TYPE_SYMMETRIC,
                        crypto_ops, BURST_SIZE) == 0)
    rte_exit(EXIT_FAILURE, "Not enough crypto operations available\n");

/* Get a burst of mbufs. */
struct rte_mbuf *mbufs[BURST_SIZE];
if (rte_pktmbuf_alloc_bulk(mbuf_pool, mbufs, BURST_SIZE) < 0)
    rte_exit(EXIT_FAILURE, "Not enough mbufs available");

/* Initialize the mbufs and append them to the crypto operations. */
unsigned int i;
for (i = 0; i < BURST_SIZE; i++) {
    if (rte_pktmbuf_append(mbufs[i], BUFFER_SIZE) == NULL)
        rte_exit(EXIT_FAILURE, "Not enough room in the mbuf\n");
    crypto_ops[i]->sym->m_src = mbufs[i];
}

/* Set up the crypto operations. */
for (i = 0; i < BURST_SIZE; i++) {
    struct rte_crypto_op *op = crypto_ops[i];
    /* Modify bytes of the IV at the end of the crypto operation */
    uint8_t *iv_ptr = rte_crypto_op_ctod_offset(op, uint8_t *,
                                            IV_OFFSET);

    generate_random_bytes(iv_ptr, AES_CBC_IV_LENGTH);

    op->sym->cipher.data.offset = 0;
    op->sym->cipher.data.length = BUFFER_SIZE;

    /* Attach the crypto session to the operation */
    rte_crypto_op_attach_sym_session(op, session);
}

/* Enqueue the crypto operations in the crypto device. */
uint16_t num_enqueued_ops = rte_cryptodev_enqueue_burst(cdev_id, 0,
                                        crypto_ops, BURST_SIZE);

/*
 * Dequeue the crypto operations until all the operations
 * are proccessed in the crypto device.
 */
uint16_t num_dequeued_ops, total_num_dequeued_ops = 0;
do {
    struct rte_crypto_op *dequeued_ops[BURST_SIZE];
    num_dequeued_ops = rte_cryptodev_dequeue_burst(cdev_id, 0,
                                    dequeued_ops, BURST_SIZE);
    total_num_dequeued_ops += num_dequeued_ops;

    /* Check if operation was processed successfully */
    for (i = 0; i < num_dequeued_ops; i++) {
        if (dequeued_ops[i]->status != RTE_CRYPTO_OP_STATUS_SUCCESS)
            rte_exit(EXIT_FAILURE,
                    "Some operations were not processed correctly");
    }

    rte_mempool_put_bulk(crypto_op_pool, (void **)dequeued_ops,
                                        num_dequeued_ops);
} while (total_num_dequeued_ops < num_enqueued_ops);

Asymmetric Cryptography

Asymmetric functionality is currently not supported by the cryptodev API.

Crypto Device API

The cryptodev Library API is described in the DPDK API Reference document.

Security Library

The security library provides a framework for management and provisioning of security protocol operations offloaded to hardware based devices. The library defines generic APIs to create and free security sessions which can support full protocol offload as well as inline crypto operation with NIC or crypto devices. The framework currently only supports the IPSec protocol and associated operations, other protocols will be added in future.

Design Principles

The security library provides an additional offload capability to an existing crypto device and/or ethernet device.

       +---------------+
       | rte_security  |
       +---------------+
         \            /
+-----------+    +--------------+
|  NIC PMD  |    |  CRYPTO PMD  |
+-----------+    +--------------+

Note

Currently, the security library does not support the case of multi-process. It will be updated in the future releases.

The supported offload types are explained in the sections below.

Inline Crypto

RTE_SECURITY_ACTION_TYPE_INLINE_CRYPTO: The crypto processing for security protocol (e.g. IPSec) is processed inline during receive and transmission on NIC port. The flow based security action should be configured on the port.

Ingress Data path - The packet is decrypted in RX path and relevant crypto status is set in Rx descriptors. After the successful inline crypto processing the packet is presented to host as a regular Rx packet however all security protocol related headers are still attached to the packet. e.g. In case of IPSec, the IPSec tunnel headers (if any), ESP/AH headers will remain in the packet but the received packet contains the decrypted data where the encrypted data was when the packet arrived. The driver Rx path check the descriptors and and based on the crypto status sets additional flags in the rte_mbuf.ol_flags field.

Note

The underlying device may not support crypto processing for all ingress packet matching to a particular flow (e.g. fragmented packets), such packets will be passed as encrypted packets. It is the responsibility of application to process such encrypted packets using other crypto driver instance.

Egress Data path - The software prepares the egress packet by adding relevant security protocol headers. Only the data will not be encrypted by the software. The driver will accordingly configure the tx descriptors. The hardware device will encrypt the data before sending the the packet out.

Note

The underlying device may support post encryption TSO.

  Egress Data Path
         |
+--------|--------+
|  egress IPsec   |
|        |        |
| +------V------+ |
| | SADB lookup | |
| +------|------+ |
| +------V------+ |
| |   Tunnel    | |   <------ Add tunnel header to packet
| +------|------+ |
| +------V------+ |
| |     ESP     | |   <------ Add ESP header without trailer to packet
| |             | |   <------ Mark packet to be offloaded, add trailer
| +------|------+ |            meta-data to mbuf
+--------V--------+
         |
+--------V--------+
|    L2 Stack     |
+--------|--------+
         |
+--------V--------+
|                 |
|     NIC PMD     |   <------ Set hw context for inline crypto offload
|                 |
+--------|--------+
         |
+--------|--------+
|  HW ACCELERATED |   <------ Packet Encryption and
|        NIC      |           Authentication happens inline
|                 |
+-----------------+
Inline protocol offload

RTE_SECURITY_ACTION_TYPE_INLINE_PROTOCOL: The crypto and protocol processing for security protocol (e.g. IPSec) is processed inline during receive and transmission. The flow based security action should be configured on the port.

Ingress Data path - The packet is decrypted in the RX path and relevant crypto status is set in the Rx descriptors. After the successful inline crypto processing the packet is presented to the host as a regular Rx packet but all security protocol related headers are optionally removed from the packet. e.g. in the case of IPSec, the IPSec tunnel headers (if any), ESP/AH headers will be removed from the packet and the received packet will contains the decrypted packet only. The driver Rx path checks the descriptors and based on the crypto status sets additional flags in rte_mbuf.ol_flags field.

Note

The underlying device in this case is stateful. It is expected that the device shall support crypto processing for all kind of packets matching to a given flow, this includes fragmented packets (post reassembly). E.g. in case of IPSec the device may internally manage anti-replay etc. It will provide a configuration option for anti-replay behavior i.e. to drop the packets or pass them to driver with error flags set in the descriptor.

Egress Data path - The software will send the plain packet without any security protocol headers added to the packet. The driver will configure the security index and other requirement in tx descriptors. The hardware device will do security processing on the packet that includes adding the relevant protocol headers and encrypting the data before sending the packet out. The software should make sure that the buffer has required head room and tail room for any protocol header addition. The software may also do early fragmentation if the resultant packet is expected to cross the MTU size.

Note

The underlying device will manage state information required for egress processing. E.g. in case of IPSec, the seq number will be added to the packet, however the device shall provide indication when the sequence number is about to overflow. The underlying device may support post encryption TSO.

 Egress Data Path
         |
+--------|--------+
|  egress IPsec   |
|        |        |
| +------V------+ |
| | SADB lookup | |
| +------|------+ |
| +------V------+ |
| |   Desc      | |   <------ Mark packet to be offloaded
| +------|------+ |
+--------V--------+
         |
+--------V--------+
|    L2 Stack     |
+--------|--------+
         |
+--------V--------+
|                 |
|     NIC PMD     |   <------ Set hw context for inline crypto offload
|                 |
+--------|--------+
         |
+--------|--------+
|  HW ACCELERATED |   <------ Add tunnel, ESP header etc header to
|        NIC      |           packet. Packet Encryption and
|                 |           Authentication happens inline.
+-----------------+
Lookaside protocol offload

RTE_SECURITY_ACTION_TYPE_LOOKASIDE_PROTOCOL: This extends librte_cryptodev to support the programming of IPsec Security Association (SA) as part of a crypto session creation including the definition. In addition to standard crypto processing, as defined by the cryptodev, the security protocol processing is also offloaded to the crypto device.

Decryption: The packet is sent to the crypto device for security protocol processing. The device will decrypt the packet and it will also optionally remove additional security headers from the packet. E.g. in case of IPSec, IPSec tunnel headers (if any), ESP/AH headers will be removed from the packet and the decrypted packet may contain plain data only.

Note

In case of IPSec the device may internally manage anti-replay etc. It will provide a configuration option for anti-replay behavior i.e. to drop the packets or pass them to driver with error flags set in descriptor.

Encryption: The software will submit the packet to cryptodev as usual for encryption, the hardware device in this case will also add the relevant security protocol header along with encrypting the packet. The software should make sure that the buffer has required head room and tail room for any protocol header addition.

Note

In the case of IPSec, the seq number will be added to the packet, It shall provide an indication when the sequence number is about to overflow.

  Egress Data Path
         |
+--------|--------+
|  egress IPsec   |
|        |        |
| +------V------+ |
| | SADB lookup | |   <------ SA maps to cryptodev session
| +------|------+ |
| +------|------+ |
| |      \--------------------\
| |    Crypto   | |           |  <- Crypto processing through
| |      /----------------\   |     inline crypto PMD
| +------|------+ |       |   |
+--------V--------+       |   |
         |                |   |
+--------V--------+       |   |  create   <-- SA is added to hw
|    L2 Stack     |       |   |  inline       using existing create
+--------|--------+       |   |  session      sym session APIs
         |                |   |    |
+--------V--------+   +---|---|----V---+
|                 |   |   \---/    |   | <--- Add tunnel, ESP header etc
|     NIC PMD     |   |   INLINE   |   |      header to packet.Packet
|                 |   | CRYPTO PMD |   |      Encryption/Decryption and
+--------|--------+   +----------------+      Authentication happens
         |                                    inline.
+--------|--------+
|       NIC       |
+--------|--------+
         V

Device Features and Capabilities

Device Capabilities For Security Operations

The device (crypto or ethernet) capabilities which support security operations, are defined by the security action type, security protocol, protocol capabilities and corresponding crypto capabilities for security. For the full scope of the Security capability see definition of rte_security_capability structure in the DPDK API Reference.

struct rte_security_capability;

Each driver (crypto or ethernet) defines its own private array of capabilities for the operations it supports. Below is an example of the capabilities for a PMD which supports the IPSec protocol.

static const struct rte_security_capability pmd_security_capabilities[] = {
    { /* IPsec Lookaside Protocol offload ESP Tunnel Egress */
            .action = RTE_SECURITY_ACTION_TYPE_LOOKASIDE_PROTOCOL,
            .protocol = RTE_SECURITY_PROTOCOL_IPSEC,
            .ipsec = {
                    .proto = RTE_SECURITY_IPSEC_SA_PROTO_ESP,
                    .mode = RTE_SECURITY_IPSEC_SA_MODE_TUNNEL,
                    .direction = RTE_SECURITY_IPSEC_SA_DIR_EGRESS,
                    .options = { 0 }
            },
            .crypto_capabilities = pmd_capabilities
    },
    { /* IPsec Lookaside Protocol offload ESP Tunnel Ingress */
            .action = RTE_SECURITY_ACTION_TYPE_LOOKASIDE_PROTOCOL,
            .protocol = RTE_SECURITY_PROTOCOL_IPSEC,
            .ipsec = {
                    .proto = RTE_SECURITY_IPSEC_SA_PROTO_ESP,
                    .mode = RTE_SECURITY_IPSEC_SA_MODE_TUNNEL,
                    .direction = RTE_SECURITY_IPSEC_SA_DIR_INGRESS,
                    .options = { 0 }
            },
            .crypto_capabilities = pmd_capabilities
    },
    {
            .action = RTE_SECURITY_ACTION_TYPE_NONE
    }
};
static const struct rte_cryptodev_capabilities pmd_capabilities[] = {
    {    /* SHA1 HMAC */
        .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC,
        .sym = {
            .xform_type = RTE_CRYPTO_SYM_XFORM_AUTH,
            .auth = {
                .algo = RTE_CRYPTO_AUTH_SHA1_HMAC,
                .block_size = 64,
                .key_size = {
                    .min = 64,
                    .max = 64,
                    .increment = 0
                },
                .digest_size = {
                    .min = 12,
                    .max = 12,
                    .increment = 0
                },
                .aad_size = { 0 },
                .iv_size = { 0 }
            }
        }
    },
    {    /* AES CBC */
        .op = RTE_CRYPTO_OP_TYPE_SYMMETRIC,
        .sym = {
            .xform_type = RTE_CRYPTO_SYM_XFORM_CIPHER,
            .cipher = {
                .algo = RTE_CRYPTO_CIPHER_AES_CBC,
                .block_size = 16,
                .key_size = {
                    .min = 16,
                    .max = 32,
                    .increment = 8
                },
                .iv_size = {
                    .min = 16,
                    .max = 16,
                    .increment = 0
                }
            }
        }
    }
}
Capabilities Discovery

Discovering the features and capabilities of a driver (crypto/ethernet) is achieved through the rte_security_capabilities_get() function.

const struct rte_security_capability *rte_security_capabilities_get(uint16_t id);

This allows the user to query a specific driver and get all device security capabilities. It returns an array of rte_security_capability structures which contains all the capabilities for that device.

Security Session Create/Free

Security Sessions are created to store the immutable fields of a particular Security Association for a particular protocol which is defined by a security session configuration structure which is used in the operation processing of a packet flow. Sessions are used to manage protocol specific information as well as crypto parameters. Security sessions cache this immutable data in a optimal way for the underlying PMD and this allows further acceleration of the offload of Crypto workloads.

The Security framework provides APIs to create and free sessions for crypto/ethernet devices, where sessions are mempool objects. It is the application’s responsibility to create and manage the session mempools. The mempool object size should be able to accommodate the driver’s private data of security session.

Once the session mempools have been created, rte_security_session_create() is used to allocate and initialize a session for the required crypto/ethernet device.

Session APIs need a parameter rte_security_ctx to identify the crypto/ethernet security ops. This parameter can be retrieved using the APIs rte_cryptodev_get_sec_ctx() (for crypto device) or rte_eth_dev_get_sec_ctx (for ethernet port).

Sessions already created can be updated with rte_security_session_update().

When a session is no longer used, the user must call rte_security_session_destroy() to free the driver private session data and return the memory back to the mempool.

For look aside protocol offload to hardware crypto device, the rte_crypto_op created by the application is attached to the security session by the API rte_security_attach_session().

For Inline Crypto and Inline protocol offload, device specific defined metadata is updated in the mbuf using rte_security_set_pkt_metadata() if DEV_TX_OFFLOAD_SEC_NEED_MDATA is set.

Security session configuration

Security Session configuration structure is defined as rte_security_session_conf

struct rte_security_session_conf {
    enum rte_security_session_action_type action_type;
    /**< Type of action to be performed on the session */
    enum rte_security_session_protocol protocol;
    /**< Security protocol to be configured */
    union {
            struct rte_security_ipsec_xform ipsec;
            struct rte_security_macsec_xform macsec;
    };
    /**< Configuration parameters for security session */
    struct rte_crypto_sym_xform *crypto_xform;
    /**< Security Session Crypto Transformations */
};

The configuration structure reuses the rte_crypto_sym_xform struct for crypto related configuration. The rte_security_session_action_type struct is used to specify whether the session is configured for Lookaside Protocol offload or Inline Crypto or Inline Protocol Offload.

enum rte_security_session_action_type {
    RTE_SECURITY_ACTION_TYPE_NONE,
    /**< No security actions */
    RTE_SECURITY_ACTION_TYPE_INLINE_CRYPTO,
    /**< Crypto processing for security protocol is processed inline
     * during transmission */
    RTE_SECURITY_ACTION_TYPE_INLINE_PROTOCOL,
    /**< All security protocol processing is performed inline during
     * transmission */
    RTE_SECURITY_ACTION_TYPE_LOOKASIDE_PROTOCOL
    /**< All security protocol processing including crypto is performed
     * on a lookaside accelerator */
};

The rte_security_session_protocol is defined as

enum rte_security_session_protocol {
    RTE_SECURITY_PROTOCOL_IPSEC,
    /**< IPsec Protocol */
    RTE_SECURITY_PROTOCOL_MACSEC,
    /**< MACSec Protocol */
};

Currently the library defines configuration parameters for IPSec only. For other protocols like MACSec, structures and enums are defined as place holders which will be updated in the future.

IPsec related configuration parameters are defined in rte_security_ipsec_xform

struct rte_security_ipsec_xform {
    uint32_t spi;
    /**< SA security parameter index */
    uint32_t salt;
    /**< SA salt */
    struct rte_security_ipsec_sa_options options;
    /**< various SA options */
    enum rte_security_ipsec_sa_direction direction;
    /**< IPSec SA Direction - Egress/Ingress */
    enum rte_security_ipsec_sa_protocol proto;
    /**< IPsec SA Protocol - AH/ESP */
    enum rte_security_ipsec_sa_mode mode;
    /**< IPsec SA Mode - transport/tunnel */
    struct rte_security_ipsec_tunnel_param tunnel;
    /**< Tunnel parameters, NULL for transport mode */
};
Security API

The rte_security Library API is described in the DPDK API Reference document.

Flow based Security Session

In the case of NIC based offloads, the security session specified in the ‘rte_flow_action_security’ must be created on the same port as the flow action that is being specified.

The ingress/egress flow attribute should match that specified in the security session if the security session supports the definition of the direction.

Multiple flows can be configured to use the same security session. For example if the security session specifies an egress IPsec SA, then multiple flows can be specified to that SA. In the case of an ingress IPsec SA then it is only valid to have a single flow to map to that security session.

 Configuration Path
         |
+--------|--------+
|    Add/Remove   |
|     IPsec SA    |   <------ Build security flow action of
|        |        |           ipsec transform
|--------|--------|
         |
+--------V--------+
|   Flow API      |
+--------|--------+
         |
+--------V--------+
|                 |
|     NIC PMD     |   <------ Add/Remove SA to/from hw context
|                 |
+--------|--------+
         |
+--------|--------+
|  HW ACCELERATED |
|        NIC      |
|                 |
+--------|--------+
  • Add/Delete SA flow: To add a new inline SA construct a rte_flow_item for Ethernet + IP + ESP using the SA selectors and the rte_crypto_ipsec_xform as the rte_flow_action. Note that any rte_flow_items may be empty, which means it is not checked.
In its most basic form, IPsec flow specification is as follows:
    +-------+     +----------+    +--------+    +-----+
    |  Eth  | ->  |   IP4/6  | -> |   ESP  | -> | END |
    +-------+     +----------+    +--------+    +-----+

However, the API can represent, IPsec crypto offload with any encapsulation:
    +-------+            +--------+    +-----+
    |  Eth  | ->  ... -> |   ESP  | -> | END |
    +-------+            +--------+    +-----+

Timer Library

The Timer library provides a timer service to DPDK execution units to enable execution of callback functions asynchronously. Features of the library are:

  • Timers can be periodic (multi-shot) or single (one-shot).
  • Timers can be loaded from one core and executed on another. It has to be specified in the call to rte_timer_reset().
  • Timers provide high precision (depends on the call frequency to rte_timer_manage() that checks timer expiration for the local core).
  • If not required in the application, timers can be disabled at compilation time by not calling the rte_timer_manage() to increase performance.

The timer library uses the rte_get_timer_cycles() function that uses the High Precision Event Timer (HPET) or the CPUs Time Stamp Counter (TSC) to provide a reliable time reference.

This library provides an interface to add, delete and restart a timer. The API is based on BSD callout() with a few differences. Refer to the callout manual.

Implementation Details

Timers are tracked on a per-lcore basis, with all pending timers for a core being maintained in order of timer expiry in a skiplist data structure. The skiplist used has ten levels and each entry in the table appears in each level with probability ¼^level. This means that all entries are present in level 0, 1 in every 4 entries is present at level 1, one in every 16 at level 2 and so on up to level 9. This means that adding and removing entries from the timer list for a core can be done in log(n) time, up to 4^10 entries, that is, approximately 1,000,000 timers per lcore.

A timer structure contains a special field called status, which is a union of a timer state (stopped, pending, running, config) and an owner (lcore id). Depending on the timer state, we know if a timer is present in a list or not:

  • STOPPED: no owner, not in a list
  • CONFIG: owned by a core, must not be modified by another core, maybe in a list or not, depending on previous state
  • PENDING: owned by a core, present in a list
  • RUNNING: owned by a core, must not be modified by another core, present in a list

Resetting or stopping a timer while it is in a CONFIG or RUNNING state is not allowed. When modifying the state of a timer, a Compare And Swap instruction should be used to guarantee that the status (state+owner) is modified atomically.

Inside the rte_timer_manage() function, the skiplist is used as a regular list by iterating along the level 0 list, which contains all timer entries, until an entry which has not yet expired has been encountered. To improve performance in the case where there are entries in the timer list but none of those timers have yet expired, the expiry time of the first list entry is maintained within the per-core timer list structure itself. On 64-bit platforms, this value can be checked without the need to take a lock on the overall structure. (Since expiry times are maintained as 64-bit values, a check on the value cannot be done on 32-bit platforms without using either a compare-and-swap (CAS) instruction or using a lock, so this additional check is skipped in favor of checking as normal once the lock has been taken.) On both 64-bit and 32-bit platforms, a call to rte_timer_manage() returns without taking a lock in the case where the timer list for the calling core is empty.

Use Cases

The timer library is used for periodic calls, such as garbage collectors, or some state machines (ARP, bridging, and so on).

References

  • callout manual - The callout facility that provides timers with a mechanism to execute a function at a given time.
  • HPET - Information about the High Precision Event Timer (HPET).

Hash Library

The DPDK provides a Hash Library for creating hash table for fast lookup. The hash table is a data structure optimized for searching through a set of entries that are each identified by a unique key. For increased performance the DPDK Hash requires that all the keys have the same number of bytes which is set at the hash creation time.

Hash API Overview

The main configuration parameters for the hash are:

  • Total number of hash entries
  • Size of the key in bytes

The hash also allows the configuration of some low-level implementation related parameters such as:

  • Hash function to translate the key into a bucket index

The main methods exported by the hash are:

  • Add entry with key: The key is provided as input. If a new entry is successfully added to the hash for the specified key, or there is already an entry in the hash for the specified key, then the position of the entry is returned. If the operation was not successful, for example due to lack of free entries in the hash, then a negative value is returned;
  • Delete entry with key: The key is provided as input. If an entry with the specified key is found in the hash, then the entry is removed from the hash and the position where the entry was found in the hash is returned. If no entry with the specified key exists in the hash, then a negative value is returned
  • Lookup for entry with key: The key is provided as input. If an entry with the specified key is found in the hash (lookup hit), then the position of the entry is returned, otherwise (lookup miss) a negative value is returned.

Apart from these method explained above, the API allows the user three more options:

  • Add / lookup / delete with key and precomputed hash: Both the key and its precomputed hash are provided as input. This allows the user to perform these operations faster, as hash is already computed.
  • Add / lookup with key and data: A pair of key-value is provided as input. This allows the user to store not only the key, but also data which may be either a 8-byte integer or a pointer to external data (if data size is more than 8 bytes).
  • Combination of the two options above: User can provide key, precomputed hash and data.

Also, the API contains a method to allow the user to look up entries in bursts, achieving higher performance than looking up individual entries, as the function prefetches next entries at the time it is operating with the first ones, which reduces significantly the impact of the necessary memory accesses. Notice that this method uses a pipeline of 8 entries (4 stages of 2 entries), so it is highly recommended to use at least 8 entries per burst.

The actual data associated with each key can be either managed by the user using a separate table that mirrors the hash in terms of number of entries and position of each entry, as shown in the Flow Classification use case describes in the following sections, or stored in the hash table itself.

The example hash tables in the L2/L3 Forwarding sample applications defines which port to forward a packet to based on a packet flow identified by the five-tuple lookup. However, this table could also be used for more sophisticated features and provide many other functions and actions that could be performed on the packets and flows.

Multi-process support

The hash library can be used in a multi-process environment, minding that only lookups are thread-safe. The only function that can only be used in single-process mode is rte_hash_set_cmp_func(), which sets up a custom compare function, which is assigned to a function pointer (therefore, it is not supported in multi-process mode).

Implementation Details

The hash table has two main tables:

  • First table is an array of entries which is further divided into buckets, with the same number of consecutive array entries in each bucket. Each entry contains the computed primary and secondary hashes of a given key (explained below), and an index to the second table.
  • The second table is an array of all the keys stored in the hash table and its data associated to each key.

The hash library uses the cuckoo hash method to resolve collisions. For any input key, there are two possible buckets (primary and secondary/alternative location) where that key can be stored in the hash, therefore only the entries within those bucket need to be examined when the key is looked up. The lookup speed is achieved by reducing the number of entries to be scanned from the total number of hash entries down to the number of entries in the two hash buckets, as opposed to the basic method of linearly scanning all the entries in the array. The hash uses a hash function (configurable) to translate the input key into a 4-byte key signature. The bucket index is the key signature modulo the number of hash buckets.

Once the buckets are identified, the scope of the hash add, delete and lookup operations is reduced to the entries in those buckets (it is very likely that entries are in the primary bucket).

To speed up the search logic within the bucket, each hash entry stores the 4-byte key signature together with the full key for each hash entry. For large key sizes, comparing the input key against a key from the bucket can take significantly more time than comparing the 4-byte signature of the input key against the signature of a key from the bucket. Therefore, the signature comparison is done first and the full key comparison done only when the signatures matches. The full key comparison is still necessary, as two input keys from the same bucket can still potentially have the same 4-byte hash signature, although this event is relatively rare for hash functions providing good uniform distributions for the set of input keys.

Example of lookup:

First of all, the primary bucket is identified and entry is likely to be stored there. If signature was stored there, we compare its key against the one provided and return the position where it was stored and/or the data associated to that key if there is a match. If signature is not in the primary bucket, the secondary bucket is looked up, where same procedure is carried out. If there is no match there either, key is considered not to be in the table.

Example of addition:

Like lookup, the primary and secondary buckets are identified. If there is an empty slot in the primary bucket, primary and secondary signatures are stored in that slot, key and data (if any) are added to the second table and an index to the position in the second table is stored in the slot of the first table. If there is no space in the primary bucket, one of the entries on that bucket is pushed to its alternative location, and the key to be added is inserted in its position. To know where the alternative bucket of the evicted entry is, the secondary signature is looked up and alternative bucket index is calculated from doing the modulo, as seen above. If there is room in the alternative bucket, the evicted entry is stored in it. If not, same process is repeated (one of the entries gets pushed) until a non full bucket is found. Notice that despite all the entry movement in the first table, the second table is not touched, which would impact greatly in performance.

In the very unlikely event that table enters in a loop where same entries are being evicted indefinitely, key is considered not able to be stored. With random keys, this method allows the user to get around 90% of the table utilization, without having to drop any stored entry (LRU) or allocate more memory (extended buckets).

Entry distribution in hash table

As mentioned above, Cuckoo hash implementation pushes elements out of their bucket, if there is a new entry to be added which primary location coincides with their current bucket, being pushed to their alternative location. Therefore, as user adds more entries to the hash table, distribution of the hash values in the buckets will change, being most of them in their primary location and a few in their secondary location, which the later will increase, as table gets busier. This information is quite useful, as performance may be lower as more entries are evicted to their secondary location.

See the tables below showing example entry distribution as table utilization increases.

Entry distribution measured with an example table with 1024 random entries using jhash algorithm
% Table used % In Primary location % In Secondary location
25 100 0
50 96.1 3.9
75 88.2 11.8
80 86.3 13.7
85 83.1 16.9
90 77.3 22.7
95.8 64.5 35.5

Entry distribution measured with an example table with 1 million random entries using jhash algorithm
% Table used % In Primary location % In Secondary location
50 96 4
75 86.9 13.1
80 83.9 16.1
85 80.1 19.9
90 74.8 25.2
94.5 67.4 32.6

Note

Last values on the tables above are the average maximum table utilization with random keys and using Jenkins hash function.

Use Case: Flow Classification

Flow classification is used to map each input packet to the connection/flow it belongs to. This operation is necessary as the processing of each input packet is usually done in the context of their connection, so the same set of operations is applied to all the packets from the same flow.

Applications using flow classification typically have a flow table to manage, with each separate flow having an entry associated with it in this table. The size of the flow table entry is application specific, with typical values of 4, 16, 32 or 64 bytes.

Each application using flow classification typically has a mechanism defined to uniquely identify a flow based on a number of fields read from the input packet that make up the flow key. One example is to use the DiffServ 5-tuple made up of the following fields of the IP and transport layer packet headers: Source IP Address, Destination IP Address, Protocol, Source Port, Destination Port.

The DPDK hash provides a generic method to implement an application specific flow classification mechanism. Given a flow table implemented as an array, the application should create a hash object with the same number of entries as the flow table and with the hash key size set to the number of bytes in the selected flow key.

The flow table operations on the application side are described below:

  • Add flow: Add the flow key to hash. If the returned position is valid, use it to access the flow entry in the flow table for adding a new flow or updating the information associated with an existing flow. Otherwise, the flow addition failed, for example due to lack of free entries for storing new flows.
  • Delete flow: Delete the flow key from the hash. If the returned position is valid, use it to access the flow entry in the flow table to invalidate the information associated with the flow.
  • Lookup flow: Lookup for the flow key in the hash. If the returned position is valid (flow lookup hit), use the returned position to access the flow entry in the flow table. Otherwise (flow lookup miss) there is no flow registered for the current packet.

References

  • Donald E. Knuth, The Art of Computer Programming, Volume 3: Sorting and Searching (2nd Edition), 1998, Addison-Wesley Professional

Elastic Flow Distributor Library

Introduction

In Data Centers today, clustering and scheduling of distributed workloads is a very common task. Many workloads require a deterministic partitioning of a flat key space among a cluster of machines. When a packet enters the cluster, the ingress node will direct the packet to its handling node. For example, data-centers with disaggregated storage use storage metadata tables to forward I/O requests to the correct back end storage cluster, stateful packet inspection will use match incoming flows to signatures in flow tables to send incoming packets to their intended deep packet inspection (DPI) devices, and so on.

EFD is a distributor library that uses perfect hashing to determine a target/value for a given incoming flow key. It has the following advantages: first, because it uses perfect hashing it does not store the key itself and hence lookup performance is not dependent on the key size. Second, the target/value can be any arbitrary value hence the system designer and/or operator can better optimize service rates and inter-cluster network traffic locating. Third, since the storage requirement is much smaller than a hash-based flow table (i.e. better fit for CPU cache), EFD can scale to millions of flow keys. Finally, with the current optimized library implementation, performance is fully scalable with any number of CPU cores.

Flow Based Distribution

Computation Based Schemes

Flow distribution and/or load balancing can be simply done using a stateless computation, for instance using round-robin or a simple computation based on the flow key as an input. For example, a hash function can be used to direct a certain flow to a target based on the flow key (e.g. h(key) mod n) where h(key) is the hash value of the flow key and n is the number of possible targets.

_images/efd_i1.svg

Load Balancing Using Front End Node

In this scheme (Fig. 17.3), the front end server/distributor/load balancer extracts the flow key from the input packet and applies a computation to determine where this flow should be directed. Intuitively, this scheme is very simple and requires no state to be kept at the front end node, and hence, storage requirements are minimum.

_images/efd_i2.svg

Consistent Hashing

A widely used flow distributor that belongs to the same category of computation-based schemes is consistent hashing, shown in Fig. 17.4. Target destinations (shown in red) are hashed into the same space as the flow keys (shown in blue), and keys are mapped to the nearest target in a clockwise fashion. Dynamically adding and removing targets with consistent hashing requires only K/n keys to be remapped on average, where K is the number of keys, and n is the number of targets. In contrast, in a traditional hash-based scheme, a change in the number of targets causes nearly all keys to be remapped.

Although computation-based schemes are simple and need very little storage requirement, they suffer from the drawback that the system designer/operator can’t fully control the target to assign a specific key, as this is dictated by the hash function. Deterministically co-locating of keys together (for example, to minimize inter-server traffic or to optimize for network traffic conditions, target load, etc.) is simply not possible.

Flow-Table Based Schemes

When using a Flow-Table based scheme to handle flow distribution/load balancing, in contrast with computation-based schemes, the system designer has the flexibility of assigning a given flow to any given target. The flow table (e.g. DPDK RTE Hash Library) will simply store both the flow key and the target value.

_images/efd_i3.svg

Table Based Flow Distribution

As shown in Fig. 17.5, when doing a lookup, the flow-table is indexed with the hash of the flow key and the keys (more than one is possible, because of hash collision) stored in this index and corresponding values are retrieved. The retrieved key(s) is matched with the input flow key and if there is a match the value (target id) is returned.

The drawback of using a hash table for flow distribution/load balancing is the storage requirement, since the flow table need to store keys, signatures and target values. This doesn’t allow this scheme to scale to millions of flow keys. Large tables will usually not fit in the CPU cache, and hence, the lookup performance is degraded because of the latency to access the main memory.

EFD Based Scheme

EFD combines the advantages of both flow-table based and computation-based schemes. It doesn’t require the large storage necessary for flow-table based schemes (because EFD doesn’t store the key as explained below), and it supports any arbitrary value for any given key.

_images/efd_i4.svg

Searching for Perfect Hash Function

The basic idea of EFD is when a given key is to be inserted, a family of hash functions is searched until the correct hash function that maps the input key to the correct value is found, as shown in Fig. 17.6. However, rather than explicitly storing all keys and their associated values, EFD stores only indices of hash functions that map keys to values, and thereby consumes much less space than conventional flow-based tables. The lookup operation is very simple, similar to a computational-based scheme: given an input key the lookup operation is reduced to hashing that key with the correct hash function.

_images/efd_i5.svg

Divide and Conquer for Millions of Keys

Intuitively, finding a hash function that maps each of a large number (millions) of input keys to the correct output value is effectively impossible, as a result EFD, as shown in Fig. 17.7, breaks the problem into smaller pieces (divide and conquer). EFD divides the entire input key set into many small groups. Each group consists of approximately 20-28 keys (a configurable parameter for the library), then, for each small group, a brute force search to find a hash function that produces the correct outputs for each key in the group.

It should be mentioned that, since the online lookup table for EFD doesn’t store the key itself, the size of the EFD table is independent of the key size and hence EFD lookup performance which is almost constant irrespective of the length of the key which is a highly desirable feature especially for longer keys.

In summary, EFD is a set separation data structure that supports millions of keys. It is used to distribute a given key to an intended target. By itself EFD is not a FIB data structure with an exact match the input flow key.

Example of EFD Library Usage

EFD can be used along the data path of many network functions and middleboxes. As previously mentioned, it can used as an index table for <key,value> pairs, meta-data for objects, a flow-level load balancer, etc. Fig. 17.8 shows an example of using EFD as a flow-level load balancer, where flows are received at a front end server before being forwarded to the target back end server for processing. The system designer would deterministically co-locate flows together in order to minimize cross-server interaction. (For example, flows requesting certain webpage objects are co-located together, to minimize forwarding of common objects across servers).

_images/efd_i6.svg

EFD as a Flow-Level Load Balancer

As shown in Fig. 17.8, the front end server will have an EFD table that stores for each group what is the perfect hash index that satisfies the correct output. Because the table size is small and fits in cache (since keys are not stored), it sustains a large number of flows (N*X, where N is the maximum number of flows served by each back end server of the X possible targets).

With an input flow key, the group id is computed (for example, using last few bits of CRC hash) and then the EFD table is indexed with the group id to retrieve the corresponding hash index to use. Once the index is retrieved the key is hashed using this hash function and the result will be the intended correct target where this flow is supposed to be processed.

It should be noted that as a result of EFD not matching the exact key but rather distributing the flows to a target back end node based on the perfect hash index, a key that has not been inserted before will be distributed to a valid target. Hence, a local table which stores the flows served at each node is used and is exact matched with the input key to rule out new never seen before flows.

Library API Overview

The EFD library API is created with a very similar semantics of a hash-index or a flow table. The application creates an EFD table for a given maximum number of flows, a function is called to insert a flow key with a specific target value, and another function is used to retrieve target values for a given individual flow key or a bulk of keys.

EFD Table Create

The function rte_efd_create() is used to create and return a pointer to an EFD table that is sized to hold up to num_flows key. The online version of the EFD table (the one that does not store the keys and is used for lookups) will be allocated and created in the last level cache (LLC) of the socket defined by the online_socket_bitmask, while the offline EFD table (the one that stores the keys and is used for key inserts and for computing the perfect hashing) is allocated and created in the LLC of the socket defined by offline_socket_bitmask. It should be noted, that for highest performance the socket id should match that where the thread is running, i.e. the online EFD lookup table should be created on the same socket as where the lookup thread is running.

EFD Insert and Update

The EFD function to insert a key or update a key to a new value is rte_efd_update(). This function will update an existing key to a new value (target) if the key has already been inserted before, or will insert the <key,value> pair if this key has not been inserted before. It will return 0 upon success. It will return EFD_UPDATE_WARN_GROUP_FULL (1) if the operation is insert, and the last available space in the key’s group was just used. It will return EFD_UPDATE_FAILED (2) when the insertion or update has failed (either it failed to find a suitable perfect hash or the group was full). The function will return EFD_UPDATE_NO_CHANGE (3) if there is no change to the EFD table (i.e, same value already exists).

Note

This function is not multi-thread safe and should only be called from one thread.

EFD Lookup

To lookup a certain key in an EFD table, the function rte_efd_lookup() is used to return the value associated with single key. As previously mentioned, if the key has been inserted, the correct value inserted is returned, if the key has not been inserted before, a ‘random’ value (based on hashing of the key) is returned. For better performance and to decrease the overhead of function calls per key, it is always recommended to use a bulk lookup function (simultaneous lookup of multiple keys) instead of a single key lookup function. rte_efd_lookup_bulk() is the bulk lookup function, that looks up num_keys simultaneously stored in the key_list and the corresponding return values will be returned in the value_list.

Note

This function is multi-thread safe, but there should not be other threads writing in the EFD table, unless locks are used.

EFD Delete

To delete a certain key in an EFD table, the function rte_efd_delete() can be used. The function returns zero upon success when the key has been found and deleted. Socket_id is the parameter to use to lookup the existing value, which is ideally the caller’s socket id. The previous value associated with this key will be returned in the prev_value argument.

Note

This function is not multi-thread safe and should only be called from one thread.

Library Internals

This section provides the brief high-level idea and an overview of the library internals to accompany the RFC. The intent of this section is to explain to readers the high-level implementation of insert, lookup and group rebalancing in the EFD library.

Insert Function Internals

As previously mentioned the EFD divides the whole set of keys into groups of a manageable size (e.g. 28 keys) and then searches for the perfect hash that satisfies the intended target value for each key. EFD stores two version of the <key,value> table:

  • Offline Version (in memory): Only used for the insertion/update operation, which is less frequent than the lookup operation. In the offline version the exact keys for each group is stored. When a new key is added, the hash function is updated that will satisfy the value for the new key together with the all old keys already inserted in this group.
  • Online Version (in cache): Used for the frequent lookup operation. In the online version, as previously mentioned, the keys are not stored but rather only the hash index for each group.
_images/efd_i7.svg

Group Assignment

Fig. 17.9 depicts the group assignment for 7 flow keys as an example. Given a flow key, a hash function (in our implementation CRC hash) is used to get the group id. As shown in the figure, the groups can be unbalanced. (We highlight group rebalancing further below).

_images/efd_i8.svg

Perfect Hash Search - Assigned Keys & Target Value

Focusing on one group that has four keys, Fig. 17.10 depicts the search algorithm to find the perfect hash function. Assuming that the target value bit for the keys is as shown in the figure, then the online EFD table will store a 16 bit hash index and 16 bit lookup table per group per value bit.

_images/efd_i9.svg

Perfect Hash Search - Satisfy Target Values

For a given keyX, a hash function (h(keyX, seed1) + index * h(keyX, seed2)) is used to point to certain bit index in the 16bit lookup_table value, as shown in Fig. 17.11. The insert function will brute force search for all possible values for the hash index until a non conflicting lookup_table is found.

_images/efd_i10.svg

Finding Hash Index for Conflict Free lookup_table

For example, since both key3 and key7 have a target bit value of 1, it is okay if the hash function of both keys point to the same bit in the lookup table. A conflict will occur if a hash index is used that maps both Key4 and Key7 to the same index in the lookup_table, as shown in Fig. 17.12, since their target value bit are not the same. Once a hash index is found that produces a lookup_table with no contradictions, this index is stored for this group. This procedure is repeated for each bit of target value.

Lookup Function Internals

The design principle of EFD is that lookups are much more frequent than inserts, and hence, EFD’s design optimizes for the lookups which are faster and much simpler than the slower insert procedure (inserts are slow, because of perfect hash search as previously discussed).

_images/efd_i11.svg

EFD Lookup Operation

Fig. 17.13 depicts the lookup operation for EFD. Given an input key, the group id is computed (using CRC hash) and then the hash index for this group is retrieved from the EFD table. Using the retrieved hash index, the hash function h(key, seed1) + index *h(key, seed2) is used which will result in an index in the lookup_table, the bit corresponding to this index will be the target value bit. This procedure is repeated for each bit of the target value.

Group Rebalancing Function Internals

When discussing EFD inserts and lookups, the discussion is simplified by assuming that a group id is simply a result of hash function. However, since hashing in general is not perfect and will not always produce a uniform output, this simplified assumption will lead to unbalanced groups, i.e., some group will have more keys than other groups. Typically, and to minimize insert time with an increasing number of keys, it is preferable that all groups will have a balanced number of keys, so the brute force search for the perfect hash terminates with a valid hash index. In order to achieve this target, groups are rebalanced during runtime inserts, and keys are moved around from a busy group to a less crowded group as the more keys are inserted.

_images/efd_i12.svg

Runtime Group Rebalancing

Fig. 17.14 depicts the high level idea of group rebalancing, given an input key the hash result is split into two parts a chunk id and 8-bit bin id. A chunk contains 64 different groups and 256 bins (i.e. for any given bin it can map to 4 distinct groups). When a key is inserted, the bin id is computed, for example in Fig. 17.14 bin_id=2, and since each bin can be mapped to one of four different groups (2 bit storage), the four possible mappings are evaluated and the one that will result in a balanced key distribution across these four is selected the mapping result is stored in these two bits.

References

1- EFD is based on collaborative research work between Intel and Carnegie Mellon University (CMU), interested readers can refer to the paper “Scaling Up Clustered Network Appliances with ScaleBricks;” Dong Zhou et al. at SIGCOMM 2015 (http://conferences.sigcomm.org/sigcomm/2015/pdf/papers/p241.pdf) for more information.

Membership Library

Introduction

The DPDK Membership Library provides an API for DPDK applications to insert a new member, delete an existing member, or query the existence of a member in a given set, or a group of sets. For the case of a group of sets, the library will return not only whether the element has been inserted before in one of the sets but also which set it belongs to. The Membership Library is an extension and generalization of a traditional filter structure (for example Bloom Filter [Member-bloom]) that has multiple usages in a wide variety of workloads and applications. In general, the Membership Library is a data structure that provides a “set-summary” on whether a member belongs to a set, and as discussed in detail later, there are two advantages of using such a set-summary rather than operating on a “full-blown” complete list of elements: first, it has a much smaller storage requirement than storing the whole list of elements themselves, and secondly checking an element membership (or other operations) in this set-summary is much faster than checking it for the original full-blown complete list of elements.

We use the term “Set-Summary” in this guide to refer to the space-efficient, probabilistic membership data structure that is provided by the library. A membership test for an element will return the set this element belongs to or that the element is “not-found” with very high probability of accuracy. Set-summary is a fundamental data aggregation component that can be used in many network (and other) applications. It is a crucial structure to address performance and scalability issues of diverse network applications including overlay networks, data-centric networks, flow table summaries, network statistics and traffic monitoring. A set-summary is useful for applications who need to include a list of elements while a complete list requires too much space and/or too much processing cost. In these situations, the set-summary works as a lossy hash-based representation of a set of members. It can dramatically reduce space requirement and significantly improve the performance of set membership queries at the cost of introducing a very small membership test error probability.

_images/member_i1.svg

Example Usages of Membership Library

There are various usages for a Membership Library in a very large set of applications and workloads. Interested readers can refer to [Member-survey] for a survey of possible networking usages. The above figure provide a small set of examples of using the Membership Library:

  • Sub-figure (a) depicts a distributed web cache architecture where a collection of proxies attempt to share their web caches (cached from a set of back-end web servers) to provide faster responses to clients, and the proxies use the Membership Library to share summaries of what web pages/objects they are caching. With the Membership Library, a proxy receiving an http request will inquire the set-summary to find its location and quickly determine whether to retrieve the requested web page from a nearby proxy or from a back-end web server.
  • Sub-figure (b) depicts another example for using the Membership Library to prevent routing loops which is typically done using slow TTL countdown and dropping packets when TTL expires. As shown in Sub-figure (b), an embedded set-summary in the packet header itself can be used to summarize the set of nodes a packet has gone through, and each node upon receiving a packet can check whether its id is a member of the set of visited nodes, and if it is, then a routing loop is detected.
  • Sub-Figure (c) presents another usage of the Membership Library to load-balance flows to worker threads with in-order guarantee where a set-summary is used to query if a packet belongs to an existing flow or a new flow. Packets belonging to a new flow are forwarded to the current least loaded worker thread, while those belonging to an existing flow are forwarded to the pre-assigned thread to guarantee in-order processing.
  • Sub-figure (d) highlights yet another usage example in the database domain where a set-summary is used to determine joins between sets instead of creating a join by comparing each element of a set against the other elements in a different set, a join is done on the summaries since they can efficiently encode members of a given set.

Membership Library is a configurable library that is optimized to cover set membership functionality for both a single set and multi-set scenarios. Two set-summary schemes are presented including (a) vector of Bloom Filters and (b) Hash-Table based set-summary schemes with and without false negative probability. This guide first briefly describes these different types of set-summaries, usage examples for each, and then it highlights the Membership Library API.

Vector of Bloom Filters

Bloom Filter (BF) [Member-bloom] is a well-known space-efficient probabilistic data structure that answers set membership queries (test whether an element is a member of a set) with some probability of false positives and zero false negatives; a query for an element returns either it is “possibly in a set” (with very high probability) or “definitely not in a set”.

The BF is a method for representing a set of n elements (for example flow keys in network applications domain) to support membership queries. The idea of BF is to allocate a bit-vector v with m bits, which are initially all set to 0. Then it chooses k independent hash functions h1, h2, ... hk with hash values range from 0 to m-1 to perform hashing calculations on each element to be inserted. Every time when an element X being inserted into the set, the bits at positions h1(X), h2(X), ... hk(X) in v are set to 1 (any particular bit might be set to 1 multiple times for multiple different inserted elements). Given a query for any element Y, the bits at positions h1(Y), h2(Y), ... hk(Y) are checked. If any of them is 0, then Y is definitely not in the set. Otherwise there is a high probability that Y is a member of the set with certain false positive probability. As shown in the next equation, the false positive probability can be made arbitrarily small by changing the number of hash functions (k) and the vector length (m).

_images/member_i2.svg

Bloom Filter False Positive Probability

Without BF, an accurate membership testing could involve a costly hash table lookup and full element comparison. The advantage of using a BF is to simplify the membership test into a series of hash calculations and memory accesses for a small bit-vector, which can be easily optimized. Hence the lookup throughput (set membership test) can be significantly faster than a normal hash table lookup with element comparison.

_images/member_i3.svg

Detecting Routing Loops Using BF

BF is used for applications that need only one set, and the membership of elements is checked against the BF. The example discussed in the above figure is one example of potential applications that uses only one set to capture the node IDs that have been visited so far by the packet. Each node will then check this embedded BF in the packet header for its own id, and if the BF indicates that the current node is definitely not in the set then a loop-free route is guaranteed.

_images/member_i4.svg

Vector Bloom Filter (vBF) Overview

To support membership test for both multiple sets and a single set, the library implements a Vector Bloom Filter (vBF) scheme. vBF basically composes multiple bloom filters into a vector of bloom filers. The membership test is conducted on all of the bloom filters concurrently to determine which set(s) it belongs to or none of them. The basic idea of vBF is shown in the above figure where an element is used to address multiple bloom filters concurrently and the bloom filter index(es) with a hit is returned.

_images/member_i5.svg

vBF for Flow Scheduling to Worker Thread

As previously mentioned, there are many usages of such structures. vBF is used for applications that need to check membership against multiple sets simultaneously. The example shown in the above figure uses a set to capture all flows being assigned for processing at a given worker thread. Upon receiving a packet the vBF is used to quickly figure out if this packet belongs to a new flow so as to be forwarded to the current least loaded worker thread, or otherwise it should be queued for an existing thread to guarantee in-order processing (i.e. the property of vBF to indicate right away that a given flow is a new one or not is critical to minimize response time latency).

It should be noted that vBF can be implemented using a set of single bloom filters with sequential lookup of each BF. However, being able to concurrently search all set-summaries is a big throughput advantage. In the library, certain parallelism is realized by the implementation of checking all bloom filters together.

Hash-Table based Set-Summaries

Hash-table based set-summary (HTSS) is another scheme in the membership library. Cuckoo filter [Member-cfilter] is an example of HTSS. HTSS supports multi-set membership testing like vBF does. However, while vBF is better for a small number of targets, HTSS is more suitable and can easily outperform vBF when the number of sets is large, since HTSS uses a single hash table for membership testing while vBF requires testing a series of Bloom Filters each corresponding to one set. As a result, generally speaking vBF is more adequate for the case of a small limited number of sets while HTSS should be used with a larger number of sets.

_images/member_i6.svg

Using HTSS for Attack Signature Matching

As shown in the above figure, attack signature matching where each set represents a certain signature length (for correctness of this example, an attack signature should not be a subset of another one) in the payload is a good example for using HTSS with 0% false negative (i.e., when an element returns not found, it has a 100% certainty that it is not a member of any set). The packet inspection application benefits from knowing right away that the current payload does not match any attack signatures in the database to establish its legitimacy, otherwise a deep inspection of the packet is needed.

HTSS employs a similar but simpler data structure to a traditional hash table, and the major difference is that HTSS stores only the signatures but not the full keys/elements which can significantly reduce the footprint of the table. Along with the signature, HTSS also stores a value to indicate the target set. When looking up an element, the element is hashed and the HTSS is addressed to retrieve the signature stored. If the signature matches then the value is retrieved corresponding to the index of the target set which the element belongs to. Because signatures can collide, HTSS can still has false positive probability. Furthermore, if elements are allowed to be overwritten or evicted when the hash table becomes full, it will also have a false negative probability. We discuss this case in the next section.

Set-Summaries with False Negative Probability

As previously mentioned, traditional set-summaries (e.g. Bloom Filters) do not have a false negative probability, i.e., it is 100% certain when an element returns “not to be present” for a given set. However, the Membership Library also supports a set-summary probabilistic data structure based on HTSS which allows for false negative probability.

In HTSS, when the hash table becomes full, keys/elements will fail to be added into the table and the hash table has to be resized to accommodate for these new elements, which can be expensive. However, if we allow new elements to overwrite or evict existing elements (as a cache typically does), then the resulting set-summary will begin to have false negative probability. This is because the element that was evicted from the set-summary may still be present in the target set. For subsequent inquiries the set-summary will falsely report the element not being in the set, hence having a false negative probability.

The major usage of HTSS with false negative is to use it as a cache for distributing elements to different target sets. By allowing HTSS to evict old elements, the set-summary can keep track of the most recent elements (i.e. active) as a cache typically does. Old inactive elements (infrequently used elements) will automatically and eventually get evicted from the set-summary. It is worth noting that the set-summary still has false positive probability, which means the application either can tolerate certain false positive or it has fall-back path when false positive happens.

_images/member_i7.svg

Using HTSS with False Negatives for Wild Card Classification

HTSS with false negative (i.e. a cache) also has its wide set of applications. For example wild card flow classification (e.g. ACL rules) highlighted in the above figure is an example of such application. In that case each target set represents a sub-table with rules defined by a certain flow mask. The flow masks are non-overlapping, and for flows matching more than one rule only the highest priority one is inserted in the corresponding sub-table (interested readers can refer to the Open vSwitch (OvS) design of Mega Flow Cache (MFC) [Member-OvS] for further details). Typically the rules will have a large number of distinct unique masks and hence, a large number of target sets each corresponding to one mask. Because the active set of flows varies widely based on the network traffic, HTSS with false negative will act as a cache for <flowid, target ACL sub-table> pair for the current active set of flows. When a miss occurs (as shown in red in the above figure) the sub-tables will be searched sequentially one by one for a possible match, and when found the flow key and target sub-table will be inserted into the set-summary (i.e. cache insertion) so subsequent packets from the same flow don’t incur the overhead of the sequential search of sub-tables.

Library API Overview

The design goal of the Membership Library API is to be as generic as possible to support all the different types of set-summaries we discussed in previous sections and beyond. Fundamentally, the APIs need to include creation, insertion, deletion, and lookup.

Set-summary Create

The rte_member_create() function is used to create a set-summary structure, the input parameter is a struct to pass in parameters that needed to initialize the set-summary, while the function returns the pointer to the created set-summary or NULL if the creation failed.

The general input arguments used when creating the set-summary should include name which is the name of the created set-summary, type which is one of the types supported by the library (e.g. RTE_MEMBER_TYPE_HT for HTSS or RTE_MEMBER_TYPE_VBF for vBF), and key_len which is the length of the element/key. There are other parameters are only used for certain type of set-summary, or which have a slightly different meaning for different types of set-summary. For example, num_keys parameter means the maximum number of entries for Hash table based set-summary. However, for bloom filter, this value means the expected number of keys that could be inserted into the bloom filter(s). The value is used to calculate the size of each bloom filter.

We also pass two seeds: prim_hash_seed and sec_hash_seed for the primary and secondary hash functions to calculate two independent hash values. socket_id parameter is the NUMA socket ID for the memory used to create the set-summary. For HTSS, another parameter is_cache is used to indicate if this set-summary is a cache (i.e. with false negative probability) or not. For vBF, extra parameters are needed. For example, num_set is the number of sets needed to initialize the vector bloom filters. This number is equal to the number of bloom filters will be created. false_pos_rate is the false positive rate. num_keys and false_pos_rate will be used to determine the number of hash functions and the bloom filter size.

Set-summary Element Insertion

The rte_member_add() function is used to insert an element/key into a set-summary structure. If it fails an error is returned. For success the returned value is dependent on the set-summary mode to provide extra information for the users. For vBF mode, a return value of 0 means a successful insert. For HTSS mode without false negative, the insert could fail with -ENOSPC if the table is full. With false negative (i.e. cache mode), for insert that does not cause any eviction (i.e. no overwriting happens to an existing entry) the return value is 0. For insertion that causes eviction, the return value is 1 to indicate such situation, but it is not an error.

The input arguments for the function should include the key which is a pointer to the element/key that needs to be added to the set-summary, and set_id which is the set id associated with the key that needs to be added.

Set-summary Element Lookup

The rte_member_lookup() function looks up a single key/element in the set-summary structure. It returns as soon as the first match is found. The return value is 1 if a match is found and 0 otherwise. The arguments for the function include key which is a pointer to the element/key that needs to be looked up, and set_id which is used to return the first target set id where the key has matched, if any.

The rte_member_lookup_bulk() function is used to look up a bulk of keys/elements in the set-summary structure for their first match. Each key lookup returns as soon as the first match is found. The return value is the number of keys that find a match. The arguments of the function include keys which is a pointer to a bulk of keys that are to be looked up, num_keys is the number of keys that will be looked up, and set_ids are the return target set ids for the first match found for each of the input keys. set_ids is an array needs to be sized according to the num_keys. If there is no match, the set id for that key will be set to RTE_MEMBER_NO_MATCH.

The rte_member_lookup_multi() function looks up a single key/element in the set-summary structure for multiple matches. It returns ALL the matches (possibly more than one) found for this key when it is matched against all target sets (it is worth noting that for cache mode HTSS, the current implementation matches at most one target set). The return value is the number of matches that was found for this key (for cache mode HTSS the return value should be at most 1). The arguments for the function include key which is a pointer to the element/key that needs to be looked up, max_match_per_key which is to indicate the maximum number of matches the user expects to find for each key, and set_id which is used to return all target set ids where the key has matched, if any. The set_id array should be sized according to max_match_per_key. For vBF, the maximum number of matches per key is equal to the number of sets. For HTSS, the maximum number of matches per key is equal to two time entry count per bucket. max_match_per_key should be equal or smaller than the maximum number of possible matches.

The rte_membership_lookup_multi_bulk() function looks up a bulk of keys/elements in the set-summary structure for multiple matches, each key lookup returns ALL the matches (possibly more than one) found for this key when it is matched against all target sets (cache mode HTSS matches at most one target set). The return value is the number of keys that find one or more matches in the set-summary structure. The arguments of the function include keys which is a pointer to a bulk of keys that are to be looked up, num_keys is the number of keys that will be looked up, max_match_per_key is the possible maximum number of matches for each key, match_count which is the returned number of matches for each key, and set_ids are the returned target set ids for all matches found for each keys. set_ids is 2-D array containing a 1-D array for each key (the size of 1-D array per key should be set by the user according to max_match_per_key). max_match_per_key should be equal or smaller than the maximum number of possible matches, similar to rte_member_lookup_multi.

Set-summary Element Delete

The rte_membership_delete() function deletes an element/key from a set-summary structure, if it fails an error is returned. The input arguments should include key which is a pointer to the element/key that needs to be deleted from the set-summary, and set_id which is the set id associated with the key to delete. It is worth noting that current implementation of vBF does not support deletion [1]. An error code -EINVAL will be returned.

[1]Traditional bloom filter does not support proactive deletion. Supporting proactive deletion require additional implementation and performance overhead.

References

[Member-bloom] B H Bloom, “Space/Time Trade-offs in Hash Coding with Allowable Errors,” Communications of the ACM, 1970.

[Member-survey] A Broder and M Mitzenmacher, “Network Applications of Bloom Filters: A Survey,” in Internet Mathematics, 2005.

[Member-cfilter] B Fan, D G Andersen and M Kaminsky, “Cuckoo Filter: Practically Better Than Bloom,” in Conference on emerging Networking Experiments and Technologies, 2014.

[Member-OvS] B Pfaff, “The Design and Implementation of Open vSwitch,” in NSDI, 2015.

LPM Library

The DPDK LPM library component implements the Longest Prefix Match (LPM) table search method for 32-bit keys that is typically used to find the best route match in IP forwarding applications.

LPM API Overview

The main configuration parameter for LPM component instances is the maximum number of rules to support. An LPM prefix is represented by a pair of parameters (32- bit key, depth), with depth in the range of 1 to 32. An LPM rule is represented by an LPM prefix and some user data associated with the prefix. The prefix serves as the unique identifier of the LPM rule. In this implementation, the user data is 1-byte long and is called next hop, in correlation with its main use of storing the ID of the next hop in a routing table entry.

The main methods exported by the LPM component are:

  • Add LPM rule: The LPM rule is provided as input. If there is no rule with the same prefix present in the table, then the new rule is added to the LPM table. If a rule with the same prefix is already present in the table, the next hop of the rule is updated. An error is returned when there is no available rule space left.
  • Delete LPM rule: The prefix of the LPM rule is provided as input. If a rule with the specified prefix is present in the LPM table, then it is removed.
  • Lookup LPM key: The 32-bit key is provided as input. The algorithm selects the rule that represents the best match for the given key and returns the next hop of that rule. In the case that there are multiple rules present in the LPM table that have the same 32-bit key, the algorithm picks the rule with the highest depth as the best match rule, which means that the rule has the highest number of most significant bits matching between the input key and the rule key.

Implementation Details

The current implementation uses a variation of the DIR-24-8 algorithm that trades memory usage for improved LPM lookup speed. The algorithm allows the lookup operation to be performed with typically a single memory read access. In the statistically rare case when the best match rule is having a depth bigger than 24, the lookup operation requires two memory read accesses. Therefore, the performance of the LPM lookup operation is greatly influenced by whether the specific memory location is present in the processor cache or not.

The main data structure is built using the following elements:

  • A table with 2^24 entries.
  • A number of tables (RTE_LPM_TBL8_NUM_GROUPS) with 2^8 entries.

The first table, called tbl24, is indexed using the first 24 bits of the IP address to be looked up, while the second table(s), called tbl8, is indexed using the last 8 bits of the IP address. This means that depending on the outcome of trying to match the IP address of an incoming packet to the rule stored in the tbl24 we might need to continue the lookup process in the second level.

Since every entry of the tbl24 can potentially point to a tbl8, ideally, we would have 2^24 tbl8s, which would be the same as having a single table with 2^32 entries. This is not feasible due to resource restrictions. Instead, this approach takes advantage of the fact that rules longer than 24 bits are very rare. By splitting the process in two different tables/levels and limiting the number of tbl8s, we can greatly reduce memory consumption while maintaining a very good lookup speed (one memory access, most of the times).

_images/tbl24_tbl8.png

Table split into different levels

An entry in tbl24 contains the following fields:

  • next hop / index to the tbl8
  • valid flag
  • external entry flag
  • depth of the rule (length)

The first field can either contain a number indicating the tbl8 in which the lookup process should continue or the next hop itself if the longest prefix match has already been found. The two flags are used to determine whether the entry is valid or not and whether the search process have finished or not respectively. The depth or length of the rule is the number of bits of the rule that is stored in a specific entry.

An entry in a tbl8 contains the following fields:

  • next hop
  • valid
  • valid group
  • depth

Next hop and depth contain the same information as in the tbl24. The two flags show whether the entry and the table are valid respectively.

The other main data structure is a table containing the main information about the rules (IP and next hop). This is a higher level table, used for different things:

  • Check whether a rule already exists or not, prior to addition or deletion, without having to actually perform a lookup.
  • When deleting, to check whether there is a rule containing the one that is to be deleted. This is important, since the main data structure will have to be updated accordingly.
Addition

When adding a rule, there are different possibilities. If the rule’s depth is exactly 24 bits, then:

  • Use the rule (IP address) as an index to the tbl24.
  • If the entry is invalid (i.e. it doesn’t already contain a rule) then set its next hop to its value, the valid flag to 1 (meaning this entry is in use), and the external entry flag to 0 (meaning the lookup process ends at this point, since this is the longest prefix that matches).

If the rule’s depth is exactly 32 bits, then:

  • Use the first 24 bits of the rule as an index to the tbl24.
  • If the entry is invalid (i.e. it doesn’t already contain a rule) then look for a free tbl8, set the index to the tbl8 to this value, the valid flag to 1 (meaning this entry is in use), and the external entry flag to 1 (meaning the lookup process must continue since the rule hasn’t been explored completely).

If the rule’s depth is any other value, prefix expansion must be performed. This means the rule is copied to all the entries (as long as they are not in use) which would also cause a match.

As a simple example, let’s assume the depth is 20 bits. This means that there are 2^(24 - 20) = 16 different combinations of the first 24 bits of an IP address that would cause a match. Hence, in this case, we copy the exact same entry to every position indexed by one of these combinations.

By doing this we ensure that during the lookup process, if a rule matching the IP address exists, it is found in either one or two memory accesses, depending on whether we need to move to the next table or not. Prefix expansion is one of the keys of this algorithm, since it improves the speed dramatically by adding redundancy.

Lookup

The lookup process is much simpler and quicker. In this case:

  • Use the first 24 bits of the IP address as an index to the tbl24. If the entry is not in use, then it means we don’t have a rule matching this IP. If it is valid and the external entry flag is set to 0, then the next hop is returned.
  • If it is valid and the external entry flag is set to 1, then we use the tbl8 index to find out the tbl8 to be checked, and the last 8 bits of the IP address as an index to this table. Similarly, if the entry is not in use, then we don’t have a rule matching this IP address. If it is valid then the next hop is returned.
Limitations in the Number of Rules

There are different things that limit the number of rules that can be added. The first one is the maximum number of rules, which is a parameter passed through the API. Once this number is reached, it is not possible to add any more rules to the routing table unless one or more are removed.

The second reason is an intrinsic limitation of the algorithm. As explained before, to avoid high memory consumption, the number of tbl8s is limited in compilation time (this value is by default 256). If we exhaust tbl8s, we won’t be able to add any more rules. How many of them are necessary for a specific routing table is hard to determine in advance.

A tbl8 is consumed whenever we have a new rule with depth bigger than 24, and the first 24 bits of this rule are not the same as the first 24 bits of a rule previously added. If they are, then the new rule will share the same tbl8 than the previous one, since the only difference between the two rules is within the last byte.

With the default value of 256, we can have up to 256 rules longer than 24 bits that differ on their first three bytes. Since routes longer than 24 bits are unlikely, this shouldn’t be a problem in most setups. Even if it is, however, the number of tbl8s can be modified.

Use Case: IPv4 Forwarding

The LPM algorithm is used to implement Classless Inter-Domain Routing (CIDR) strategy used by routers implementing IPv4 forwarding.

References

LPM6 Library

The LPM6 (LPM for IPv6) library component implements the Longest Prefix Match (LPM) table search method for 128-bit keys that is typically used to find the best match route in IPv6 forwarding applications.

LPM6 API Overview

The main configuration parameters for the LPM6 library are:

  • Maximum number of rules: This defines the size of the table that holds the rules, and therefore the maximum number of rules that can be added.
  • Number of tbl8s: A tbl8 is a node of the trie that the LPM6 algorithm is based on.

This parameter is related to the number of rules you can have, but there is no way to accurately predict the number needed to hold a specific number of rules, since it strongly depends on the depth and IP address of every rule. One tbl8 consumes 1 kb of memory. As a recommendation, 65536 tbl8s should be sufficient to store several thousand IPv6 rules, but the number can vary depending on the case.

An LPM prefix is represented by a pair of parameters (128-bit key, depth), with depth in the range of 1 to 128. An LPM rule is represented by an LPM prefix and some user data associated with the prefix. The prefix serves as the unique identifier for the LPM rule. In this implementation, the user data is 21-bits long and is called “next hop”, which corresponds to its main use of storing the ID of the next hop in a routing table entry.

The main methods exported for the LPM component are:

  • Add LPM rule: The LPM rule is provided as input. If there is no rule with the same prefix present in the table, then the new rule is added to the LPM table. If a rule with the same prefix is already present in the table, the next hop of the rule is updated. An error is returned when there is no available space left.
  • Delete LPM rule: The prefix of the LPM rule is provided as input. If a rule with the specified prefix is present in the LPM table, then it is removed.
  • Lookup LPM key: The 128-bit key is provided as input. The algorithm selects the rule that represents the best match for the given key and returns the next hop of that rule. In the case that there are multiple rules present in the LPM table that have the same 128-bit value, the algorithm picks the rule with the highest depth as the best match rule, which means the rule has the highest number of most significant bits matching between the input key and the rule key.
Implementation Details

This is a modification of the algorithm used for IPv4 (see Implementation Details). In this case, instead of using two levels, one with a tbl24 and a second with a tbl8, 14 levels are used.

The implementation can be seen as a multi-bit trie where the stride or number of bits inspected on each level varies from level to level. Specifically, 24 bits are inspected on the root node, and the remaining 104 bits are inspected in groups of 8 bits. This effectively means that the trie has 14 levels at the most, depending on the rules that are added to the table.

The algorithm allows the lookup operation to be performed with a number of memory accesses that directly depends on the length of the rule and whether there are other rules with bigger depths and the same key in the data structure. It can vary from 1 to 14 memory accesses, with 5 being the average value for the lengths that are most commonly used in IPv6.

The main data structure is built using the following elements:

  • A table with 224 entries
  • A number of tables, configurable by the user through the API, with 28 entries

The first table, called tbl24, is indexed using the first 24 bits of the IP address be looked up, while the rest of the tables, called tbl8s, are indexed using the rest of the bytes of the IP address, in chunks of 8 bits. This means that depending on the outcome of trying to match the IP address of an incoming packet to the rule stored in the tbl24 or the subsequent tbl8s we might need to continue the lookup process in deeper levels of the tree.

Similar to the limitation presented in the algorithm for IPv4, to store every possible IPv6 rule, we would need a table with 2^128 entries. This is not feasible due to resource restrictions.

By splitting the process in different tables/levels and limiting the number of tbl8s, we can greatly reduce memory consumption while maintaining a very good lookup speed (one memory access per level).

_images/tbl24_tbl8_tbl8.png

Table split into different levels

An entry in a table contains the following fields:

  • next hop / index to the tbl8
  • depth of the rule (length)
  • valid flag
  • valid group flag
  • external entry flag

The first field can either contain a number indicating the tbl8 in which the lookup process should continue or the next hop itself if the longest prefix match has already been found. The depth or length of the rule is the number of bits of the rule that is stored in a specific entry. The flags are used to determine whether the entry/table is valid or not and whether the search process have finished or not respectively.

Both types of tables share the same structure.

The other main data structure is a table containing the main information about the rules (IP, next hop and depth). This is a higher level table, used for different things:

  • Check whether a rule already exists or not, prior to addition or deletion, without having to actually perform a lookup.

When deleting, to check whether there is a rule containing the one that is to be deleted. This is important, since the main data structure will have to be updated accordingly.

Addition

When adding a rule, there are different possibilities. If the rule’s depth is exactly 24 bits, then:

  • Use the rule (IP address) as an index to the tbl24.
  • If the entry is invalid (i.e. it doesn’t already contain a rule) then set its next hop to its value, the valid flag to 1 (meaning this entry is in use), and the external entry flag to 0 (meaning the lookup process ends at this point, since this is the longest prefix that matches).

If the rule’s depth is bigger than 24 bits but a multiple of 8, then:

  • Use the first 24 bits of the rule as an index to the tbl24.
  • If the entry is invalid (i.e. it doesn’t already contain a rule) then look for a free tbl8, set the index to the tbl8 to this value, the valid flag to 1 (meaning this entry is in use), and the external entry flag to 1 (meaning the lookup process must continue since the rule hasn’t been explored completely).
  • Use the following 8 bits of the rule as an index to the next tbl8.
  • Repeat the process until the tbl8 at the right level (depending on the depth) has been reached and fill it with the next hop, setting the next entry flag to 0.

If the rule’s depth is any other value, prefix expansion must be performed. This means the rule is copied to all the entries (as long as they are not in use) which would also cause a match.

As a simple example, let’s assume the depth is 20 bits. This means that there are 2^(24-20) = 16 different combinations of the first 24 bits of an IP address that would cause a match. Hence, in this case, we copy the exact same entry to every position indexed by one of these combinations.

By doing this we ensure that during the lookup process, if a rule matching the IP address exists, it is found in, at the most, 14 memory accesses, depending on how many times we need to move to the next table. Prefix expansion is one of the keys of this algorithm, since it improves the speed dramatically by adding redundancy.

Prefix expansion can be performed at any level. So, for example, is the depth is 34 bits, it will be performed in the third level (second tbl8-based level).

Lookup

The lookup process is much simpler and quicker. In this case:

  • Use the first 24 bits of the IP address as an index to the tbl24. If the entry is not in use, then it means we don’t have a rule matching this IP. If it is valid and the external entry flag is set to 0, then the next hop is returned.
  • If it is valid and the external entry flag is set to 1, then we use the tbl8 index to find out the tbl8 to be checked, and the next 8 bits of the IP address as an index to this table. Similarly, if the entry is not in use, then we don’t have a rule matching this IP address. If it is valid then check the external entry flag for a new tbl8 to be inspected.
  • Repeat the process until either we find an invalid entry (lookup miss) or a valid entry with the external entry flag set to 0. Return the next hop in the latter case.
Limitations in the Number of Rules

There are different things that limit the number of rules that can be added. The first one is the maximum number of rules, which is a parameter passed through the API. Once this number is reached, it is not possible to add any more rules to the routing table unless one or more are removed.

The second limitation is in the number of tbl8s available. If we exhaust tbl8s, we won’t be able to add any more rules. How to know how many of them are necessary for a specific routing table is hard to determine in advance.

In this algorithm, the maximum number of tbl8s a single rule can consume is 13, which is the number of levels minus one, since the first three bytes are resolved in the tbl24. However:

  • Typically, on IPv6, routes are not longer than 48 bits, which means rules usually take up to 3 tbl8s.

As explained in the LPM for IPv4 algorithm, it is possible and very likely that several rules will share one or more tbl8s, depending on what their first bytes are. If they share the same first 24 bits, for instance, the tbl8 at the second level will be shared. This might happen again in deeper levels, so, effectively, two 48 bit-long rules may use the same three tbl8s if the only difference is in their last byte.

The number of tbl8s is a parameter exposed to the user through the API in this version of the algorithm, due to its impact in memory consumption and the number or rules that can be added to the LPM table. One tbl8 consumes 1 kilobyte of memory.

Use Case: IPv6 Forwarding

The LPM algorithm is used to implement the Classless Inter-Domain Routing (CIDR) strategy used by routers implementing IP forwarding.

Flow Classification Library

DPDK provides a Flow Classification library that provides the ability to classify an input packet by matching it against a set of Flow rules.

The initial implementation supports counting of IPv4 5-tuple packets which match a particular Flow rule only.

Please refer to the Generic flow API (rte_flow) for more information.

The Flow Classification library uses the librte_table API for managing Flow rules and matching packets against the Flow rules. The library is table agnostic and can use the following tables: Access Control List, Hash and Longest Prefix Match(LPM). The Access Control List table is used in the initial implementation.

Please refer to the Packet Framework for more information.on librte_table.

DPDK provides an Access Control List library that provides the ability to classify an input packet based on a set of classification rules.

Please refer to the Packet Classification and Access Control library for more information on librte_acl.

There is also a Flow Classify sample application which demonstrates the use of the Flow Classification Library API’s.

Please refer to the Flow Classify Sample Application for more information on the flow_classify sample application.

Overview

The library has the following API’s

/**
 * Flow classifier create
 *
 * @param params
 *   Parameters for flow classifier creation
 * @return
 *   Handle to flow classifier instance on success or NULL otherwise
 */
struct rte_flow_classifier *
rte_flow_classifier_create(struct rte_flow_classifier_params *params);

/**
 * Flow classifier free
 *
 * @param cls
 *   Handle to flow classifier instance
 * @return
 *   0 on success, error code otherwise
 */
int
rte_flow_classifier_free(struct rte_flow_classifier *cls);

/**
 * Flow classify table create
 *
 * @param cls
 *   Handle to flow classifier instance
 * @param params
 *   Parameters for flow_classify table creation
 * @param table_id
 *   Table ID. Valid only within the scope of table IDs of the current
 *   classifier. Only returned after a successful invocation.
 * @return
 *   0 on success, error code otherwise
 */
int
rte_flow_classify_table_create(struct rte_flow_classifier *cls,
       struct rte_flow_classify_table_params *params,
       uint32_t *table_id);

/**
 * Add a flow classify rule to the flow_classifier table.
 *
 * @param[in] cls
 *   Flow classifier handle
 * @param[in] table_id
 *   id of table
 * @param[in] attr
 *   Flow rule attributes
 * @param[in] pattern
 *   Pattern specification (list terminated by the END pattern item).
 * @param[in] actions
 *   Associated actions (list terminated by the END pattern item).
 * @param[out] error
 *   Perform verbose error reporting if not NULL. Structure
 *   initialised in case of error only.
 * @return
 *   A valid handle in case of success, NULL otherwise.
 */
struct rte_flow_classify_rule *
rte_flow_classify_table_entry_add(struct rte_flow_classifier *cls,
        uint32_t table_id,
        const struct rte_flow_attr *attr,
        const struct rte_flow_item pattern[],
        const struct rte_flow_action actions[],
        struct rte_flow_error *error);

/**
 * Delete a flow classify rule from the flow_classifier table.
 *
 * @param[in] cls
 *   Flow classifier handle
 * @param[in] table_id
 *   id of table
 * @param[in] rule
 *   Flow classify rule
 * @return
 *   0 on success, error code otherwise.
 */
int
rte_flow_classify_table_entry_delete(struct rte_flow_classifier *cls,
        uint32_t table_id,
        struct rte_flow_classify_rule *rule);

/**
 * Query flow classifier for given rule.
 *
 * @param[in] cls
 *   Flow classifier handle
 * @param[in] table_id
 *   id of table
 * @param[in] pkts
 *   Pointer to packets to process
 * @param[in] nb_pkts
 *   Number of packets to process
 * @param[in] rule
 *   Flow classify rule
 * @param[in] stats
 *   Flow classify stats
 *
 * @return
 *   0 on success, error code otherwise.
 */
int
rte_flow_classifier_query(struct rte_flow_classifier *cls,
        uint32_t table_id,
        struct rte_mbuf **pkts,
        const uint16_t nb_pkts,
        struct rte_flow_classify_rule *rule,
        struct rte_flow_classify_stats *stats);
Classifier creation

The application creates the Classifier using the rte_flow_classifier_create API. The rte_flow_classify_params structure must be initialised by the application before calling the API.

struct rte_flow_classifier_params {
    /** flow classifier name */
    const char *name;

    /** CPU socket ID where memory for the flow classifier and its */
    /** elements (tables) should be allocated */
    int socket_id;

    /** Table type */
    enum rte_flow_classify_table_type type;
};

The Classifier has the following internal structures:

struct rte_table {
    /* Input parameters */
    struct rte_table_ops ops;
    uint32_t entry_size;
    enum rte_flow_classify_table_type type;

    /* Handle to the low-level table object */
    void *h_table;
};

#define RTE_FLOW_CLASSIFIER_MAX_NAME_SZ 256

struct rte_flow_classifier {
    /* Input parameters */
    char name[RTE_FLOW_CLASSIFIER_MAX_NAME_SZ];
    int socket_id;
    enum rte_flow_classify_table_type type;

    /* Internal tables */
    struct rte_table tables[RTE_FLOW_CLASSIFY_TABLE_MAX];
    uint32_t num_tables;
    uint16_t nb_pkts;
    struct rte_flow_classify_table_entry
        *entries[RTE_PORT_IN_BURST_SIZE_MAX];
} __rte_cache_aligned;
Adding a table to the Classifier

The application adds a table to the Classifier using the rte_flow_classify_table_create API. The rte_flow_classify_table_params structure must be initialised by the application before calling the API.

struct rte_flow_classify_table_params {
    /** Table operations (specific to each table type) */
    struct rte_table_ops *ops;

    /** Opaque param to be passed to the table create operation */
    void *arg_create;

    /** Memory size to be reserved per classifier object entry for */
    /** storing meta data */
    uint32_t table_metadata_size;
 };

To create an ACL table the rte_table_acl_params structure must be initialised and assigned to arg_create in the rte_flow_classify_table_params structure.

struct rte_table_acl_params {
    /** Name */
    const char *name;

    /** Maximum number of ACL rules in the table */
    uint32_t n_rules;

    /** Number of fields in the ACL rule specification */
    uint32_t n_rule_fields;

    /** Format specification of the fields of the ACL rule */
    struct rte_acl_field_def field_format[RTE_ACL_MAX_FIELDS];
};

The fields for the ACL rule must also be initialised by the application.

An ACL table can be added to the Classifier for each ACL rule, for example another table could be added for the IPv6 5-tuple rule.

Flow Parsing

The library currently supports three IPv4 5-tuple flow patterns, for UDP, TCP and SCTP.

/* Pattern for IPv4 5-tuple UDP filter */
static enum rte_flow_item_type pattern_ntuple_1[] = {
    RTE_FLOW_ITEM_TYPE_ETH,
    RTE_FLOW_ITEM_TYPE_IPV4,
    RTE_FLOW_ITEM_TYPE_UDP,
    RTE_FLOW_ITEM_TYPE_END,
};

/* Pattern for IPv4 5-tuple TCP filter */
static enum rte_flow_item_type pattern_ntuple_2[] = {
    RTE_FLOW_ITEM_TYPE_ETH,
    RTE_FLOW_ITEM_TYPE_IPV4,
    RTE_FLOW_ITEM_TYPE_TCP,
    RTE_FLOW_ITEM_TYPE_END,
};

/* Pattern for IPv4 5-tuple SCTP filter */
static enum rte_flow_item_type pattern_ntuple_3[] = {
    RTE_FLOW_ITEM_TYPE_ETH,
    RTE_FLOW_ITEM_TYPE_IPV4,
    RTE_FLOW_ITEM_TYPE_SCTP,
    RTE_FLOW_ITEM_TYPE_END,
};

The internal function flow_classify_parse_flow parses the IPv4 5-tuple pattern, attributes and actions and returns the 5-tuple data in the rte_eth_ntuple_filter structure.

static int
flow_classify_parse_flow(
               const struct rte_flow_attr *attr,
               const struct rte_flow_item pattern[],
               const struct rte_flow_action actions[],
               struct rte_flow_error *error)
Adding Flow Rules

The rte_flow_classify_table_entry_add API creates an rte_flow_classify object which contains the flow_classify id and type, the action, a union of add and delete keys and a union of rules. It uses the flow_classify_parse_flow internal function for parsing the flow parameters. The 5-tuple ACL key data is obtained from the rte_eth_ntuple_filter structure populated by the classify_parse_ntuple_filter function which parses the Flow rule.

struct acl_keys {
    struct rte_table_acl_rule_add_params key_add; /* add key */
    struct rte_table_acl_rule_delete_params key_del; /* delete key */
};

struct classify_rules {
    enum rte_flow_classify_rule_type type;
    union {
        struct rte_flow_classify_ipv4_5tuple ipv4_5tuple;
    } u;
};

struct rte_flow_classify {
    uint32_t id;  /* unique ID of classify object */
    struct rte_flow_action action; /* action when match found */
    struct classify_rules rules; /* union of rules */
    union {
        struct acl_keys key;
    } u;
    int key_found; /* rule key found in table */
    void *entry; /* pointer to buffer to hold rule meta data */
    void *entry_ptr; /* handle to the table entry for rule meta data */
};

It then calls the table[table_id].ops.f_add API to add the rule to the ACL table.

Deleting Flow Rules

The rte_flow_classify_table_entry_delete API calls the table[table_id].ops.f_delete API to delete a rule from the ACL table.

Packet Matching

The rte_flow_classifier_query API is used to find packets which match a given flow Flow rule in the table. This API calls the flow_classify_run internal function which calls the table[table_id].ops.f_lookup API to see if any packets in a burst match any of the Flow rules in the table. The meta data for the highest priority rule matched for each packet is returned in the entries array in the rte_flow_classify object. The internal function action_apply implements the Count action which is used to return data which matches a particular Flow rule.

The rte_flow_classifier_query API uses the following structures to return data to the application.

/** IPv4 5-tuple data */
struct rte_flow_classify_ipv4_5tuple {
    uint32_t dst_ip;         /**< Destination IP address in big endian. */
    uint32_t dst_ip_mask;    /**< Mask of destination IP address. */
    uint32_t src_ip;         /**< Source IP address in big endian. */
    uint32_t src_ip_mask;    /**< Mask of destination IP address. */
    uint16_t dst_port;       /**< Destination port in big endian. */
    uint16_t dst_port_mask;  /**< Mask of destination port. */
    uint16_t src_port;       /**< Source Port in big endian. */
    uint16_t src_port_mask;  /**< Mask of source port. */
    uint8_t proto;           /**< L4 protocol. */
    uint8_t proto_mask;      /**< Mask of L4 protocol. */
};

/**
 * Flow stats
 *
 * For the count action, stats can be returned by the query API.
 *
 * Storage for stats is provided by the application.
 *
 *
 */
struct rte_flow_classify_stats {
    void *stats;
};

struct rte_flow_classify_5tuple_stats {
    /** count of packets that match IPv4 5tuple pattern */
    uint64_t counter1;
    /** IPv4 5tuple data */
    struct rte_flow_classify_ipv4_5tuple ipv4_5tuple;
};

Packet Distributor Library

The DPDK Packet Distributor library is a library designed to be used for dynamic load balancing of traffic while supporting single packet at a time operation. When using this library, the logical cores in use are to be considered in two roles: firstly a distributor lcore, which is responsible for load balancing or distributing packets, and a set of worker lcores which are responsible for receiving the packets from the distributor and operating on them. The model of operation is shown in the diagram below.

_images/packet_distributor1.png

Packet Distributor mode of operation

There are two modes of operation of the API in the distributor library, one which sends one packet at a time to workers using 32-bits for flow_id, and an optimized mode which sends bursts of up to 8 packets at a time to workers, using 15 bits of flow_id. The mode is selected by the type field in the rte_distributor_create() function.

Distributor Core Operation

The distributor core does the majority of the processing for ensuring that packets are fairly shared among workers. The operation of the distributor is as follows:

  1. Packets are passed to the distributor component by having the distributor lcore thread call the “rte_distributor_process()” API
  2. The worker lcores all share a single cache line with the distributor core in order to pass messages and packets to and from the worker. The process API call will poll all the worker cache lines to see what workers are requesting packets.
  3. As workers request packets, the distributor takes packets from the set of packets passed in and distributes them to the workers. As it does so, it examines the “tag” – stored in the RSS hash field in the mbuf – for each packet and records what tags are being processed by each worker.
  4. If the next packet in the input set has a tag which is already being processed by a worker, then that packet will be queued up for processing by that worker and given to it in preference to other packets when that work next makes a request for work. This ensures that no two packets with the same tag are processed in parallel, and that all packets with the same tag are processed in input order.
  5. Once all input packets passed to the process API have either been distributed to workers or been queued up for a worker which is processing a given tag, then the process API returns to the caller.

Other functions which are available to the distributor lcore are:

  • rte_distributor_returned_pkts()
  • rte_distributor_flush()
  • rte_distributor_clear_returns()

Of these the most important API call is “rte_distributor_returned_pkts()” which should only be called on the lcore which also calls the process API. It returns to the caller all packets which have finished processing by all worker cores. Within this set of returned packets, all packets sharing the same tag will be returned in their original order.

NOTE: If worker lcores buffer up packets internally for transmission in bulk afterwards, the packets sharing a tag will likely get out of order. Once a worker lcore requests a new packet, the distributor assumes that it has completely finished with the previous packet and therefore that additional packets with the same tag can safely be distributed to other workers – who may then flush their buffered packets sooner and cause packets to get out of order.

NOTE: No packet ordering guarantees are made about packets which do not share a common packet tag.

Using the process and returned_pkts API, the following application workflow can be used, while allowing packet order within a packet flow – identified by a tag – to be maintained.

_images/packet_distributor2.png

Application workflow

The flush and clear_returns API calls, mentioned previously, are likely of less use that the process and returned_pkts APIS, and are principally provided to aid in unit testing of the library. Descriptions of these functions and their use can be found in the DPDK API Reference document.

Worker Operation

Worker cores are the cores which do the actual manipulation of the packets distributed by the packet distributor. Each worker calls “rte_distributor_get_pkt()” API to request a new packet when it has finished processing the previous one. [The previous packet should be returned to the distributor component by passing it as the final parameter to this API call.]

Since it may be desirable to vary the number of worker cores, depending on the traffic load i.e. to save power at times of lighter load, it is possible to have a worker stop processing packets by calling “rte_distributor_return_pkt()” to indicate that it has finished the current packet and does not want a new one.

Reorder Library

The Reorder Library provides a mechanism for reordering mbufs based on their sequence number.

Operation

The reorder library is essentially a buffer that reorders mbufs. The user inserts out of order mbufs into the reorder buffer and pulls in-order mbufs from it.

At a given time, the reorder buffer contains mbufs whose sequence number are inside the sequence window. The sequence window is determined by the minimum sequence number and the number of entries that the buffer was configured to hold. For example, given a reorder buffer with 200 entries and a minimum sequence number of 350, the sequence window has low and high limits of 350 and 550 respectively.

When inserting mbufs, the reorder library differentiates between valid, early and late mbufs depending on the sequence number of the inserted mbuf:

  • valid: the sequence number is inside the window.
  • late: the sequence number is outside the window and less than the low limit.
  • early: the sequence number is outside the window and greater than the high limit.

The reorder buffer directly returns late mbufs and tries to accommodate early mbufs.

Implementation Details

The reorder library is implemented as a pair of buffers, which referred to as the Order buffer and the Ready buffer.

On an insert call, valid mbufs are inserted directly into the Order buffer and late mbufs are returned to the user with an error.

In the case of early mbufs, the reorder buffer will try to move the window (incrementing the minimum sequence number) so that the mbuf becomes a valid one. To that end, mbufs in the Order buffer are moved into the Ready buffer. Any mbufs that have not arrived yet are ignored and therefore will become late mbufs. This means that as long as there is room in the Ready buffer, the window will be moved to accommodate early mbufs that would otherwise be outside the reordering window.

For example, assuming that we have a buffer of 200 entries with a 350 minimum sequence number, and we need to insert an early mbuf with 565 sequence number. That means that we would need to move the windows at least 15 positions to accommodate the mbuf. The reorder buffer would try to move mbufs from at least the next 15 slots in the Order buffer to the Ready buffer, as long as there is room in the Ready buffer. Any gaps in the Order buffer at that point are skipped, and those packet will be reported as late packets when they arrive. The process of moving packets to the Ready buffer continues beyond the minimum required until a gap, i.e. missing mbuf, in the Order buffer is encountered.

When draining mbufs, the reorder buffer would return mbufs in the Ready buffer first and then from the Order buffer until a gap is found (mbufs that have not arrived yet).

Use Case: Packet Distributor

An application using the DPDK packet distributor could make use of the reorder library to transmit packets in the same order they were received.

A basic packet distributor use case would consist of a distributor with multiple workers cores. The processing of packets by the workers is not guaranteed to be in order, hence a reorder buffer can be used to order as many packets as possible.

In such a scenario, the distributor assigns a sequence number to mbufs before delivering them to the workers. As the workers finish processing the packets, the distributor inserts those mbufs into the reorder buffer and finally transmit drained mbufs.

NOTE: Currently the reorder buffer is not thread safe so the same thread is responsible for inserting and draining mbufs.

IP Fragmentation and Reassembly Library

The IP Fragmentation and Reassembly Library implements IPv4 and IPv6 packet fragmentation and reassembly.

Packet fragmentation

Packet fragmentation routines divide input packet into number of fragments. Both rte_ipv4_fragment_packet() and rte_ipv6_fragment_packet() functions assume that input mbuf data points to the start of the IP header of the packet (i.e. L2 header is already stripped out). To avoid copying of the actual packet’s data zero-copy technique is used (rte_pktmbuf_attach). For each fragment two new mbufs are created:

  • Direct mbuf – mbuf that will contain L3 header of the new fragment.
  • Indirect mbuf – mbuf that is attached to the mbuf with the original packet. It’s data field points to the start of the original packets data plus fragment offset.

Then L3 header is copied from the original mbuf into the ‘direct’ mbuf and updated to reflect new fragmented status. Note that for IPv4, header checksum is not recalculated and is set to zero.

Finally ‘direct’ and ‘indirect’ mbufs for each fragment are linked together via mbuf’s next filed to compose a packet for the new fragment.

The caller has an ability to explicitly specify which mempools should be used to allocate ‘direct’ and ‘indirect’ mbufs from.

For more information about direct and indirect mbufs, refer to Direct and Indirect Buffers.

Packet reassembly

IP Fragment Table

Fragment table maintains information about already received fragments of the packet.

Each IP packet is uniquely identified by triple <Source IP address>, <Destination IP address>, <ID>.

Note that all update/lookup operations on Fragment Table are not thread safe. So if different execution contexts (threads/processes) will access the same table simultaneously, then some external syncing mechanism have to be provided.

Each table entry can hold information about packets consisting of up to RTE_LIBRTE_IP_FRAG_MAX (by default: 4) fragments.

Code example, that demonstrates creation of a new Fragment table:

frag_cycles = (rte_get_tsc_hz() + MS_PER_S - 1) / MS_PER_S * max_flow_ttl;
bucket_num = max_flow_num + max_flow_num / 4;
frag_tbl = rte_ip_frag_table_create(max_flow_num, bucket_entries, max_flow_num, frag_cycles, socket_id);

Internally Fragment table is a simple hash table. The basic idea is to use two hash functions and <bucket_entries> * associativity. This provides 2 * <bucket_entries> possible locations in the hash table for each key. When the collision occurs and all 2 * <bucket_entries> are occupied, instead of reinserting existing keys into alternative locations, ip_frag_tbl_add() just returns a failure.

Also, entries that resides in the table longer then <max_cycles> are considered as invalid, and could be removed/replaced by the new ones.

Note that reassembly demands a lot of mbuf’s to be allocated. At any given time up to (2 * bucket_entries * RTE_LIBRTE_IP_FRAG_MAX * <maximum number of mbufs per packet>) can be stored inside Fragment Table waiting for remaining fragments.

Packet Reassembly

Fragmented packets processing and reassembly is done by the rte_ipv4_frag_reassemble_packet()/rte_ipv6_frag_reassemble_packet. Functions. They either return a pointer to valid mbuf that contains reassembled packet, or NULL (if the packet can’t be reassembled for some reason).

These functions are responsible for:

  1. Search the Fragment Table for entry with packet’s <IPv4 Source Address, IPv4 Destination Address, Packet ID>.
  2. If the entry is found, then check if that entry already timed-out. If yes, then free all previously received fragments, and remove information about them from the entry.
  3. If no entry with such key is found, then try to create a new one by one of two ways:
    1. Use as empty entry.
    2. Delete a timed-out entry, free mbufs associated with it mbufs and store a new entry with specified key in it.
  4. Update the entry with new fragment information and check if a packet can be reassembled (the packet’s entry contains all fragments).
    1. If yes, then, reassemble the packet, mark table’s entry as empty and return the reassembled mbuf to the caller.
    2. If no, then return a NULL to the caller.

If at any stage of packet processing an error is encountered (e.g: can’t insert new entry into the Fragment Table, or invalid/timed-out fragment), then the function will free all associated with the packet fragments, mark the table entry as invalid and return NULL to the caller.

Debug logging and Statistics Collection

The RTE_LIBRTE_IP_FRAG_TBL_STAT config macro controls statistics collection for the Fragment Table. This macro is not enabled by default.

The RTE_LIBRTE_IP_FRAG_DEBUG controls debug logging of IP fragments processing and reassembling. This macro is disabled by default. Note that while logging contains a lot of detailed information, it slows down packet processing and might cause the loss of a lot of packets.

Generic Receive Offload Library

Generic Receive Offload (GRO) is a widely used SW-based offloading technique to reduce per-packet processing overhead. It gains performance by reassembling small packets into large ones. To enable more flexibility to applications, DPDK implements GRO as a standalone library. Applications explicitly use the GRO library to merge small packets into large ones.

The GRO library assumes all input packets have correct checksums. In addition, the GRO library doesn’t re-calculate checksums for merged packets. If input packets are IP fragmented, the GRO library assumes they are complete packets (i.e. with L4 headers).

Currently, the GRO library implements TCP/IPv4 packet reassembly.

Reassembly Modes

The GRO library provides two reassembly modes: lightweight and heavyweight mode. If applications want to merge packets in a simple way, they can use the lightweight mode API. If applications want more fine-grained controls, they can choose the heavyweight mode API.

Lightweight Mode

The rte_gro_reassemble_burst() function is used for reassembly in lightweight mode. It tries to merge N input packets at a time, where N should be less than or equal to RTE_GRO_MAX_BURST_ITEM_NUM.

In each invocation, rte_gro_reassemble_burst() allocates temporary reassembly tables for the desired GRO types. Note that the reassembly table is a table structure used to reassemble packets and different GRO types (e.g. TCP/IPv4 GRO and TCP/IPv6 GRO) have different reassembly table structures. The rte_gro_reassemble_burst() function uses the reassembly tables to merge the N input packets.

For applications, performing GRO in lightweight mode is simple. They just need to invoke rte_gro_reassemble_burst(). Applications can get GROed packets as soon as rte_gro_reassemble_burst() returns.

Heavyweight Mode

The rte_gro_reassemble() function is used for reassembly in heavyweight mode. Compared with the lightweight mode, performing GRO in heavyweight mode is relatively complicated.

Before performing GRO, applications need to create a GRO context object by calling rte_gro_ctx_create(). A GRO context object holds the reassembly tables of desired GRO types. Note that all update/lookup operations on the context object are not thread safe. So if different processes or threads want to access the same context object simultaneously, some external syncing mechanisms must be used.

Once the GRO context is created, applications can then use the rte_gro_reassemble() function to merge packets. In each invocation, rte_gro_reassemble() tries to merge input packets with the packets in the reassembly tables. If an input packet is an unsupported GRO type, or other errors happen (e.g. SYN bit is set), rte_gro_reassemble() returns the packet to applications. Otherwise, the input packet is either merged or inserted into a reassembly table.

When applications want to get GRO processed packets, they need to use rte_gro_timeout_flush() to flush them from the tables manually.

TCP/IPv4 GRO

TCP/IPv4 GRO supports merging small TCP/IPv4 packets into large ones, using a table structure called the TCP/IPv4 reassembly table.

TCP/IPv4 Reassembly Table

A TCP/IPv4 reassembly table includes a “key” array and an “item” array. The key array keeps the criteria to merge packets and the item array keeps the packet information.

Each key in the key array points to an item group, which consists of packets which have the same criteria values but can’t be merged. A key in the key array includes two parts:

  • criteria: the criteria to merge packets. If two packets can be merged, they must have the same criteria values.
  • start_index: the item array index of the first packet in the item group.

Each element in the item array keeps the information of a packet. An item in the item array mainly includes three parts:

  • firstseg: the mbuf address of the first segment of the packet.
  • lastseg: the mbuf address of the last segment of the packet.
  • next_pkt_index: the item array index of the next packet in the same item group. TCP/IPv4 GRO uses next_pkt_index to chain the packets that have the same criteria value but can’t be merged together.
Procedure to Reassemble a Packet

To reassemble an incoming packet needs three steps:

  1. Check if the packet should be processed. Packets with one of the following properties aren’t processed and are returned immediately:
    • FIN, SYN, RST, URG, PSH, ECE or CWR bit is set.
    • L4 payload length is 0.
  2. Traverse the key array to find a key which has the same criteria value with the incoming packet. If found, go to the next step. Otherwise, insert a new key and a new item for the packet.
  3. Locate the first packet in the item group via start_index. Then traverse all packets in the item group via next_pkt_index. If a packet is found which can be merged with the incoming one, merge them together. If one isn’t found, insert the packet into this item group. Note that to merge two packets is to link them together via mbuf’s next field.

When packets are flushed from the reassembly table, TCP/IPv4 GRO updates packet header fields for the merged packets. Note that before reassembling the packet, TCP/IPv4 GRO doesn’t check if the checksums of packets are correct. Also, TCP/IPv4 GRO doesn’t re-calculate checksums for merged packets.

Generic Segmentation Offload Library

Overview

Generic Segmentation Offload (GSO) is a widely used software implementation of TCP Segmentation Offload (TSO), which reduces per-packet processing overhead. Much like TSO, GSO gains performance by enabling upper layer applications to process a smaller number of large packets (e.g. MTU size of 64KB), instead of processing higher numbers of small packets (e.g. MTU size of 1500B), thus reducing per-packet overhead.

For example, GSO allows guest kernel stacks to transmit over-sized TCP segments that far exceed the kernel interface’s MTU; this eliminates the need to segment packets within the guest, and improves the data-to-overhead ratio of both the guest-host link, and PCI bus. The expectation of the guest network stack in this scenario is that segmentation of egress frames will take place either in the NIC HW, or where that hardware capability is unavailable, either in the host application, or network stack.

Bearing that in mind, the GSO library enables DPDK applications to segment packets in software. Note however, that GSO is implemented as a standalone library, and not via a ‘fallback’ mechanism (i.e. for when TSO is unsupported in the underlying hardware); that is, applications must explicitly invoke the GSO library to segment packets. The size of GSO segments (segsz) is configurable by the application.

Limitations

  1. The GSO library doesn’t check if input packets have correct checksums.
  2. In addition, the GSO library doesn’t re-calculate checksums for segmented packets (that task is left to the application).
  3. IP fragments are unsupported by the GSO library.
  4. The egress interface’s driver must support multi-segment packets.
  5. Currently, the GSO library supports the following IPv4 packet types:
  • TCP
  • VxLAN
  • GRE
See Supported GSO Packet Types for further details.

Packet Segmentation

The rte_gso_segment() function is the GSO library’s primary segmentation API.

Before performing segmentation, an application must create a GSO context object (struct rte_gso_ctx), which provides the library with some of the information required to understand how the packet should be segmented. Refer to How to Segment a Packet for additional details on same. Once the GSO context has been created, and populated, the application can then use the rte_gso_segment() function to segment packets.

The GSO library typically stores each segment that it creates in two parts: the first part contains a copy of the original packet’s headers, while the second part contains a pointer to an offset within the original packet. This mechanism is explained in more detail in GSO Output Segment Format.

The GSO library supports both single- and multi-segment input mbufs.

GSO Output Segment Format

To reduce the number of expensive memcpy operations required when segmenting a packet, the GSO library typically stores each segment that it creates as a two-part mbuf (technically, this is termed a ‘two-segment’ mbuf; however, since the elements produced by the API are also called ‘segments’, for clarity the term ‘part’ is used here instead).

The first part of each output segment is a direct mbuf and contains a copy of the original packet’s headers, which must be prepended to each output segment. These headers are copied from the original packet into each output segment.

The second part of each output segment, represents a section of data from the original packet, i.e. a data segment. Rather than copy the data directly from the original packet into the output segment (which would impact performance considerably), the second part of each output segment is an indirect mbuf, which contains no actual data, but simply points to an offset within the original packet.

The combination of the ‘header’ segment and the ‘data’ segment constitutes a single logical output GSO segment of the original packet. This is illustrated in Fig. 26.2.

_images/gso-output-segment-format.svg

Two-part GSO output segment

In one situation, the output segment may contain additional ‘data’ segments. This only occurs when:

  • the input packet on which GSO is to be performed is represented by a multi-segment mbuf.
  • the output segment is required to contain data that spans the boundaries between segments of the input multi-segment mbuf.

The GSO library traverses each segment of the input packet, and produces numerous output segments; for optimal performance, the number of output segments is kept to a minimum. Consequently, the GSO library maximizes the amount of data contained within each output segment; i.e. each output segment segsz bytes of data. The only exception to this is in the case of the very final output segment; if pkt_len % segsz, then the final segment is smaller than the rest.

In order for an output segment to meet its MSS, it may need to include data from multiple input segments. Due to the nature of indirect mbufs (each indirect mbuf can point to only one direct mbuf), the solution here is to add another indirect mbuf to the output segment; this additional segment then points to the next input segment. If necessary, this chaining process is repeated, until the sum of all of the data ‘contained’ in the output segment reaches segsz. This ensures that the amount of data contained within each output segment is uniform, with the possible exception of the last segment, as previously described.

Fig. 26.3 illustrates an example of a three-part output segment. In this example, the output segment needs to include data from the end of one input segment, and the beginning of another. To achieve this, an additional indirect mbuf is chained to the second part of the output segment, and is attached to the next input segment (i.e. it points to the data in the next input segment).

_images/gso-three-seg-mbuf.svg

Three-part GSO output segment

Supported GSO Packet Types

TCP/IPv4 GSO

TCP/IPv4 GSO supports segmentation of suitably large TCP/IPv4 packets, which may also contain an optional VLAN tag.

VxLAN GSO

VxLAN packets GSO supports segmentation of suitably large VxLAN packets, which contain an outer IPv4 header, inner TCP/IPv4 headers, and optional inner and/or outer VLAN tag(s).

GRE GSO

GRE GSO supports segmentation of suitably large GRE packets, which contain an outer IPv4 header, inner TCP/IPv4 headers, and an optional VLAN tag.

How to Segment a Packet

To segment an outgoing packet, an application must:

  1. First create a GSO context (struct rte_gso_ctx); this contains:

    • a pointer to the mbuf pool for allocating the direct buffers, which are used to store the GSO segments’ packet headers.

    • a pointer to the mbuf pool for allocating indirect buffers, which are used to locate GSO segments’ packet payloads.

      Note

      An application may use the same pool for both direct and indirect buffers. However, since indirect mbufs simply store a pointer, the application may reduce its memory consumption by creating a separate memory pool, containing smaller elements, for the indirect pool.

    • the size of each output segment, including packet headers and payload, measured in bytes.

    • the bit mask of required GSO types. The GSO library uses the same macros as those that describe a physical device’s TX offloading capabilities (i.e. DEV_TX_OFFLOAD_*_TSO) for gso_types. For example, if an application wants to segment TCP/IPv4 packets, it should set gso_types to DEV_TX_OFFLOAD_TCP_TSO. The only other supported values currently supported for gso_types are DEV_TX_OFFLOAD_VXLAN_TNL_TSO, and DEV_TX_OFFLOAD_GRE_TNL_TSO; a combination of these macros is also allowed.

    • a flag, that indicates whether the IPv4 headers of output segments should contain fixed or incremental ID values.

  1. Set the appropriate ol_flags in the mbuf.
    • The GSO library use the value of an mbuf’s ol_flags attribute to to determine how a packet should be segmented. It is the application’s responsibility to ensure that these flags are set.
    • For example, in order to segment TCP/IPv4 packets, the application should add the PKT_TX_IPV4 and PKT_TX_TCP_SEG flags to the mbuf’s ol_flags.
    • If checksum calculation in hardware is required, the application should also add the PKT_TX_TCP_CKSUM and PKT_TX_IP_CKSUM flags.
  2. Check if the packet should be processed. Packets with one of the following properties are not processed and are returned immediately:
    • Packet length is less than segsz (i.e. GSO is not required).
    • Packet type is not supported by GSO library (see Supported GSO Packet Types).
    • Application has not enabled GSO support for the packet type.
    • Packet’s ol_flags have been incorrectly set.
  3. Allocate space in which to store the output GSO segments. If the amount of space allocated by the application is insufficient, segmentation will fail.
  4. Invoke the GSO segmentation API, rte_gso_segment().
  5. If required, update the L3 and L4 checksums of the newly-created segments. For tunneled packets, the outer IPv4 headers’ checksums should also be updated. Alternatively, the application may offload checksum calculation to HW.

The librte_pdump Library

The librte_pdump library provides a framework for packet capturing in DPDK. The library does the complete copy of the Rx and Tx mbufs to a new mempool and hence it slows down the performance of the applications, so it is recommended to use this library for debugging purposes.

The library provides the following APIs to initialize the packet capture framework, to enable or disable the packet capture, and to uninitialize it:

  • rte_pdump_init(): This API initializes the packet capture framework.
  • rte_pdump_enable(): This API enables the packet capture on a given port and queue. Note: The filter option in the API is a place holder for future enhancements.
  • rte_pdump_enable_by_deviceid(): This API enables the packet capture on a given device id (vdev name or pci address) and queue. Note: The filter option in the API is a place holder for future enhancements.
  • rte_pdump_disable(): This API disables the packet capture on a given port and queue.
  • rte_pdump_disable_by_deviceid(): This API disables the packet capture on a given device id (vdev name or pci address) and queue.
  • rte_pdump_uninit(): This API uninitializes the packet capture framework.
  • rte_pdump_set_socket_dir(): This API sets the server and client socket paths. Note: This API is not thread-safe.

Operation

The librte_pdump library works on a client/server model. The server is responsible for enabling or disabling the packet capture and the clients are responsible for requesting the enabling or disabling of the packet capture.

The packet capture framework, as part of its initialization, creates the pthread and the server socket in the pthread. The application that calls the framework initialization will have the server socket created, either under the path that the application has passed or under the default path i.e. either /var/run/.dpdk for root user or ~/.dpdk for non root user.

Applications that request enabling or disabling of the packet capture will have the client socket created either under the path that the application has passed or under the default path i.e. either /var/run/.dpdk for root user or ~/.dpdk for not root user to send the requests to the server. The server socket will listen for client requests for enabling or disabling the packet capture.

Implementation Details

The library API rte_pdump_init(), initializes the packet capture framework by creating the pthread and the server socket. The server socket in the pthread context will be listening to the client requests to enable or disable the packet capture.

The library APIs rte_pdump_enable() and rte_pdump_enable_by_deviceid() enables the packet capture. On each call to these APIs, the library creates a separate client socket, creates the “pdump enable” request and sends the request to the server. The server that is listening on the socket will take the request and enable the packet capture by registering the Ethernet RX and TX callbacks for the given port or device_id and queue combinations. Then the server will mirror the packets to the new mempool and enqueue them to the rte_ring that clients have passed to these APIs. The server also sends the response back to the client about the status of the request that was processed. After the response is received from the server, the client socket is closed.

The library APIs rte_pdump_disable() and rte_pdump_disable_by_deviceid() disables the packet capture. On each call to these APIs, the library creates a separate client socket, creates the “pdump disable” request and sends the request to the server. The server that is listening on the socket will take the request and disable the packet capture by removing the Ethernet RX and TX callbacks for the given port or device_id and queue combinations. The server also sends the response back to the client about the status of the request that was processed. After the response is received from the server, the client socket is closed.

The library API rte_pdump_uninit(), uninitializes the packet capture framework by closing the pthread and the server socket.

The library API rte_pdump_set_socket_dir(), sets the given path as either server socket path or client socket path based on the type argument of the API. If the given path is NULL, default path will be selected, i.e. either /var/run/.dpdk for root user or ~/.dpdk for non root user. Clients also need to call this API to set their server socket path if the server socket path is different from default path.

Use Case: Packet Capturing

The DPDK app/pdump tool is developed based on this library to capture packets in DPDK. Users can use this as an example to develop their own packet capturing tools.

Multi-process Support

In the DPDK, multi-process support is designed to allow a group of DPDK processes to work together in a simple transparent manner to perform packet processing, or other workloads. To support this functionality, a number of additions have been made to the core DPDK Environment Abstraction Layer (EAL).

The EAL has been modified to allow different types of DPDK processes to be spawned, each with different permissions on the hugepage memory used by the applications. For now, there are two types of process specified:

  • primary processes, which can initialize and which have full permissions on shared memory
  • secondary processes, which cannot initialize shared memory, but can attach to pre- initialized shared memory and create objects in it.

Standalone DPDK processes are primary processes, while secondary processes can only run alongside a primary process or after a primary process has already configured the hugepage shared memory for them.

To support these two process types, and other multi-process setups described later, two additional command-line parameters are available to the EAL:

  • --proc-type: for specifying a given process instance as the primary or secondary DPDK instance
  • --file-prefix: to allow processes that do not want to co-operate to have different memory regions

A number of example applications are provided that demonstrate how multiple DPDK processes can be used together. These are more fully documented in the “Multi- process Sample Application” chapter in the DPDK Sample Application’s User Guide.

Memory Sharing

The key element in getting a multi-process application working using the DPDK is to ensure that memory resources are properly shared among the processes making up the multi-process application. Once there are blocks of shared memory available that can be accessed by multiple processes, then issues such as inter-process communication (IPC) becomes much simpler.

On application start-up in a primary or standalone process, the DPDK records to memory-mapped files the details of the memory configuration it is using - hugepages in use, the virtual addresses they are mapped at, the number of memory channels present, etc. When a secondary process is started, these files are read and the EAL recreates the same memory configuration in the secondary process so that all memory zones are shared between processes and all pointers to that memory are valid, and point to the same objects, in both processes.

Note

Refer to Multi-process Limitations for details of how Linux kernel Address-Space Layout Randomization (ASLR) can affect memory sharing.

_images/multi_process_memory.svg

Memory Sharing in the DPDK Multi-process Sample Application

The EAL also supports an auto-detection mode (set by EAL --proc-type=auto flag ), whereby an DPDK process is started as a secondary instance if a primary instance is already running.

Deployment Models

Symmetric/Peer Processes

DPDK multi-process support can be used to create a set of peer processes where each process performs the same workload. This model is equivalent to having multiple threads each running the same main-loop function, as is done in most of the supplied DPDK sample applications. In this model, the first of the processes spawned should be spawned using the --proc-type=primary EAL flag, while all subsequent instances should be spawned using the --proc-type=secondary flag.

The simple_mp and symmetric_mp sample applications demonstrate this usage model. They are described in the “Multi-process Sample Application” chapter in the DPDK Sample Application’s User Guide.

Asymmetric/Non-Peer Processes

An alternative deployment model that can be used for multi-process applications is to have a single primary process instance that acts as a load-balancer or server distributing received packets among worker or client threads, which are run as secondary processes. In this case, extensive use of rte_ring objects is made, which are located in shared hugepage memory.

The client_server_mp sample application shows this usage model. It is described in the “Multi-process Sample Application” chapter in the DPDK Sample Application’s User Guide.

Running Multiple Independent DPDK Applications

In addition to the above scenarios involving multiple DPDK processes working together, it is possible to run multiple DPDK processes side-by-side, where those processes are all working independently. Support for this usage scenario is provided using the --file-prefix parameter to the EAL.

By default, the EAL creates hugepage files on each hugetlbfs filesystem using the rtemap_X filename, where X is in the range 0 to the maximum number of hugepages -1. Similarly, it creates shared configuration files, memory mapped in each process, using the /var/run/.rte_config filename, when run as root (or $HOME/.rte_config when run as a non-root user; if filesystem and device permissions are set up to allow this). The rte part of the filenames of each of the above is configurable using the file-prefix parameter.

In addition to specifying the file-prefix parameter, any DPDK applications that are to be run side-by-side must explicitly limit their memory use. This is done by passing the -m flag to each process to specify how much hugepage memory, in megabytes, each process can use (or passing --socket-mem to specify how much hugepage memory on each socket each process can use).

Note

Independent DPDK instances running side-by-side on a single machine cannot share any network ports. Any network ports being used by one process should be blacklisted in every other process.

Running Multiple Independent Groups of DPDK Applications

In the same way that it is possible to run independent DPDK applications side- by-side on a single system, this can be trivially extended to multi-process groups of DPDK applications running side-by-side. In this case, the secondary processes must use the same --file-prefix parameter as the primary process whose shared memory they are connecting to.

Note

All restrictions and issues with multiple independent DPDK processes running side-by-side apply in this usage scenario also.

Multi-process Limitations

There are a number of limitations to what can be done when running DPDK multi-process applications. Some of these are documented below:

  • The multi-process feature requires that the exact same hugepage memory mappings be present in all applications. The Linux security feature - Address-Space Layout Randomization (ASLR) can interfere with this mapping, so it may be necessary to disable this feature in order to reliably run multi-process applications.

Warning

Disabling Address-Space Layout Randomization (ASLR) may have security implications, so it is recommended that it be disabled only when absolutely necessary, and only when the implications of this change have been understood.

  • All DPDK processes running as a single application and using shared memory must have distinct coremask/corelist arguments. It is not possible to have a primary and secondary instance, or two secondary instances, using any of the same logical cores. Attempting to do so can cause corruption of memory pool caches, among other issues.
  • The delivery of interrupts, such as Ethernet* device link status interrupts, do not work in secondary processes. All interrupts are triggered inside the primary process only. Any application needing interrupt notification in multiple processes should provide its own mechanism to transfer the interrupt information from the primary process to any secondary process that needs the information.
  • The use of function pointers between multiple processes running based of different compiled binaries is not supported, since the location of a given function in one process may be different to its location in a second. This prevents the librte_hash library from behaving properly as in a multi-threaded instance, since it uses a pointer to the hash function internally.

To work around this issue, it is recommended that multi-process applications perform the hash calculations by directly calling the hashing function from the code and then using the rte_hash_add_with_hash()/rte_hash_lookup_with_hash() functions instead of the functions which do the hashing internally, such as rte_hash_add()/rte_hash_lookup().

  • Depending upon the hardware in use, and the number of DPDK processes used, it may not be possible to have HPET timers available in each DPDK instance. The minimum number of HPET comparators available to Linux* userspace can be just a single comparator, which means that only the first, primary DPDK process instance can open and mmap /dev/hpet. If the number of required DPDK processes exceeds that of the number of available HPET comparators, the TSC (which is the default timer in this release) must be used as a time source across all processes instead of the HPET.

Kernel NIC Interface

The DPDK Kernel NIC Interface (KNI) allows userspace applications access to the Linux* control plane.

The benefits of using the DPDK KNI are:

  • Faster than existing Linux TUN/TAP interfaces (by eliminating system calls and copy_to_user()/copy_from_user() operations.
  • Allows management of DPDK ports using standard Linux net tools such as ethtool, ifconfig and tcpdump.
  • Allows an interface with the kernel network stack.

The components of an application using the DPDK Kernel NIC Interface are shown in Fig. 29.1.

_images/kernel_nic_intf.png

Components of a DPDK KNI Application

The DPDK KNI Kernel Module

The KNI kernel loadable module provides support for two types of devices:

  • A Miscellaneous device (/dev/kni) that:
    • Creates net devices (via ioctl calls).
    • Maintains a kernel thread context shared by all KNI instances (simulating the RX side of the net driver).
    • For single kernel thread mode, maintains a kernel thread context shared by all KNI instances (simulating the RX side of the net driver).
    • For multiple kernel thread mode, maintains a kernel thread context for each KNI instance (simulating the RX side of the net driver).
  • Net device:
    • Net functionality provided by implementing several operations such as netdev_ops, header_ops, ethtool_ops that are defined by struct net_device, including support for DPDK mbufs and FIFOs.
    • The interface name is provided from userspace.
    • The MAC address can be the real NIC MAC address or random.

KNI Creation and Deletion

The KNI interfaces are created by a DPDK application dynamically. The interface name and FIFO details are provided by the application through an ioctl call using the rte_kni_device_info struct which contains:

  • The interface name.
  • Physical addresses of the corresponding memzones for the relevant FIFOs.
  • Mbuf mempool details, both physical and virtual (to calculate the offset for mbuf pointers).
  • PCI information.
  • Core affinity.

Refer to rte_kni_common.h in the DPDK source code for more details.

The physical addresses will be re-mapped into the kernel address space and stored in separate KNI contexts.

The affinity of kernel RX thread (both single and multi-threaded modes) is controlled by force_bind and core_id config parameters.

The KNI interfaces can be deleted by a DPDK application dynamically after being created. Furthermore, all those KNI interfaces not deleted will be deleted on the release operation of the miscellaneous device (when the DPDK application is closed).

DPDK mbuf Flow

To minimize the amount of DPDK code running in kernel space, the mbuf mempool is managed in userspace only. The kernel module will be aware of mbufs, but all mbuf allocation and free operations will be handled by the DPDK application only.

Fig. 29.2 shows a typical scenario with packets sent in both directions.

_images/pkt_flow_kni.png

Packet Flow via mbufs in the DPDK KNI

Use Case: Ingress

On the DPDK RX side, the mbuf is allocated by the PMD in the RX thread context. This thread will enqueue the mbuf in the rx_q FIFO. The KNI thread will poll all KNI active devices for the rx_q. If an mbuf is dequeued, it will be converted to a sk_buff and sent to the net stack via netif_rx(). The dequeued mbuf must be freed, so the same pointer is sent back in the free_q FIFO.

The RX thread, in the same main loop, polls this FIFO and frees the mbuf after dequeuing it.

Use Case: Egress

For packet egress the DPDK application must first enqueue several mbufs to create an mbuf cache on the kernel side.

The packet is received from the Linux net stack, by calling the kni_net_tx() callback. The mbuf is dequeued (without waiting due the cache) and filled with data from sk_buff. The sk_buff is then freed and the mbuf sent in the tx_q FIFO.

The DPDK TX thread dequeues the mbuf and sends it to the PMD (via rte_eth_tx_burst()). It then puts the mbuf back in the cache.

Ethtool

Ethtool is a Linux-specific tool with corresponding support in the kernel where each net device must register its own callbacks for the supported operations. The current implementation uses the igb/ixgbe modified Linux drivers for ethtool support. Ethtool is not supported in i40e and VMs (VF or EM devices).

Thread Safety of DPDK Functions

The DPDK is comprised of several libraries. Some of the functions in these libraries can be safely called from multiple threads simultaneously, while others cannot. This section allows the developer to take these issues into account when building their own application.

The run-time environment of the DPDK is typically a single thread per logical core. In some cases, it is not only multi-threaded, but multi-process. Typically, it is best to avoid sharing data structures between threads and/or processes where possible. Where this is not possible, then the execution blocks must access the data in a thread- safe manner. Mechanisms such as atomics or locking can be used that will allow execution blocks to operate serially. However, this can have an effect on the performance of the application.

Fast-Path APIs

Applications operating in the data plane are performance sensitive but certain functions within those libraries may not be safe to call from multiple threads simultaneously. The hash, LPM and mempool libraries and RX/TX in the PMD are examples of this.

The hash and LPM libraries are, by design, thread unsafe in order to maintain performance. However, if required the developer can add layers on top of these libraries to provide thread safety. Locking is not needed in all situations, and in both the hash and LPM libraries, lookups of values can be performed in parallel in multiple threads. Adding, removing or modifying values, however, cannot be done in multiple threads without using locking when a single hash or LPM table is accessed. Another alternative to locking would be to create multiple instances of these tables allowing each thread its own copy.

The RX and TX of the PMD are the most critical aspects of a DPDK application and it is recommended that no locking be used as it will impact performance. Note, however, that these functions can safely be used from multiple threads when each thread is performing I/O on a different NIC queue. If multiple threads are to use the same hardware queue on the same NIC port, then locking, or some other form of mutual exclusion, is necessary.

The ring library is based on a lockless ring-buffer algorithm that maintains its original design for thread safety. Moreover, it provides high performance for either multi- or single-consumer/producer enqueue/dequeue operations. The mempool library is based on the DPDK lockless ring library and therefore is also multi-thread safe.

Performance Insensitive API

Outside of the performance sensitive areas described in Section 25.1, the DPDK provides a thread-safe API for most other libraries. For example, malloc and memzone functions are safe for use in multi-threaded and multi-process environments.

The setup and configuration of the PMD is not performance sensitive, but is not thread safe either. It is possible that the multiple read/writes during PMD setup and configuration could be corrupted in a multi-thread environment. Since this is not performance sensitive, the developer can choose to add their own layer to provide thread-safe setup and configuration. It is expected that, in most applications, the initial configuration of the network ports would be done by a single thread at startup.

Library Initialization

It is recommended that DPDK libraries are initialized in the main thread at application startup rather than subsequently in the forwarding threads. However, the DPDK performs checks to ensure that libraries are only initialized once. If initialization is attempted more than once, an error is returned.

In the multi-process case, the configuration information of shared memory will only be initialized by the master process. Thereafter, both master and secondary processes can allocate/release any objects of memory that finally rely on rte_malloc or memzones.

Interrupt Thread

The DPDK works almost entirely in Linux user space in polling mode. For certain infrequent operations, such as receiving a PMD link status change notification, callbacks may be called in an additional thread outside the main DPDK processing threads. These function callbacks should avoid manipulating DPDK objects that are also managed by the normal DPDK threads, and if they need to do so, it is up to the application to provide the appropriate locking or mutual exclusion restrictions around those objects.

Event Device Library

The DPDK Event device library is an abstraction that provides the application with features to schedule events. This is achieved using the PMD architecture similar to the ethdev or cryptodev APIs, which may already be familiar to the reader.

The eventdev framework introduces the event driven programming model. In a polling model, lcores poll ethdev ports and associated Rx queues directly to look for a packet. By contrast in an event driven model, lcores call the scheduler that selects packets for them based on programmer-specified criteria. The Eventdev library adds support for an event driven programming model, which offers applications automatic multicore scaling, dynamic load balancing, pipelining, packet ingress order maintenance and synchronization services to simplify application packet processing.

By introducing an event driven programming model, DPDK can support both polling and event driven programming models for packet processing, and applications are free to choose whatever model (or combination of the two) best suits their needs.

Step-by-step instructions of the eventdev design is available in the API Walk-through section later in this document.

Event struct

The eventdev API represents each event with a generic struct, which contains a payload and metadata required for scheduling by an eventdev. The rte_event struct is a 16 byte C structure, defined in libs/librte_eventdev/rte_eventdev.h.

Event Metadata

The rte_event structure contains the following metadata fields, which the application fills in to have the event scheduled as required:

  • flow_id - The targeted flow identifier for the enq/deq operation.
  • event_type - The source of this event, eg RTE_EVENT_TYPE_ETHDEV or CPU.
  • sub_event_type - Distinguishes events inside the application, that have the same event_type (see above)
  • op - This field takes one of the RTE_EVENT_OP_* values, and tells the eventdev about the status of the event - valid values are NEW, FORWARD or RELEASE.
  • sched_type - Represents the type of scheduling that should be performed on this event, valid values are the RTE_SCHED_TYPE_ORDERED, ATOMIC and PARALLEL.
  • queue_id - The identifier for the event queue that the event is sent to.
  • priority - The priority of this event, see RTE_EVENT_DEV_PRIORITY.
Event Payload

The rte_event struct contains a union for payload, allowing flexibility in what the actual event being scheduled is. The payload is a union of the following:

  • uint64_t u64
  • void *event_ptr
  • struct rte_mbuf *mbuf

These three items in a union occupy the same 64 bits at the end of the rte_event structure. The application can utilize the 64 bits directly by accessing the u64 variable, while the event_ptr and mbuf are provided as convenience variables. For example the mbuf pointer in the union can used to schedule a DPDK packet.

Queues

An event queue is a queue containing events that are scheduled by the event device. An event queue contains events of different flows associated with scheduling types, such as atomic, ordered, or parallel.

Queue All Types Capable

If RTE_EVENT_DEV_CAP_QUEUE_ALL_TYPES capability bit is set in the event device, then events of any type may be sent to any queue. Otherwise, the queues only support events of the type that it was created with.

Queue All Types Incapable

In this case, each stage has a specified scheduling type. The application configures each queue for a specific type of scheduling, and just enqueues all events to the eventdev. An example of a PMD of this type is the eventdev software PMD.

The Eventdev API supports the following scheduling types per queue:

  • Atomic
  • Ordered
  • Parallel

Atomic, Ordered and Parallel are load-balanced scheduling types: the output of the queue can be spread out over multiple CPU cores.

Atomic scheduling on a queue ensures that a single flow is not present on two different CPU cores at the same time. Ordered allows sending all flows to any core, but the scheduler must ensure that on egress the packets are returned to ingress order on downstream queue enqueue. Parallel allows sending all flows to all CPU cores, without any re-ordering guarantees.

Ports

Ports are the points of contact between worker cores and the eventdev. The general use-case will see one CPU core using one port to enqueue and dequeue events from an eventdev. Ports are linked to queues in order to retrieve events from those queues (more details in Linking Queues and Ports below).

API Walk-through

This section will introduce the reader to the eventdev API, showing how to create and configure an eventdev and use it for a two-stage atomic pipeline with a single core for TX. The diagram below shows the final state of the application after this walk-through:

_images/eventdev_usage.svg

Sample eventdev usage, with RX, two atomic stages and a single-link to TX.

A high level overview of the setup steps are:

  • rte_event_dev_configure()
  • rte_event_queue_setup()
  • rte_event_port_setup()
  • rte_event_port_link()
  • rte_event_dev_start()
Init and Config

The eventdev library uses vdev options to add devices to the DPDK application. The --vdev EAL option allows adding eventdev instances to your DPDK application, using the name of the eventdev PMD as an argument.

For example, to create an instance of the software eventdev scheduler, the following vdev arguments should be provided to the application EAL command line:

./dpdk_application --vdev="event_sw0"

In the following code, we configure eventdev instance with 3 queues and 6 ports as follows. The 3 queues consist of 2 Atomic and 1 Single-Link, while the 6 ports consist of 4 workers, 1 RX and 1 TX.

const struct rte_event_dev_config config = {
        .nb_event_queues = 3,
        .nb_event_ports = 6,
        .nb_events_limit  = 4096,
        .nb_event_queue_flows = 1024,
        .nb_event_port_dequeue_depth = 128,
        .nb_event_port_enqueue_depth = 128,
};
int err = rte_event_dev_configure(dev_id, &config);

The remainder of this walk-through assumes that dev_id is 0.

Setting up Queues

Once the eventdev itself is configured, the next step is to configure queues. This is done by setting the appropriate values in a queue_conf structure, and calling the setup function. Repeat this step for each queue, starting from 0 and ending at nb_event_queues - 1 from the event_dev config above.

struct rte_event_queue_conf atomic_conf = {
        .schedule_type = RTE_SCHED_TYPE_ATOMIC,
        .priority = RTE_EVENT_DEV_PRIORITY_NORMAL,
        .nb_atomic_flows = 1024,
        .nb_atomic_order_sequences = 1024,
};
int dev_id = 0;
int queue_id = 0;
int err = rte_event_queue_setup(dev_id, queue_id, &atomic_conf);

The remainder of this walk-through assumes that the queues are configured as follows:

  • id 0, atomic queue #1
  • id 1, atomic queue #2
  • id 2, single-link queue
Setting up Ports

Once queues are set up successfully, create the ports as required. Each port should be set up with its corresponding port_conf type, worker for worker cores, rx and tx for the RX and TX cores:

struct rte_event_port_conf rx_conf = {
        .dequeue_depth = 128,
        .enqueue_depth = 128,
        .new_event_threshold = 1024,
};
struct rte_event_port_conf worker_conf = {
        .dequeue_depth = 16,
        .enqueue_depth = 64,
        .new_event_threshold = 4096,
};
struct rte_event_port_conf tx_conf = {
        .dequeue_depth = 128,
        .enqueue_depth = 128,
        .new_event_threshold = 4096,
};
int dev_id = 0;
int port_id = 0;
int err = rte_event_port_setup(dev_id, port_id, &CORE_FUNCTION_conf);

It is now assumed that:

  • port 0: RX core
  • ports 1,2,3,4: Workers
  • port 5: TX core
Linking Queues and Ports

The final step is to “wire up” the ports to the queues. After this, the eventdev is capable of scheduling events, and when cores request work to do, the correct events are provided to that core. Note that the RX core takes input from eg: a NIC so it is not linked to any eventdev queues.

Linking all workers to atomic queues, and the TX core to the single-link queue can be achieved like this:

uint8_t port_id = 0;
uint8_t atomic_qs[] = {0, 1};
uint8_t single_link_q = 2;
uint8_t tx_port_id = 5;
uin8t_t priority = RTE_EVENT_DEV_PRIORITY_NORMAL;

for(int i = 0; i < 4; i++) {
        int worker_port = i + 1;
        int links_made = rte_event_port_link(dev_id, worker_port, atomic_qs, NULL, 2);
}
int links_made = rte_event_port_link(dev_id, tx_port_id, &single_link_q, &priority, 1);
Starting the EventDev

A single function call tells the eventdev instance to start processing events. Note that all queues must be linked to for the instance to start, as if any queue is not linked to, enqueuing to that queue will cause the application to backpressure and eventually stall due to no space in the eventdev.

int err = rte_event_dev_start(dev_id);
Ingress of New Events

Now that the eventdev is set up, and ready to receive events, the RX core must enqueue some events into the system for it to schedule. The events to be scheduled are ordinary DPDK packets, received from an eth_rx_burst() as normal. The following code shows how those packets can be enqueued into the eventdev:

const uint16_t nb_rx = rte_eth_rx_burst(eth_port, 0, mbufs, BATCH_SIZE);

for (i = 0; i < nb_rx; i++) {
        ev[i].flow_id = mbufs[i]->hash.rss;
        ev[i].op = RTE_EVENT_OP_NEW;
        ev[i].sched_type = RTE_SCHED_TYPE_ATOMIC;
        ev[i].queue_id = 0;
        ev[i].event_type = RTE_EVENT_TYPE_ETHDEV;
        ev[i].sub_event_type = 0;
        ev[i].priority = RTE_EVENT_DEV_PRIORITY_NORMAL;
        ev[i].mbuf = mbufs[i];
}

const int nb_tx = rte_event_enqueue_burst(dev_id, port_id, ev, nb_rx);
if (nb_tx != nb_rx) {
        for(i = nb_tx; i < nb_rx; i++)
                rte_pktmbuf_free(mbufs[i]);
}
Forwarding of Events

Now that the RX core has injected events, there is work to be done by the workers. Note that each worker will dequeue as many events as it can in a burst, process each one individually, and then burst the packets back into the eventdev.

The worker can lookup the events source from event.queue_id, which should indicate to the worker what workload needs to be performed on the event. Once done, the worker can update the event.queue_id to a new value, to send the event to the next stage in the pipeline.

int timeout = 0;
struct rte_event events[BATCH_SIZE];
uint16_t nb_rx = rte_event_dequeue_burst(dev_id, worker_port_id, events, BATCH_SIZE, timeout);

for (i = 0; i < nb_rx; i++) {
        /* process mbuf using events[i].queue_id as pipeline stage */
        struct rte_mbuf *mbuf = events[i].mbuf;
        /* Send event to next stage in pipeline */
        events[i].queue_id++;
}

uint16_t nb_tx = rte_event_enqueue_burst(dev_id, port_id, events, nb_rx);
Egress of Events

Finally, when the packet is ready for egress or needs to be dropped, we need to inform the eventdev that the packet is no longer being handled by the application. This can be done by calling dequeue() or dequeue_burst(), which indicates that the previous burst of packets is no longer in use by the application.

An event driven worker thread has following typical workflow on fastpath:

while (1) {
        rte_event_dequeue_burst(...);
        (event processing)
        rte_event_enqueue_burst(...);
}

Summary

The eventdev library allows an application to easily schedule events as it requires, either using a run-to-completion or pipeline processing model. The queues and ports abstract the logical functionality of an eventdev, providing the application with a generic method to schedule events. With the flexible PMD infrastructure applications benefit of improvements in existing eventdevs and additions of new ones without modification.

Event Ethernet Rx Adapter Library

The DPDK Eventdev API allows the application to use an event driven programming model for packet processing. In this model, the application polls an event device port for receiving events that reference packets instead of polling Rx queues of ethdev ports. Packet transfer between ethdev and the event device can be supported in hardware or require a software thread to receive packets from the ethdev port using ethdev poll mode APIs and enqueue these as events to the event device using the eventdev API. Both transfer mechanisms may be present on the same platform depending on the particular combination of the ethdev and the event device.

The Event Ethernet Rx Adapter library is intended for the application code to configure both transfer mechanisms using a common API. A capability API allows the eventdev PMD to advertise features supported for a given ethdev and allows the application to perform configuration as per supported features.

API Walk-through

This section will introduce the reader to the adapter API. The application has to first instantiate an adapter which is associated with a single eventdev, next the adapter instance is configured with Rx queues that are either polled by a SW thread or linked using hardware support. Finally the adapter is started.

For SW based packet transfers from ethdev to eventdev, the adapter uses a DPDK service function and the application is also required to assign a core to the service function.

Creating an Adapter Instance

An adapter instance is created using rte_event_eth_rx_adapter_create(). This function is passed the event device to be associated with the adapter and port configuration for the adapter to setup an event port if the adapter needs to use a service function.

int err;
uint8_t dev_id;
struct rte_event_dev_info dev_info;
struct rte_event_port_conf rx_p_conf;

err = rte_event_dev_info_get(id, &dev_info);

rx_p_conf.new_event_threshold = dev_info.max_num_events;
rx_p_conf.dequeue_depth = dev_info.max_event_port_dequeue_depth;
rx_p_conf.enqueue_depth = dev_info.max_event_port_enqueue_depth;
err = rte_event_eth_rx_adapter_create(id, dev_id, &rx_p_conf);

If the application desires to have finer control of eventdev port allocation and setup, it can use the rte_event_eth_rx_adapter_create_ext() function. The rte_event_eth_rx_adapter_create_ext() function is passed a callback function. The callback function is invoked if the adapter needs to use a service function and needs to create an event port for it. The callback is expected to fill the struct rte_event_eth_rx_adapter_conf structure passed to it.

Adding Rx Queues to the Adapter Instance

Ethdev Rx queues are added to the instance using the rte_event_eth_rx_adapter_queue_add() function. Configuration for the Rx queue is passed in using a struct rte_event_eth_rx_adapter_queue_conf parameter. Event information for packets from this Rx queue is encoded in the ev field of struct rte_event_eth_rx_adapter_queue_conf. The servicing_weight member of the struct rte_event_eth_rx_adapter_queue_conf is the relative polling frequency of the Rx queue and is applicable when the adapter uses a service core function.

ev.queue_id = 0;
ev.sched_type = RTE_SCHED_TYPE_ATOMIC;
ev.priority = 0;

queue_config.rx_queue_flags = 0;
queue_config.ev = ev;
queue_config.servicing_weight = 1;

err = rte_event_eth_rx_adapter_queue_add(id,
                                        eth_dev_id,
                                        0, &queue_config);
Querying Adapter Capabilities

The rte_event_eth_rx_adapter_caps_get() function allows the application to query the adapter capabilities for an eventdev and ethdev combination. For e.g, if the RTE_EVENT_ETH_RX_ADAPTER_CAP_OVERRIDE_FLOW_ID is set, the application can override the adapter generated flow ID in the event using rx_queue_flags field in struct rte_event_eth_rx_adapter_queue_conf which is passed as a parameter to the rte_event_eth_rx_adapter_queue_add() function.

err = rte_event_eth_rx_adapter_caps_get(dev_id, eth_dev_id, &cap);

queue_config.rx_queue_flags = 0;
if (cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_OVERRIDE_FLOW_ID) {
        ev.flow_id = 1;
        queue_config.rx_queue_flags =
                RTE_EVENT_ETH_RX_ADAPTER_QUEUE_FLOW_ID_VALID;
}
Configuring the Service Function

If the adapter uses a service function, the application is required to assign a service core to the service function as show below.

uint32_t service_id;

if (rte_event_eth_rx_adapter_service_id_get(0, &service_id) == 0)
        rte_service_map_lcore_set(service_id, RX_CORE_ID);
Starting the Adapter Instance

The application calls rte_event_eth_rx_adapter_start() to start the adapter. This function calls the start callbacks of the eventdev PMDs for hardware based eventdev-ethdev connections and rte_service_run_state_set() to enable the service function if one exists.

Getting Adapter Statistics

The rte_event_eth_rx_adapter_stats_get() function reports counters defined in struct rte_event_eth_rx_adapter_stats. The received packet and enqueued event counts are a sum of the counts from the eventdev PMD callbacks if the callback is supported, and the counts maintained by the service function, if one exists. The service function also maintains a count of cycles for which it was not able to enqueue to the event device.

Quality of Service (QoS) Framework

This chapter describes the DPDK Quality of Service (QoS) framework.

Packet Pipeline with QoS Support

An example of a complex packet processing pipeline with QoS support is shown in the following figure.

_images/pkt_proc_pipeline_qos.png

Complex Packet Processing Pipeline with QoS Support

This pipeline can be built using reusable DPDK software libraries. The main blocks implementing QoS in this pipeline are: the policer, the dropper and the scheduler. A functional description of each block is provided in the following table.

Packet Processing Pipeline Implementing QoS
# Block Functional Description
1 Packet I/O RX & TX Packet reception/ transmission from/to multiple NIC ports. Poll mode drivers (PMDs) for Intel 1 GbE/10 GbE NICs.
2 Packet parser Identify the protocol stack of the input packet. Check the integrity of the packet headers.
3 Flow classification Map the input packet to one of the known traffic flows. Exact match table lookup using configurable hash function (jhash, CRC and so on) and bucket logic to handle collisions.
4 Policer Packet metering using srTCM (RFC 2697) or trTCM (RFC2698) algorithms.
5 Load Balancer Distribute the input packets to the application workers. Provide uniform load to each worker. Preserve the affinity of traffic flows to workers and the packet order within each flow.
6 Worker threads Placeholders for the customer specific application workload (for example, IP stack and so on).
7 Dropper Congestion management using the Random Early Detection (RED) algorithm (specified by the Sally Floyd - Van Jacobson paper) or Weighted RED (WRED). Drop packets based on the current scheduler queue load level and packet priority. When congestion is experienced, lower priority packets are dropped first.
8 Hierarchical Scheduler 5-level hierarchical scheduler (levels are: output port, subport, pipe, traffic class and queue) with thousands (typically 64K) leaf nodes (queues). Implements traffic shaping (for subport and pipe levels), strict priority (for traffic class level) and Weighted Round Robin (WRR) (for queues within each pipe traffic class).

The infrastructure blocks used throughout the packet processing pipeline are listed in the following table.

Infrastructure Blocks Used by the Packet Processing Pipeline
# Block Functional Description
1 Buffer manager Support for global buffer pools and private per-thread buffer caches.
2 Queue manager Support for message passing between pipeline blocks.
3 Power saving Support for power saving during low activity periods.

The mapping of pipeline blocks to CPU cores is configurable based on the performance level required by each specific application and the set of features enabled for each block. Some blocks might consume more than one CPU core (with each CPU core running a different instance of the same block on different input packets), while several other blocks could be mapped to the same CPU core.

Hierarchical Scheduler

The hierarchical scheduler block, when present, usually sits on the TX side just before the transmission stage. Its purpose is to prioritize the transmission of packets from different users and different traffic classes according to the policy specified by the Service Level Agreements (SLAs) of each network node.

Overview

The hierarchical scheduler block is similar to the traffic manager block used by network processors that typically implement per flow (or per group of flows) packet queuing and scheduling. It typically acts like a buffer that is able to temporarily store a large number of packets just before their transmission (enqueue operation); as the NIC TX is requesting more packets for transmission, these packets are later on removed and handed over to the NIC TX with the packet selection logic observing the predefined SLAs (dequeue operation).

_images/hier_sched_blk.png

Hierarchical Scheduler Block Internal Diagram

The hierarchical scheduler is optimized for a large number of packet queues. When only a small number of queues are needed, message passing queues should be used instead of this block. See Worst Case Scenarios for Performance for a more detailed discussion.

Scheduling Hierarchy

The scheduling hierarchy is shown in Fig. 33.3. The first level of the hierarchy is the Ethernet TX port 1/10/40 GbE, with subsequent hierarchy levels defined as subport, pipe, traffic class and queue.

Typically, each subport represents a predefined group of users, while each pipe represents an individual user/subscriber. Each traffic class is the representation of a different traffic type with specific loss rate, delay and jitter requirements, such as voice, video or data transfers. Each queue hosts packets from one or multiple connections of the same type belonging to the same user.

_images/sched_hier_per_port.png

Scheduling Hierarchy per Port

The functionality of each hierarchical level is detailed in the following table.

Port Scheduling Hierarchy
# Level Siblings per Parent Functional Description
1 Port
  1. Output Ethernet port 1/10/40 GbE.
  2. Multiple ports are scheduled in round robin order with all ports having equal priority.
2 Subport Configurable (default: 8)
  1. Traffic shaping using token bucket algorithm (one token bucket per subport).
  2. Upper limit enforced per Traffic Class (TC) at the subport level.
  3. Lower priority TCs able to reuse subport bandwidth currently unused by higher priority TCs.
3 Pipe Configurable (default: 4K)
  1. Traffic shaping using the token bucket algorithm (one token bucket per pipe.
4 Traffic Class (TC) 4
  1. TCs of the same pipe handled in strict priority order.
  2. Upper limit enforced per TC at the pipe level.
  3. Lower priority TCs able to reuse pipe bandwidth currently unused by higher priority TCs.
  4. When subport TC is oversubscribed (configuration time event), pipe TC upper limit is capped to a dynamically adjusted value that is shared by all the subport pipes.
5 Queue 4
  1. Queues of the same TC are serviced using Weighted Round Robin (WRR) according to predefined weights.
Application Programming Interface (API)
Port Scheduler Configuration API

The rte_sched.h file contains configuration functions for port, subport and pipe.

Port Scheduler Enqueue API

The port scheduler enqueue API is very similar to the API of the DPDK PMD TX function.

int rte_sched_port_enqueue(struct rte_sched_port *port, struct rte_mbuf **pkts, uint32_t n_pkts);
Port Scheduler Dequeue API

The port scheduler dequeue API is very similar to the API of the DPDK PMD RX function.

int rte_sched_port_dequeue(struct rte_sched_port *port, struct rte_mbuf **pkts, uint32_t n_pkts);
Usage Example
/* File "application.c" */

#define N_PKTS_RX   64
#define N_PKTS_TX   48
#define NIC_RX_PORT 0
#define NIC_RX_QUEUE 0
#define NIC_TX_PORT 1
#define NIC_TX_QUEUE 0

struct rte_sched_port *port = NULL;
struct rte_mbuf *pkts_rx[N_PKTS_RX], *pkts_tx[N_PKTS_TX];
uint32_t n_pkts_rx, n_pkts_tx;

/* Initialization */

<initialization code>

/* Runtime */
while (1) {
    /* Read packets from NIC RX queue */

    n_pkts_rx = rte_eth_rx_burst(NIC_RX_PORT, NIC_RX_QUEUE, pkts_rx, N_PKTS_RX);

    /* Hierarchical scheduler enqueue */

    rte_sched_port_enqueue(port, pkts_rx, n_pkts_rx);

    /* Hierarchical scheduler dequeue */

    n_pkts_tx = rte_sched_port_dequeue(port, pkts_tx, N_PKTS_TX);

    /* Write packets to NIC TX queue */

    rte_eth_tx_burst(NIC_TX_PORT, NIC_TX_QUEUE, pkts_tx, n_pkts_tx);
}
Implementation
Internal Data Structures per Port

A schematic of the internal data structures in shown in with details in.

_images/data_struct_per_port.png

Internal Data Structures per Port

Scheduler Internal Data Structures per Port
# Data structure Size (bytes) # per port Access type Description
Enq Deq  
1 Subport table entry 64 # subports per port
Rd, Wr Persistent subport data (credits, etc).
2 Pipe table entry 64 # pipes per port
Rd, Wr

Persistent data for pipe, its TCs and its queues (credits, etc) that is updated during run-time.

The pipe configuration parameters do not change during run-time. The same pipe configuration parameters are shared by multiple pipes, therefore they are not part of pipe table entry.

3 Queue table entry 4 #queues per port Rd, Wr Rd, Wr

Persistent queue data (read and write pointers). The queue size is the same per TC for all queues, allowing the queue base address to be computed using a fast formula, so these two parameters are not part of queue table entry.

The queue table entries for any given pipe are stored in the same cache line.

4 Queue storage area Config (default: 64 x8) # queues per port Wr Rd Array of elements per queue; each element is 8 byte in size (mbuf pointer).
5 Active queues bitmap 1 bit per queue 1 Wr (Set) Rd, Wr (Clear)

The bitmap maintains one status bit per queue: queue not active (queue is empty) or queue active (queue is not empty).

Queue bit is set by the scheduler enqueue and cleared by the scheduler dequeue when queue becomes empty.

Bitmap scan operation returns the next non-empty pipe and its status (16-bit mask of active queue in the pipe).

6 Grinder ~128 Config (default: 8)
Rd, Wr

Short list of active pipes currently under processing. The grinder contains temporary data during pipe processing.

Once the current pipe exhausts packets or credits, it is replaced with another active pipe from the bitmap.

Multicore Scaling Strategy

The multicore scaling strategy is:

  1. Running different physical ports on different threads. The enqueue and dequeue of the same port are run by the same thread.
  2. Splitting the same physical port to different threads by running different sets of subports of the same physical port (virtual ports) on different threads. Similarly, a subport can be split into multiple subports that are each run by a different thread. The enqueue and dequeue of the same port are run by the same thread. This is only required if, for performance reasons, it is not possible to handle a full port with a single core.
Enqueue and Dequeue for the Same Output Port

Running enqueue and dequeue operations for the same output port from different cores is likely to cause significant impact on scheduler’s performance and it is therefore not recommended.

The port enqueue and dequeue operations share access to the following data structures:

  1. Packet descriptors
  2. Queue table
  3. Queue storage area
  4. Bitmap of active queues

The expected drop in performance is due to:

  1. Need to make the queue and bitmap operations thread safe, which requires either using locking primitives for access serialization (for example, spinlocks/ semaphores) or using atomic primitives for lockless access (for example, Test and Set, Compare And Swap, an so on). The impact is much higher in the former case.
  2. Ping-pong of cache lines storing the shared data structures between the cache hierarchies of the two cores (done transparently by the MESI protocol cache coherency CPU hardware).

Therefore, the scheduler enqueue and dequeue operations have to be run from the same thread, which allows the queues and the bitmap operations to be non-thread safe and keeps the scheduler data structures internal to the same core.

Performance Scaling

Scaling up the number of NIC ports simply requires a proportional increase in the number of CPU cores to be used for traffic scheduling.

Enqueue Pipeline

The sequence of steps per packet:

  1. Access the mbuf to read the data fields required to identify the destination queue for the packet. These fields are: port, subport, traffic class and queue within traffic class, and are typically set by the classification stage.
  2. Access the queue structure to identify the write location in the queue array. If the queue is full, then the packet is discarded.
  3. Access the queue array location to store the packet (i.e. write the mbuf pointer).

It should be noted the strong data dependency between these steps, as steps 2 and 3 cannot start before the result from steps 1 and 2 becomes available, which prevents the processor out of order execution engine to provide any significant performance optimizations.

Given the high rate of input packets and the large amount of queues, it is expected that the data structures accessed to enqueue the current packet are not present in the L1 or L2 data cache of the current core, thus the above 3 memory accesses would result (on average) in L1 and L2 data cache misses. A number of 3 L1/L2 cache misses per packet is not acceptable for performance reasons.

The workaround is to prefetch the required data structures in advance. The prefetch operation has an execution latency during which the processor should not attempt to access the data structure currently under prefetch, so the processor should execute other work. The only other work available is to execute different stages of the enqueue sequence of operations on other input packets, thus resulting in a pipelined implementation for the enqueue operation.

Fig. 33.5 illustrates a pipelined implementation for the enqueue operation with 4 pipeline stages and each stage executing 2 different input packets. No input packet can be part of more than one pipeline stage at a given time.

_images/prefetch_pipeline.png

Prefetch Pipeline for the Hierarchical Scheduler Enqueue Operation

The congestion management scheme implemented by the enqueue pipeline described above is very basic: packets are enqueued until a specific queue becomes full, then all the packets destined to the same queue are dropped until packets are consumed (by the dequeue operation). This can be improved by enabling RED/WRED as part of the enqueue pipeline which looks at the queue occupancy and packet priority in order to yield the enqueue/drop decision for a specific packet (as opposed to enqueuing all packets / dropping all packets indiscriminately).

Dequeue State Machine

The sequence of steps to schedule the next packet from the current pipe is:

  1. Identify the next active pipe using the bitmap scan operation, prefetch pipe.
  2. Read pipe data structure. Update the credits for the current pipe and its subport. Identify the first active traffic class within the current pipe, select the next queue using WRR, prefetch queue pointers for all the 16 queues of the current pipe.
  3. Read next element from the current WRR queue and prefetch its packet descriptor.
  4. Read the packet length from the packet descriptor (mbuf structure). Based on the packet length and the available credits (of current pipe, pipe traffic class, subport and subport traffic class), take the go/no go scheduling decision for the current packet.

To avoid the cache misses, the above data structures (pipe, queue, queue array, mbufs) are prefetched in advance of being accessed. The strategy of hiding the latency of the prefetch operations is to switch from the current pipe (in grinder A) to another pipe (in grinder B) immediately after a prefetch is issued for the current pipe. This gives enough time to the prefetch operation to complete before the execution switches back to this pipe (in grinder A).

The dequeue pipe state machine exploits the data presence into the processor cache, therefore it tries to send as many packets from the same pipe TC and pipe as possible (up to the available packets and credits) before moving to the next active TC from the same pipe (if any) or to another active pipe.

_images/pipe_prefetch_sm.png

Pipe Prefetch State Machine for the Hierarchical Scheduler Dequeue Operation

Timing and Synchronization

The output port is modeled as a conveyor belt of byte slots that need to be filled by the scheduler with data for transmission. For 10 GbE, there are 1.25 billion byte slots that need to be filled by the port scheduler every second. If the scheduler is not fast enough to fill the slots, provided that enough packets and credits exist, then some slots will be left unused and bandwidth will be wasted.

In principle, the hierarchical scheduler dequeue operation should be triggered by NIC TX. Usually, once the occupancy of the NIC TX input queue drops below a predefined threshold, the port scheduler is woken up (interrupt based or polling based, by continuously monitoring the queue occupancy) to push more packets into the queue.

Internal Time Reference

The scheduler needs to keep track of time advancement for the credit logic, which requires credit updates based on time (for example, subport and pipe traffic shaping, traffic class upper limit enforcement, and so on).

Every time the scheduler decides to send a packet out to the NIC TX for transmission, the scheduler will increment its internal time reference accordingly. Therefore, it is convenient to keep the internal time reference in units of bytes, where a byte signifies the time duration required by the physical interface to send out a byte on the transmission medium. This way, as a packet is scheduled for transmission, the time is incremented with (n + h), where n is the packet length in bytes and h is the number of framing overhead bytes per packet.

Internal Time Reference Re-synchronization

The scheduler needs to align its internal time reference to the pace of the port conveyor belt. The reason is to make sure that the scheduler does not feed the NIC TX with more bytes than the line rate of the physical medium in order to prevent packet drop (by the scheduler, due to the NIC TX input queue being full, or later on, internally by the NIC TX).

The scheduler reads the current time on every dequeue invocation. The CPU time stamp can be obtained by reading either the Time Stamp Counter (TSC) register or the High Precision Event Timer (HPET) register. The current CPU time stamp is converted from number of CPU clocks to number of bytes: time_bytes = time_cycles / cycles_per_byte, where cycles_per_byte is the amount of CPU cycles that is equivalent to the transmission time for one byte on the wire (e.g. for a CPU frequency of 2 GHz and a 10GbE port,*cycles_per_byte = 1.6*).

The scheduler maintains an internal time reference of the NIC time. Whenever a packet is scheduled, the NIC time is incremented with the packet length (including framing overhead). On every dequeue invocation, the scheduler checks its internal reference of the NIC time against the current time:

  1. If NIC time is in the future (NIC time >= current time), no adjustment of NIC time is needed. This means that scheduler is able to schedule NIC packets before the NIC actually needs those packets, so the NIC TX is well supplied with packets;
  2. If NIC time is in the past (NIC time < current time), then NIC time should be adjusted by setting it to the current time. This means that the scheduler is not able to keep up with the speed of the NIC byte conveyor belt, so NIC bandwidth is wasted due to poor packet supply to the NIC TX.
Scheduler Accuracy and Granularity

The scheduler round trip delay (SRTD) is the time (number of CPU cycles) between two consecutive examinations of the same pipe by the scheduler.

To keep up with the output port (that is, avoid bandwidth loss), the scheduler should be able to schedule n packets faster than the same n packets are transmitted by NIC TX.

The scheduler needs to keep up with the rate of each individual pipe, as configured for the pipe token bucket, assuming that no port oversubscription is taking place. This means that the size of the pipe token bucket should be set high enough to prevent it from overflowing due to big SRTD, as this would result in credit loss (and therefore bandwidth loss) for the pipe.

Credit Logic
Scheduling Decision

The scheduling decision to send next packet from (subport S, pipe P, traffic class TC, queue Q) is favorable (packet is sent) when all the conditions below are met:

  • Pipe P of subport S is currently selected by one of the port grinders;
  • Traffic class TC is the highest priority active traffic class of pipe P;
  • Queue Q is the next queue selected by WRR within traffic class TC of pipe P;
  • Subport S has enough credits to send the packet;
  • Subport S has enough credits for traffic class TC to send the packet;
  • Pipe P has enough credits to send the packet;
  • Pipe P has enough credits for traffic class TC to send the packet.

If all the above conditions are met, then the packet is selected for transmission and the necessary credits are subtracted from subport S, subport S traffic class TC, pipe P, pipe P traffic class TC.

Framing Overhead

As the greatest common divisor for all packet lengths is one byte, the unit of credit is selected as one byte. The number of credits required for the transmission of a packet of n bytes is equal to (n+h), where h is equal to the number of framing overhead bytes per packet.

Ethernet Frame Overhead Fields
# Packet field Length (bytes) Comments
1 Preamble 7  
2 Start of Frame Delimiter (SFD) 1  
3 Frame Check Sequence (FCS) 4 Considered overhead only if not included in the mbuf packet length field.
4 Inter Frame Gap (IFG) 12  
5 Total 24  
Traffic Shaping

The traffic shaping for subport and pipe is implemented using a token bucket per subport/per pipe. Each token bucket is implemented using one saturated counter that keeps track of the number of available credits.

The token bucket generic parameters and operations are presented in Table 33.6 and Table 33.7.

Token Bucket Generic Operations
# Token Bucket Parameter Unit Description
1 bucket_rate Credits per second Rate of adding credits to the bucket.
2 bucket_size Credits Max number of credits that can be stored in the bucket.
Token Bucket Generic Parameters
# Token Bucket Operation Description
1 Initialization Bucket set to a predefined value, e.g. zero or half of the bucket size.
2 Credit update Credits are added to the bucket on top of existing ones, either periodically or on demand, based on the bucket_rate. Credits cannot exceed the upper limit defined by the bucket_size, so any credits to be added to the bucket while the bucket is full are dropped.
3 Credit consumption As result of packet scheduling, the necessary number of credits is removed from the bucket. The packet can only be sent if enough credits are in the bucket to send the full packet (packet bytes and framing overhead for the packet).

To implement the token bucket generic operations described above, the current design uses the persistent data structure presented in Table 33.8, while the implementation of the token bucket operations is described in Table 33.9.

Token Bucket Persistent Data Structure
# Token bucket field Unit Description
1 tb_time Bytes

Time of the last credit update. Measured in bytes instead of seconds or CPU cycles for ease of credit consumption operation (as the current time is also maintained in bytes).

See Section 26.2.4.5.1 “Internal Time Reference” for an explanation of why the time is maintained in byte units.

2 tb_period Bytes Time period that should elapse since the last credit update in order for the bucket to be awarded tb_credits_per_period worth or credits.
3 tb_credits_per_period Bytes Credit allowance per tb_period.
4 tb_size Bytes Bucket size, i.e. upper limit for the tb_credits.
5 tb_credits Bytes Number of credits currently in the bucket.

The bucket rate (in bytes per second) can be computed with the following formula:

bucket_rate = (tb_credits_per_period / tb_period) * r

where, r = port line rate (in bytes per second).

Token Bucket Operations
# Token bucket operation Description
1 Initialization tb_credits = 0; or tb_credits = tb_size / 2;
2 Credit update

Credit update options:

  • Every time a packet is sent for a port, update the credits of all the the subports and pipes of that port. Not feasible.
  • Every time a packet is sent, update the credits for the pipe and subport. Very accurate, but not needed (a lot of calculations).
  • Every time a pipe is selected (that is, picked by one of the grinders), update the credits for the pipe and its subport.

The current implementation is using option 3. According to Section Dequeue State Machine, the pipe and subport credits are updated every time a pipe is selected by the dequeue process before the pipe and subport credits are actually used.

The implementation uses a tradeoff between accuracy and speed by updating the bucket credits only when at least a full tb_period has elapsed since the last update.

  • Full accuracy can be achieved by selecting the value for tb_period for which tb_credits_per_period = 1.
  • When full accuracy is not required, better performance is achieved by setting tb_credits to a larger value.

Update operations:

  • n_periods = (time - tb_time) / tb_period;
  • tb_credits += n_periods * tb_credits_per_period;
  • tb_credits = min(tb_credits, tb_size);
  • tb_time += n_periods * tb_period;
3
Credit consumption
(on packet scheduling)

As result of packet scheduling, the necessary number of credits is removed from the bucket. The packet can only be sent if enough credits are in the bucket to send the full packet (packet bytes and framing overhead for the packet).

Scheduling operations:

pkt_credits = pkt_len + frame_overhead; if (tb_credits >= pkt_credits){tb_credits -= pkt_credits;}

Traffic Classes
Implementation of Strict Priority Scheduling

Strict priority scheduling of traffic classes within the same pipe is implemented by the pipe dequeue state machine, which selects the queues in ascending order. Therefore, queues 0..3 (associated with TC 0, highest priority TC) are handled before queues 4..7 (TC 1, lower priority than TC 0), which are handled before queues 8..11 (TC 2), which are handled before queues 12..15 (TC 3, lowest priority TC).

Upper Limit Enforcement

The traffic classes at the pipe and subport levels are not traffic shaped, so there is no token bucket maintained in this context. The upper limit for the traffic classes at the subport and pipe levels is enforced by periodically refilling the subport / pipe traffic class credit counter, out of which credits are consumed every time a packet is scheduled for that subport / pipe, as described in Table 33.10 and Table 33.11.

Subport/Pipe Traffic Class Upper Limit Enforcement Persistent Data Structure
# Subport or pipe field Unit Description
1 tc_time Bytes

Time of the next update (upper limit refill) for the 4 TCs of the current subport / pipe.

See Section Internal Time Reference for the explanation of why the time is maintained in byte units.

2 tc_period Bytes Time between two consecutive updates for the 4 TCs of the current subport / pipe. This is expected to be many times bigger than the typical value of the token bucket tb_period.
3 tc_credits_per_period Bytes Upper limit for the number of credits allowed to be consumed by the current TC during each enforcement period tc_period.
4 tc_credits Bytes Current upper limit for the number of credits that can be consumed by the current traffic class for the remainder of the current enforcement period.
Subport/Pipe Traffic Class Upper Limit Enforcement Operations
# Traffic Class Operation Description
1 Initialization

tc_credits = tc_credits_per_period;

tc_time = tc_period;

2 Credit update

Update operations:

if (time >= tc_time) {

tc_credits = tc_credits_per_period;

tc_time = time + tc_period;

}

3 Credit consumption (on packet scheduling)

As result of packet scheduling, the TC limit is decreased with the necessary number of credits. The packet can only be sent if enough credits are currently available in the TC limit to send the full packet (packet bytes and framing overhead for the packet).

Scheduling operations:

pkt_credits = pk_len + frame_overhead;

if (tc_credits >= pkt_credits) {tc_credits -= pkt_credits;}

Weighted Round Robin (WRR)

The evolution of the WRR design solution from simple to complex is shown in Table 33.12.

Weighted Round Robin (WRR)
# All Queues Active? Equal Weights for All Queues? All Packets Equal? Strategy
1 Yes Yes Yes

Byte level round robin

Next queue queue #i, i = (i + 1) % n

2 Yes Yes No

Packet level round robin

Consuming one byte from queue #i requires consuming exactly one token for queue #i.

T(i) = Accumulated number of tokens previously consumed from queue #i. Every time a packet is consumed from queue #i, T(i) is updated as: T(i) += pkt_len.

Next queue : queue with the smallest T.

3 Yes No No

Packet level weighted round robin

This case can be reduced to the previous case by introducing a cost per byte that is different for each queue. Queues with lower weights have a higher cost per byte. This way, it is still meaningful to compare the consumption amongst different queues in order to select the next queue.

w(i) = Weight of queue #i

t(i) = Tokens per byte for queue #i, defined as the inverse weight of queue #i. For example, if w[0..3] = [1:2:4:8], then t[0..3] = [8:4:2:1]; if w[0..3] = [1:4:15:20], then t[0..3] = [60:15:4:3]. Consuming one byte from queue #i requires consuming t(i) tokens for queue #i.

T(i) = Accumulated number of tokens previously consumed from queue #i. Every time a packet is consumed from queue #i, T(i) is updated as: T(i) += pkt_len * t(i). Next queue : queue with the smallest T.

4 No No No

Packet level weighted round robin with variable queue status

Reduce this case to the previous case by setting the consumption of inactive queues to a high number, so that the inactive queues will never be selected by the smallest T logic.

To prevent T from overflowing as result of successive accumulations, T(i) is truncated after each packet consumption for all queues. For example, T[0..3] = [1000, 1100, 1200, 1300] is truncated to T[0..3] = [0, 100, 200, 300] by subtracting the min T from T(i), i = 0..n.

This requires having at least one active queue in the set of input queues, which is guaranteed by the dequeue state machine never selecting an inactive traffic class.

mask(i) = Saturation mask for queue #i, defined as:

mask(i) = (queue #i is active)? 0 : 0xFFFFFFFF;

w(i) = Weight of queue #i

t(i) = Tokens per byte for queue #i, defined as the inverse weight of queue #i.

T(i) = Accumulated numbers of tokens previously consumed from queue #i.

Next queue : queue with smallest T.

Before packet consumption from queue #i:

T(i) |= mask(i)

After packet consumption from queue #i:

T(j) -= T(i), j != i

T(i) = pkt_len * t(i)

Note: T(j) uses the T(i) value before T(i) is updated.

Subport Traffic Class Oversubscription
Problem Statement

Oversubscription for subport traffic class X is a configuration-time event that occurs when more bandwidth is allocated for traffic class X at the level of subport member pipes than allocated for the same traffic class at the parent subport level.

The existence of the oversubscription for a specific subport and traffic class is solely the result of pipe and subport-level configuration as opposed to being created due to dynamic evolution of the traffic load at run-time (as congestion is).

When the overall demand for traffic class X for the current subport is low, the existence of the oversubscription condition does not represent a problem, as demand for traffic class X is completely satisfied for all member pipes. However, this can no longer be achieved when the aggregated demand for traffic class X for all subport member pipes exceeds the limit configured at the subport level.

Solution Space

summarizes some of the possible approaches for handling this problem, with the third approach selected for implementation.

Subport Traffic Class Oversubscription
No. Approach Description
1 Don’t care

First come, first served.

This approach is not fair amongst subport member pipes, as pipes that are served first will use up as much bandwidth for TC X as they need, while pipes that are served later will receive poor service due to bandwidth for TC X at the subport level being scarce.

2 Scale down all pipes

All pipes within the subport have their bandwidth limit for TC X scaled down by the same factor.

This approach is not fair among subport member pipes, as the low end pipes (that is, pipes configured with low bandwidth) can potentially experience severe service degradation that might render their service unusable (if available bandwidth for these pipes drops below the minimum requirements for a workable service), while the service degradation for high end pipes might not be noticeable at all.

3 Cap the high demand pipes Each subport member pipe receives an equal share of the bandwidth available at run-time for TC X at the subport level. Any bandwidth left unused by the low-demand pipes is redistributed in equal portions to the high-demand pipes. This way, the high-demand pipes are truncated while the low-demand pipes are not impacted.

Typically, the subport TC oversubscription feature is enabled only for the lowest priority traffic class (TC 3), which is typically used for best effort traffic, with the management plane preventing this condition from occurring for the other (higher priority) traffic classes.

To ease implementation, it is also assumed that the upper limit for subport TC 3 is set to 100% of the subport rate, and that the upper limit for pipe TC 3 is set to 100% of pipe rate for all subport member pipes.

Implementation Overview

The algorithm computes a watermark, which is periodically updated based on the current demand experienced by the subport member pipes, whose purpose is to limit the amount of traffic that each pipe is allowed to send for TC 3. The watermark is computed at the subport level at the beginning of each traffic class upper limit enforcement period and the same value is used by all the subport member pipes throughout the current enforcement period. illustrates how the watermark computed as subport level at the beginning of each period is propagated to all subport member pipes.

At the beginning of the current enforcement period (which coincides with the end of the previous enforcement period), the value of the watermark is adjusted based on the amount of bandwidth allocated to TC 3 at the beginning of the previous period that was not left unused by the subport member pipes at the end of the previous period.

If there was subport TC 3 bandwidth left unused, the value of the watermark for the current period is increased to encourage the subport member pipes to consume more bandwidth. Otherwise, the value of the watermark is decreased to enforce equality of bandwidth consumption among subport member pipes for TC 3.

The increase or decrease in the watermark value is done in small increments, so several enforcement periods might be required to reach the equilibrium state. This state can change at any moment due to variations in the demand experienced by the subport member pipes for TC 3, for example, as a result of demand increase (when the watermark needs to be lowered) or demand decrease (when the watermark needs to be increased).

When demand is low, the watermark is set high to prevent it from impeding the subport member pipes from consuming more bandwidth. The highest value for the watermark is picked as the highest rate configured for a subport member pipe. Table 33.14 and Table 33.15 illustrates the watermark operation.

Watermark Propagation from Subport Level to Member Pipes at the Beginning of Each Traffic Class Upper Limit Enforcement Period
No. Subport Traffic Class Operation Description
1 Initialization

Subport level: subport_period_id= 0

Pipe level: pipe_period_id = 0

2 Credit update

Subport Level:

if (time>=subport_tc_time)

{

subport_wm = water_mark_update();

subport_tc_time = time + subport_tc_period;

subport_period_id++;

}

Pipelevel:

if(pipe_period_id != subport_period_id)

{

pipe_ov_credits = subport_wm * pipe_weight;

pipe_period_id = subport_period_id;

}

3 Credit consumption (on packet scheduling)

Pipe level:

pkt_credits = pk_len + frame_overhead;

if(pipe_ov_credits >= pkt_credits{

pipe_ov_credits -= pkt_credits;

}

Watermark Calculation
No. Subport Traffic Class Operation Description
1 Initialization

Subport level:

wm = WM_MAX

2 Credit update

Subport level (water_mark_update):

tc0_cons = subport_tc0_credits_per_period - subport_tc0_credits;

tc1_cons = subport_tc1_credits_per_period - subport_tc1_credits;

tc2_cons = subport_tc2_credits_per_period - subport_tc2_credits;

tc3_cons = subport_tc3_credits_per_period - subport_tc3_credits;

tc3_cons_max = subport_tc3_credits_per_period - (tc0_cons + tc1_cons + tc2_cons);

if(tc3_consumption > (tc3_consumption_max - MTU)){

wm -= wm >> 7;

if(wm < WM_MIN) wm = WM_MIN;

} else {

wm += (wm >> 7) + 1;

if(wm > WM_MAX) wm = WM_MAX;

}

Worst Case Scenarios for Performance
Lots of Active Queues with Not Enough Credits

The more queues the scheduler has to examine for packets and credits in order to select one packet, the lower the performance of the scheduler is.

The scheduler maintains the bitmap of active queues, which skips the non-active queues, but in order to detect whether a specific pipe has enough credits, the pipe has to be drilled down using the pipe dequeue state machine, which consumes cycles regardless of the scheduling result (no packets are produced or at least one packet is produced).

This scenario stresses the importance of the policer for the scheduler performance: if the pipe does not have enough credits, its packets should be dropped as soon as possible (before they reach the hierarchical scheduler), thus rendering the pipe queues as not active, which allows the dequeue side to skip that pipe with no cycles being spent on investigating the pipe credits that would result in a “not enough credits” status.

Single Queue with 100% Line Rate

The port scheduler performance is optimized for a large number of queues. If the number of queues is small, then the performance of the port scheduler for the same level of active traffic is expected to be worse than the performance of a small set of message passing queues.

Dropper

The purpose of the DPDK dropper is to drop packets arriving at a packet scheduler to avoid congestion. The dropper supports the Random Early Detection (RED), Weighted Random Early Detection (WRED) and tail drop algorithms. Fig. 33.7 illustrates how the dropper integrates with the scheduler. The DPDK currently does not support congestion management so the dropper provides the only method for congestion avoidance.

_images/blk_diag_dropper.png

High-level Block Diagram of the DPDK Dropper

The dropper uses the Random Early Detection (RED) congestion avoidance algorithm as documented in the reference publication. The purpose of the RED algorithm is to monitor a packet queue, determine the current congestion level in the queue and decide whether an arriving packet should be enqueued or dropped. The RED algorithm uses an Exponential Weighted Moving Average (EWMA) filter to compute average queue size which gives an indication of the current congestion level in the queue.

For each enqueue operation, the RED algorithm compares the average queue size to minimum and maximum thresholds. Depending on whether the average queue size is below, above or in between these thresholds, the RED algorithm calculates the probability that an arriving packet should be dropped and makes a random decision based on this probability.

The dropper also supports Weighted Random Early Detection (WRED) by allowing the scheduler to select different RED configurations for the same packet queue at run-time. In the case of severe congestion, the dropper resorts to tail drop. This occurs when a packet queue has reached maximum capacity and cannot store any more packets. In this situation, all arriving packets are dropped.

The flow through the dropper is illustrated in Fig. 33.8. The RED/WRED algorithm is exercised first and tail drop second.

_images/flow_tru_droppper.png

Flow Through the Dropper

The use cases supported by the dropper are:

    • Initialize configuration data
    • Initialize run-time data
    • Enqueue (make a decision to enqueue or drop an arriving packet)
    • Mark empty (record the time at which a packet queue becomes empty)

The configuration use case is explained in Section2.23.3.1, the enqueue operation is explained in Section 2.23.3.2 and the mark empty operation is explained in Section 2.23.3.3.

Configuration

A RED configuration contains the parameters given in Table 33.16.

RED Configuration Parameters
Parameter Minimum Maximum Typical
Minimum Threshold 0 1022 1/4 x queue size
Maximum Threshold 1 1023 1/2 x queue size
Inverse Mark Probability 1 255 10
EWMA Filter Weight 1 12 9

The meaning of these parameters is explained in more detail in the following sections. The format of these parameters as specified to the dropper module API corresponds to the format used by Cisco* in their RED implementation. The minimum and maximum threshold parameters are specified to the dropper module in terms of number of packets. The mark probability parameter is specified as an inverse value, for example, an inverse mark probability parameter value of 10 corresponds to a mark probability of 1/10 (that is, 1 in 10 packets will be dropped). The EWMA filter weight parameter is specified as an inverse log value, for example, a filter weight parameter value of 9 corresponds to a filter weight of 1/29.

Enqueue Operation

In the example shown in Fig. 33.9, q (actual queue size) is the input value, avg (average queue size) and count (number of packets since the last drop) are run-time values, decision is the output value and the remaining values are configuration parameters.

_images/ex_data_flow_tru_dropper.png

Example Data Flow Through Dropper

EWMA Filter Microblock

The purpose of the EWMA Filter microblock is to filter queue size values to smooth out transient changes that result from “bursty” traffic. The output value is the average queue size which gives a more stable view of the current congestion level in the queue.

The EWMA filter has one configuration parameter, filter weight, which determines how quickly or slowly the average queue size output responds to changes in the actual queue size input. Higher values of filter weight mean that the average queue size responds more quickly to changes in actual queue size.

Average Queue Size Calculation when the Queue is not Empty

The definition of the EWMA filter is given in the following equation.

_images/ewma_filter_eq_1.png

Where:

  • avg = average queue size
  • wq = filter weight
  • q = actual queue size

Note

The filter weight, wq = 1/2^n, where n is the filter weight parameter value passed to the dropper module
on configuration (see Section2.23.3.1 ).
Average Queue Size Calculation when the Queue is Empty

The EWMA filter does not read time stamps and instead assumes that enqueue operations will happen quite regularly. Special handling is required when the queue becomes empty as the queue could be empty for a short time or a long time. When the queue becomes empty, average queue size should decay gradually to zero instead of dropping suddenly to zero or remaining stagnant at the last computed value. When a packet is enqueued on an empty queue, the average queue size is computed using the following formula:

_images/ewma_filter_eq_2.png

Where:

  • m = the number of enqueue operations that could have occurred on this queue while the queue was empty

In the dropper module, m is defined as:

_images/m_definition.png

Where:

  • time = current time
  • qtime = time the queue became empty
  • s = typical time between successive enqueue operations on this queue

The time reference is in units of bytes, where a byte signifies the time duration required by the physical interface to send out a byte on the transmission medium (see Section Internal Time Reference). The parameter s is defined in the dropper module as a constant with the value: s=2^22. This corresponds to the time required by every leaf node in a hierarchy with 64K leaf nodes to transmit one 64-byte packet onto the wire and represents the worst case scenario. For much smaller scheduler hierarchies, it may be necessary to reduce the parameter s, which is defined in the red header source file (rte_red.h) as:

#define RTE_RED_S

Since the time reference is in bytes, the port speed is implied in the expression: time-qtime. The dropper does not have to be configured with the actual port speed. It adjusts automatically to low speed and high speed links.

Implementation

A numerical method is used to compute the factor (1-wq)^m that appears in Equation 2.

This method is based on the following identity:

_images/eq2_factor.png

This allows us to express the following:

_images/eq2_expression.png

In the dropper module, a look-up table is used to compute log2(1-wq) for each value of wq supported by the dropper module. The factor (1-wq)^m can then be obtained by multiplying the table value by m and applying shift operations. To avoid overflow in the multiplication, the value, m, and the look-up table values are limited to 16 bits. The total size of the look-up table is 56 bytes. Once the factor (1-wq)^m is obtained using this method, the average queue size can be calculated from Equation 2.

Alternative Approaches

Other methods for calculating the factor (1-wq)^m in the expression for computing average queue size when the queue is empty (Equation 2) were considered. These approaches include:

  • Floating-point evaluation
  • Fixed-point evaluation using a small look-up table (512B) and up to 16 multiplications (this is the approach used in the FreeBSD* ALTQ RED implementation)
  • Fixed-point evaluation using a small look-up table (512B) and 16 SSE multiplications (SSE optimized version of the approach used in the FreeBSD* ALTQ RED implementation)
  • Large look-up table (76 KB)

The method that was finally selected (described above in Section 26.3.2.2.1) out performs all of these approaches in terms of run-time performance and memory requirements and also achieves accuracy comparable to floating-point evaluation. Table 33.17 lists the performance of each of these alternative approaches relative to the method that is used in the dropper. As can be seen, the floating-point implementation achieved the worst performance.

Relative Performance of Alternative Approaches
Method Relative Performance
Current dropper method (see Section 23.3.2.1.3) 100%
Fixed-point method with small (512B) look-up table 148%
SSE method with small (512B) look-up table 114%
Large (76KB) look-up table 118%
Floating-point 595%
Note: In this case, since performance is expressed as time spent executing the operation in a specific condition, any relative performance value above 100% runs slower than the reference method.
Drop Decision Block

The Drop Decision block:

  • Compares the average queue size with the minimum and maximum thresholds
  • Calculates a packet drop probability
  • Makes a random decision to enqueue or drop an arriving packet

The calculation of the drop probability occurs in two stages. An initial drop probability is calculated based on the average queue size, the minimum and maximum thresholds and the mark probability. An actual drop probability is then computed from the initial drop probability. The actual drop probability takes the count run-time value into consideration so that the actual drop probability increases as more packets arrive to the packet queue since the last packet was dropped.

Initial Packet Drop Probability

The initial drop probability is calculated using the following equation.

_images/drop_probability_eq3.png

Where:

  • maxp = mark probability
  • avg = average queue size
  • minth = minimum threshold
  • maxth = maximum threshold

The calculation of the packet drop probability using Equation 3 is illustrated in Fig. 33.10. If the average queue size is below the minimum threshold, an arriving packet is enqueued. If the average queue size is at or above the maximum threshold, an arriving packet is dropped. If the average queue size is between the minimum and maximum thresholds, a drop probability is calculated to determine if the packet should be enqueued or dropped.

_images/pkt_drop_probability.png

Packet Drop Probability for a Given RED Configuration

Actual Drop Probability

If the average queue size is between the minimum and maximum thresholds, then the actual drop probability is calculated from the following equation.

_images/drop_probability_eq4.png

Where:

  • Pb = initial drop probability (from Equation 3)
  • count = number of packets that have arrived since the last drop

The constant 2, in Equation 4 is the only deviation from the drop probability formulae given in the reference document where a value of 1 is used instead. It should be noted that the value pa computed from can be negative or greater than 1. If this is the case, then a value of 1 should be used instead.

The initial and actual drop probabilities are shown in Fig. 33.11. The actual drop probability is shown for the case where the formula given in the reference document1 is used (blue curve) and also for the case where the formula implemented in the dropper module, is used (red curve). The formula in the reference document results in a significantly higher drop rate compared to the mark probability configuration parameter specified by the user. The choice to deviate from the reference document is simply a design decision and one that has been taken by other RED implementations, for example, FreeBSD* ALTQ RED.

_images/drop_probability_graph.png

Initial Drop Probability (pb), Actual Drop probability (pa) Computed Using a Factor 1 (Blue Curve) and a Factor 2 (Red Curve)

Queue Empty Operation

The time at which a packet queue becomes empty must be recorded and saved with the RED run-time data so that the EWMA filter block can calculate the average queue size on the next enqueue operation. It is the responsibility of the calling application to inform the dropper module through the API that a queue has become empty.

Source Files Location

The source files for the DPDK dropper are located at:

  • DPDK/lib/librte_sched/rte_red.h
  • DPDK/lib/librte_sched/rte_red.c
Integration with the DPDK QoS Scheduler

RED functionality in the DPDK QoS scheduler is disabled by default. To enable it, use the DPDK configuration parameter:

CONFIG_RTE_SCHED_RED=y

This parameter must be set to y. The parameter is found in the build configuration files in the DPDK/config directory, for example, DPDK/config/common_linuxapp. RED configuration parameters are specified in the rte_red_params structure within the rte_sched_port_params structure that is passed to the scheduler on initialization. RED parameters are specified separately for four traffic classes and three packet colors (green, yellow and red) allowing the scheduler to implement Weighted Random Early Detection (WRED).

Integration with the DPDK QoS Scheduler Sample Application

The DPDK QoS Scheduler Application reads a configuration file on start-up. The configuration file includes a section containing RED parameters. The format of these parameters is described in Section2.23.3.1. A sample RED configuration is shown below. In this example, the queue size is 64 packets.

Note

For correct operation, the same EWMA filter weight parameter (wred weight) should be used for each packet color (green, yellow, red) in the same traffic class (tc).

 ; RED params per traffic class and color (Green / Yellow / Red)

[red]
tc 0 wred min = 28 22 16
tc 0 wred max = 32 32 32
tc 0 wred inv prob = 10 10 10
tc 0 wred weight = 9 9 9

tc 1 wred min = 28 22 16
tc 1 wred max = 32 32 32
tc 1 wred inv prob = 10 10 10
tc 1 wred weight = 9 9 9

tc 2 wred min = 28 22 16
tc 2 wred max = 32 32 32
tc 2 wred inv prob = 10 10 10
tc 2 wred weight = 9 9 9

tc 3 wred min = 28 22 16
tc 3 wred max = 32 32 32
tc 3 wred inv prob = 10 10 10
tc 3 wred weight = 9 9 9

With this configuration file, the RED configuration that applies to green, yellow and red packets in traffic class 0 is shown in Table 33.18.

RED Configuration Corresponding to RED Configuration File
RED Parameter Configuration Name Green Yellow Red
Minimum Threshold tc 0 wred min 28 22 16
Maximum Threshold tc 0 wred max 32 32 32
Mark Probability tc 0 wred inv prob 10 10 10
EWMA Filter Weight tc 0 wred weight 9 9 9
Application Programming Interface (API)
Enqueue API

The syntax of the enqueue API is as follows:

int rte_red_enqueue(const struct rte_red_config *red_cfg, struct rte_red *red, const unsigned q, const uint64_t time)

The arguments passed to the enqueue API are configuration data, run-time data, the current size of the packet queue (in packets) and a value representing the current time. The time reference is in units of bytes, where a byte signifies the time duration required by the physical interface to send out a byte on the transmission medium (see Section 26.2.4.5.1 “Internal Time Reference” ). The dropper reuses the scheduler time stamps for performance reasons.

Empty API

The syntax of the empty API is as follows:

void rte_red_mark_queue_empty(struct rte_red *red, const uint64_t time)

The arguments passed to the empty API are run-time data and the current time in bytes.

Traffic Metering

The traffic metering component implements the Single Rate Three Color Marker (srTCM) and Two Rate Three Color Marker (trTCM) algorithms, as defined by IETF RFC 2697 and 2698 respectively. These algorithms meter the stream of incoming packets based on the allowance defined in advance for each traffic flow. As result, each incoming packet is tagged as green, yellow or red based on the monitored consumption of the flow the packet belongs to.

Functional Overview

The srTCM algorithm defines two token buckets for each traffic flow, with the two buckets sharing the same token update rate:

  • Committed (C) bucket: fed with tokens at the rate defined by the Committed Information Rate (CIR) parameter (measured in IP packet bytes per second). The size of the C bucket is defined by the Committed Burst Size (CBS) parameter (measured in bytes);
  • Excess (E) bucket: fed with tokens at the same rate as the C bucket. The size of the E bucket is defined by the Excess Burst Size (EBS) parameter (measured in bytes).

The trTCM algorithm defines two token buckets for each traffic flow, with the two buckets being updated with tokens at independent rates:

  • Committed (C) bucket: fed with tokens at the rate defined by the Committed Information Rate (CIR) parameter (measured in bytes of IP packet per second). The size of the C bucket is defined by the Committed Burst Size (CBS) parameter (measured in bytes);
  • Peak (P) bucket: fed with tokens at the rate defined by the Peak Information Rate (PIR) parameter (measured in IP packet bytes per second). The size of the P bucket is defined by the Peak Burst Size (PBS) parameter (measured in bytes).

Please refer to RFC 2697 (for srTCM) and RFC 2698 (for trTCM) for details on how tokens are consumed from the buckets and how the packet color is determined.

Color Blind and Color Aware Modes

For both algorithms, the color blind mode is functionally equivalent to the color aware mode with input color set as green. For color aware mode, a packet with red input color can only get the red output color, while a packet with yellow input color can only get the yellow or red output colors.

The reason why the color blind mode is still implemented distinctly than the color aware mode is that color blind mode can be implemented with fewer operations than the color aware mode.

Implementation Overview

For each input packet, the steps for the srTCM / trTCM algorithms are:

  • Update the C and E / P token buckets. This is done by reading the current time (from the CPU timestamp counter), identifying the amount of time since the last bucket update and computing the associated number of tokens (according to the pre-configured bucket rate). The number of tokens in the bucket is limited by the pre-configured bucket size;
  • Identify the output color for the current packet based on the size of the IP packet and the amount of tokens currently available in the C and E / P buckets; for color aware mode only, the input color of the packet is also considered. When the output color is not red, a number of tokens equal to the length of the IP packet are subtracted from the C or E /P or both buckets, depending on the algorithm and the output color of the packet.

Power Management

The DPDK Power Management feature allows users space applications to save power by dynamically adjusting CPU frequency or entering into different C-States.

  • Adjusting the CPU frequency dynamically according to the utilization of RX queue.
  • Entering into different deeper C-States according to the adaptive algorithms to speculate brief periods of time suspending the application if no packets are received.

The interfaces for adjusting the operating CPU frequency are in the power management library. C-State control is implemented in applications according to the different use cases.

CPU Frequency Scaling

The Linux kernel provides a cpufreq module for CPU frequency scaling for each lcore. For example, for cpuX, /sys/devices/system/cpu/cpuX/cpufreq/ has the following sys files for frequency scaling:

  • affected_cpus
  • bios_limit
  • cpuinfo_cur_freq
  • cpuinfo_max_freq
  • cpuinfo_min_freq
  • cpuinfo_transition_latency
  • related_cpus
  • scaling_available_frequencies
  • scaling_available_governors
  • scaling_cur_freq
  • scaling_driver
  • scaling_governor
  • scaling_max_freq
  • scaling_min_freq
  • scaling_setspeed

In the DPDK, scaling_governor is configured in user space. Then, a user space application can prompt the kernel by writing scaling_setspeed to adjust the CPU frequency according to the strategies defined by the user space application.

Core-load Throttling through C-States

Core state can be altered by speculative sleeps whenever the specified lcore has nothing to do. In the DPDK, if no packet is received after polling, speculative sleeps can be triggered according the strategies defined by the user space application.

Per-core Turbo Boost

Individual cores can be allowed to enter a Turbo Boost state on a per-core basis. This is achieved by enabling Turbo Boost Technology in the BIOS, then looping through the relevant cores and enabling/disabling Turbo Boost on each core.

API Overview of the Power Library

The main methods exported by power library are for CPU frequency scaling and include the following:

  • Freq up: Prompt the kernel to scale up the frequency of the specific lcore.
  • Freq down: Prompt the kernel to scale down the frequency of the specific lcore.
  • Freq max: Prompt the kernel to scale up the frequency of the specific lcore to the maximum.
  • Freq min: Prompt the kernel to scale down the frequency of the specific lcore to the minimum.
  • Get available freqs: Read the available frequencies of the specific lcore from the sys file.
  • Freq get: Get the current frequency of the specific lcore.
  • Freq set: Prompt the kernel to set the frequency for the specific lcore.
  • Enable turbo: Prompt the kernel to enable Turbo Boost for the specific lcore.
  • Disable turbo: Prompt the kernel to disable Turbo Boost for the specific lcore.

User Cases

The power management mechanism is used to save power when performing L3 forwarding.

References

  • l3fwd-power: The sample application in DPDK that performs L3 forwarding with power management.
  • The “L3 Forwarding with Power Management Sample Application” chapter in the DPDK Sample Application’s User Guide.

Packet Classification and Access Control

The DPDK provides an Access Control library that gives the ability to classify an input packet based on a set of classification rules.

The ACL library is used to perform an N-tuple search over a set of rules with multiple categories and find the best match (highest priority) for each category. The library API provides the following basic operations:

  • Create a new Access Control (AC) context.
  • Add rules into the context.
  • For all rules in the context, build the runtime structures necessary to perform packet classification.
  • Perform input packet classifications.
  • Destroy an AC context and its runtime structures and free the associated memory.

Overview

Rule definition

The current implementation allows the user for each AC context to specify its own rule (set of fields) over which packet classification will be performed. Though there are few restrictions on the rule fields layout:

  • First field in the rule definition has to be one byte long.
  • All subsequent fields has to be grouped into sets of 4 consecutive bytes.

This is done mainly for performance reasons - search function processes the first input byte as part of the flow setup and then the inner loop of the search function is unrolled to process four input bytes at a time.

To define each field inside an AC rule, the following structure is used:

struct rte_acl_field_def {
    uint8_t type;         /*< type - ACL_FIELD_TYPE. */
    uint8_t size;         /*< size of field 1,2,4, or 8. */
    uint8_t field_index;  /*< index of field inside the rule. */
    uint8_t input_index;  /*< 0-N input index. */
    uint32_t offset;      /*< offset to start of field. */
};
  • type The field type is one of three choices:
    • _MASK - for fields such as IP addresses that have a value and a mask defining the number of relevant bits.
    • _RANGE - for fields such as ports that have a lower and upper value for the field.
    • _BITMASK - for fields such as protocol identifiers that have a value and a bit mask.
  • size The size parameter defines the length of the field in bytes. Allowable values are 1, 2, 4, or 8 bytes. Note that due to the grouping of input bytes, 1 or 2 byte fields must be defined as consecutive fields that make up 4 consecutive input bytes. Also, it is best to define fields of 8 or more bytes as 4 byte fields so that the build processes can eliminate fields that are all wild.
  • field_index A zero-based value that represents the position of the field inside the rule; 0 to N-1 for N fields.
  • input_index As mentioned above, all input fields, except the very first one, must be in groups of 4 consecutive bytes. The input index specifies to which input group that field belongs to.
  • offset The offset field defines the offset for the field. This is the offset from the beginning of the buffer parameter for the search.

For example, to define classification for the following IPv4 5-tuple structure:

struct ipv4_5tuple {
    uint8_t proto;
    uint32_t ip_src;
    uint32_t ip_dst;
    uint16_t port_src;
    uint16_t port_dst;
};

The following array of field definitions can be used:

struct rte_acl_field_def ipv4_defs[5] = {
    /* first input field - always one byte long. */
    {
        .type = RTE_ACL_FIELD_TYPE_BITMASK,
        .size = sizeof (uint8_t),
        .field_index = 0,
        .input_index = 0,
        .offset = offsetof (struct ipv4_5tuple, proto),
    },

    /* next input field (IPv4 source address) - 4 consecutive bytes. */
    {
        .type = RTE_ACL_FIELD_TYPE_MASK,
        .size = sizeof (uint32_t),
        .field_index = 1,
        .input_index = 1,
       .offset = offsetof (struct ipv4_5tuple, ip_src),
    },

    /* next input field (IPv4 destination address) - 4 consecutive bytes. */
    {
        .type = RTE_ACL_FIELD_TYPE_MASK,
        .size = sizeof (uint32_t),
        .field_index = 2,
        .input_index = 2,
       .offset = offsetof (struct ipv4_5tuple, ip_dst),
    },

    /*
     * Next 2 fields (src & dst ports) form 4 consecutive bytes.
     * They share the same input index.
     */
    {
        .type = RTE_ACL_FIELD_TYPE_RANGE,
        .size = sizeof (uint16_t),
        .field_index = 3,
        .input_index = 3,
        .offset = offsetof (struct ipv4_5tuple, port_src),
    },

    {
        .type = RTE_ACL_FIELD_TYPE_RANGE,
        .size = sizeof (uint16_t),
        .field_index = 4,
        .input_index = 3,
        .offset = offsetof (struct ipv4_5tuple, port_dst),
    },
};

A typical example of such an IPv4 5-tuple rule is a follows:

source addr/mask  destination addr/mask  source ports dest ports protocol/mask
192.168.1.0/24    192.168.2.31/32        0:65535      1234:1234  17/0xff

Any IPv4 packets with protocol ID 17 (UDP), source address 192.168.1.[0-255], destination address 192.168.2.31, source port [0-65535] and destination port 1234 matches the above rule.

To define classification for the IPv6 2-tuple: <protocol, IPv6 source address> over the following IPv6 header structure:

struct struct ipv6_hdr {
    uint32_t vtc_flow;     /* IP version, traffic class & flow label. */
    uint16_t payload_len;  /* IP packet length - includes sizeof(ip_header). */
    uint8_t proto;         /* Protocol, next header. */
    uint8_t hop_limits;    /* Hop limits. */
    uint8_t src_addr[16];  /* IP address of source host. */
    uint8_t dst_addr[16];  /* IP address of destination host(s). */
} __attribute__((__packed__));

The following array of field definitions can be used:

struct struct rte_acl_field_def ipv6_2tuple_defs[5] = {
    {
        .type = RTE_ACL_FIELD_TYPE_BITMASK,
        .size = sizeof (uint8_t),
        .field_index = 0,
        .input_index = 0,
        .offset = offsetof (struct ipv6_hdr, proto),
    },

    {
        .type = RTE_ACL_FIELD_TYPE_MASK,
        .size = sizeof (uint32_t),
        .field_index = 1,
        .input_index = 1,
        .offset = offsetof (struct ipv6_hdr, src_addr[0]),
    },

    {
        .type = RTE_ACL_FIELD_TYPE_MASK,
        .size = sizeof (uint32_t),
        .field_index = 2,
        .input_index = 2,
        .offset = offsetof (struct ipv6_hdr, src_addr[4]),
    },

    {
        .type = RTE_ACL_FIELD_TYPE_MASK,
        .size = sizeof (uint32_t),
        .field_index = 3,
        .input_index = 3,
       .offset = offsetof (struct ipv6_hdr, src_addr[8]),
    },

    {
       .type = RTE_ACL_FIELD_TYPE_MASK,
       .size = sizeof (uint32_t),
       .field_index = 4,
       .input_index = 4,
       .offset = offsetof (struct ipv6_hdr, src_addr[12]),
    },
};

A typical example of such an IPv6 2-tuple rule is a follows:

source addr/mask                              protocol/mask
2001:db8:1234:0000:0000:0000:0000:0000/48     6/0xff

Any IPv6 packets with protocol ID 6 (TCP), and source address inside the range [2001:db8:1234:0000:0000:0000:0000:0000 - 2001:db8:1234:ffff:ffff:ffff:ffff:ffff] matches the above rule.

In the following example the last element of the search key is 8-bit long. So it is a case where the 4 consecutive bytes of an input field are not fully occupied. The structure for the classification is:

struct acl_key {
    uint8_t ip_proto;
    uint32_t ip_src;
    uint32_t ip_dst;
    uint8_t tos;      /*< This is partially using a 32-bit input element */
};

The following array of field definitions can be used:

struct rte_acl_field_def ipv4_defs[4] = {
    /* first input field - always one byte long. */
    {
        .type = RTE_ACL_FIELD_TYPE_BITMASK,
        .size = sizeof (uint8_t),
        .field_index = 0,
        .input_index = 0,
        .offset = offsetof (struct acl_key, ip_proto),
    },

    /* next input field (IPv4 source address) - 4 consecutive bytes. */
    {
        .type = RTE_ACL_FIELD_TYPE_MASK,
        .size = sizeof (uint32_t),
        .field_index = 1,
        .input_index = 1,
       .offset = offsetof (struct acl_key, ip_src),
    },

    /* next input field (IPv4 destination address) - 4 consecutive bytes. */
    {
        .type = RTE_ACL_FIELD_TYPE_MASK,
        .size = sizeof (uint32_t),
        .field_index = 2,
        .input_index = 2,
       .offset = offsetof (struct acl_key, ip_dst),
    },

    /*
     * Next element of search key (Type of Service) is indeed 1 byte long.
     * Anyway we need to allocate all the 4 consecutive bytes for it.
     */
    {
        .type = RTE_ACL_FIELD_TYPE_BITMASK,
        .size = sizeof (uint32_t), /* All the 4 consecutive bytes are allocated */
        .field_index = 3,
        .input_index = 3,
        .offset = offsetof (struct acl_key, tos),
    },
};

A typical example of such an IPv4 4-tuple rule is as follows:

source addr/mask  destination addr/mask  tos/mask protocol/mask
192.168.1.0/24    192.168.2.31/32        1/0xff   6/0xff

Any IPv4 packets with protocol ID 6 (TCP), source address 192.168.1.[0-255], destination address 192.168.2.31, ToS 1 matches the above rule.

When creating a set of rules, for each rule, additional information must be supplied also:

  • priority: A weight to measure the priority of the rules (higher is better). If the input tuple matches more than one rule, then the rule with the higher priority is returned. Note that if the input tuple matches more than one rule and these rules have equal priority, it is undefined which rule is returned as a match. It is recommended to assign a unique priority for each rule.
  • category_mask: Each rule uses a bit mask value to select the relevant category(s) for the rule. When a lookup is performed, the result for each category is returned. This effectively provides a “parallel lookup” by enabling a single search to return multiple results if, for example, there were four different sets of ACL rules, one for access control, one for routing, and so on. Each set could be assigned its own category and by combining them into a single database, one lookup returns a result for each of the four sets.
  • userdata: A user-defined value. For each category, a successful match returns the userdata field of the highest priority matched rule. When no rules match, returned value is zero.

Note

When adding new rules into an ACL context, all fields must be in host byte order (LSB). When the search is performed for an input tuple, all fields in that tuple must be in network byte order (MSB).

RT memory size limit

Build phase (rte_acl_build()) creates for a given set of rules internal structure for further run-time traversal. With current implementation it is a set of multi-bit tries (with stride == 8). Depending on the rules set, that could consume significant amount of memory. In attempt to conserve some space ACL build process tries to split the given rule-set into several non-intersecting subsets and construct a separate trie for each of them. Depending on the rule-set, it might reduce RT memory requirements but might increase classification time. There is a possibility at build-time to specify maximum memory limit for internal RT structures for given AC context. It could be done via max_size field of the rte_acl_config structure. Setting it to the value greater than zero, instructs rte_acl_build() to:

  • attempt to minimize number of tries in the RT table, but
  • make sure that size of RT table wouldn’t exceed given value.

Setting it to zero makes rte_acl_build() to use the default behavior: try to minimize size of the RT structures, but doesn’t expose any hard limit on it.

That gives the user the ability to decisions about performance/space trade-off. For example:

struct rte_acl_ctx * acx;
struct rte_acl_config cfg;
int ret;

/*
 * assuming that acx points to already created and
 * populated with rules AC context and cfg filled properly.
 */

 /* try to build AC context, with RT structures less then 8MB. */
 cfg.max_size = 0x800000;
 ret = rte_acl_build(acx, &cfg);

 /*
  * RT structures can't fit into 8MB for given context.
  * Try to build without exposing any hard limit.
  */
 if (ret == -ERANGE) {
    cfg.max_size = 0;
    ret = rte_acl_build(acx, &cfg);
 }
Classification methods

After rte_acl_build() over given AC context has finished successfully, it can be used to perform classification - search for a rule with highest priority over the input data. There are several implementations of classify algorithm:

  • RTE_ACL_CLASSIFY_SCALAR: generic implementation, doesn’t require any specific HW support.
  • RTE_ACL_CLASSIFY_SSE: vector implementation, can process up to 8 flows in parallel. Requires SSE 4.1 support.
  • RTE_ACL_CLASSIFY_AVX2: vector implementation, can process up to 16 flows in parallel. Requires AVX2 support.

It is purely a runtime decision which method to choose, there is no build-time difference. All implementations operates over the same internal RT structures and use similar principles. The main difference is that vector implementations can manually exploit IA SIMD instructions and process several input data flows in parallel. At startup ACL library determines the highest available classify method for the given platform and sets it as default one. Though the user has an ability to override the default classifier function for a given ACL context or perform particular search using non-default classify method. In that case it is user responsibility to make sure that given platform supports selected classify implementation.

Application Programming Interface (API) Usage

Note

For more details about the Access Control API, please refer to the DPDK API Reference.

The following example demonstrates IPv4, 5-tuple classification for rules defined above with multiple categories in more detail.

Classify with Multiple Categories
struct rte_acl_ctx * acx;
struct rte_acl_config cfg;
int ret;

/* define a structure for the rule with up to 5 fields. */

RTE_ACL_RULE_DEF(acl_ipv4_rule, RTE_DIM(ipv4_defs));

/* AC context creation parameters. */

struct rte_acl_param prm = {
    .name = "ACL_example",
    .socket_id = SOCKET_ID_ANY,
    .rule_size = RTE_ACL_RULE_SZ(RTE_DIM(ipv4_defs)),

    /* number of fields per rule. */

    .max_rule_num = 8, /* maximum number of rules in the AC context. */
};

struct acl_ipv4_rule acl_rules[] = {

    /* matches all packets traveling to 192.168.0.0/16, applies for categories: 0,1 */
    {
        .data = {.userdata = 1, .category_mask = 3, .priority = 1},

        /* destination IPv4 */
        .field[2] = {.value.u32 = IPv4(192,168,0,0),. mask_range.u32 = 16,},

        /* source port */
        .field[3] = {.value.u16 = 0, .mask_range.u16 = 0xffff,},

        /* destination port */
       .field[4] = {.value.u16 = 0, .mask_range.u16 = 0xffff,},
    },

    /* matches all packets traveling to 192.168.1.0/24, applies for categories: 0 */
    {
        .data = {.userdata = 2, .category_mask = 1, .priority = 2},

        /* destination IPv4 */
        .field[2] = {.value.u32 = IPv4(192,168,1,0),. mask_range.u32 = 24,},

        /* source port */
        .field[3] = {.value.u16 = 0, .mask_range.u16 = 0xffff,},

        /* destination port */
        .field[4] = {.value.u16 = 0, .mask_range.u16 = 0xffff,},
    },

    /* matches all packets traveling from 10.1.1.1, applies for categories: 1 */
    {
        .data = {.userdata = 3, .category_mask = 2, .priority = 3},

        /* source IPv4 */
        .field[1] = {.value.u32 = IPv4(10,1,1,1),. mask_range.u32 = 32,},

        /* source port */
        .field[3] = {.value.u16 = 0, .mask_range.u16 = 0xffff,},

        /* destination port */
        .field[4] = {.value.u16 = 0, .mask_range.u16 = 0xffff,},
    },

};


/* create an empty AC context  */

if ((acx = rte_acl_create(&prm)) == NULL) {

    /* handle context create failure. */

}

/* add rules to the context */

ret = rte_acl_add_rules(acx, acl_rules, RTE_DIM(acl_rules));
if (ret != 0) {
   /* handle error at adding ACL rules. */
}

/* prepare AC build config. */

cfg.num_categories = 2;
cfg.num_fields = RTE_DIM(ipv4_defs);

memcpy(cfg.defs, ipv4_defs, sizeof (ipv4_defs));

/* build the runtime structures for added rules, with 2 categories. */

ret = rte_acl_build(acx, &cfg);
if (ret != 0) {
   /* handle error at build runtime structures for ACL context. */
}

For a tuple with source IP address: 10.1.1.1 and destination IP address: 192.168.1.15, once the following lines are executed:

uint32_t results[4]; /* make classify for 4 categories. */

rte_acl_classify(acx, data, results, 1, 4);

then the results[] array contains:

results[4] = {2, 3, 0, 0};
  • For category 0, both rules 1 and 2 match, but rule 2 has higher priority, therefore results[0] contains the userdata for rule 2.
  • For category 1, both rules 1 and 3 match, but rule 3 has higher priority, therefore results[1] contains the userdata for rule 3.
  • For categories 2 and 3, there are no matches, so results[2] and results[3] contain zero, which indicates that no matches were found for those categories.

For a tuple with source IP address: 192.168.1.1 and destination IP address: 192.168.2.11, once the following lines are executed:

uint32_t results[4]; /* make classify by 4 categories. */

rte_acl_classify(acx, data, results, 1, 4);

the results[] array contains:

results[4] = {1, 1, 0, 0};
  • For categories 0 and 1, only rule 1 matches.
  • For categories 2 and 3, there are no matches.

For a tuple with source IP address: 10.1.1.1 and destination IP address: 201.212.111.12, once the following lines are executed:

uint32_t results[4]; /* make classify by 4 categories. */
rte_acl_classify(acx, data, results, 1, 4);

the results[] array contains:

results[4] = {0, 3, 0, 0};
  • For category 1, only rule 3 matches.
  • For categories 0, 2 and 3, there are no matches.

Packet Framework

Design Objectives

The main design objectives for the DPDK Packet Framework are:

  • Provide standard methodology to build complex packet processing pipelines. Provide reusable and extensible templates for the commonly used pipeline functional blocks;
  • Provide capability to switch between pure software and hardware-accelerated implementations for the same pipeline functional block;
  • Provide the best trade-off between flexibility and performance. Hardcoded pipelines usually provide the best performance, but are not flexible, while developing flexible frameworks is never a problem, but performance is usually low;
  • Provide a framework that is logically similar to Open Flow.

Overview

Packet processing applications are frequently structured as pipelines of multiple stages, with the logic of each stage glued around a lookup table. For each incoming packet, the table defines the set of actions to be applied to the packet, as well as the next stage to send the packet to.

The DPDK Packet Framework minimizes the development effort required to build packet processing pipelines by defining a standard methodology for pipeline development, as well as providing libraries of reusable templates for the commonly used pipeline blocks.

The pipeline is constructed by connecting the set of input ports with the set of output ports through the set of tables in a tree-like topology. As result of lookup operation for the current packet in the current table, one of the table entries (on lookup hit) or the default table entry (on lookup miss) provides the set of actions to be applied on the current packet, as well as the next hop for the packet, which can be either another table, an output port or packet drop.

An example of packet processing pipeline is presented in Fig. 36.1:

_images/figure32.png

Example of Packet Processing Pipeline where Input Ports 0 and 1 are Connected with Output Ports 0, 1 and 2 through Tables 0 and 1

Port Library Design

Port Types

Table 36.1 is a non-exhaustive list of ports that can be implemented with the Packet Framework.

Port Types
# Port type Description
1 SW ring SW circular buffer used for message passing between the application threads. Uses the DPDK rte_ring primitive. Expected to be the most commonly used type of port.
2 HW ring Queue of buffer descriptors used to interact with NIC, switch or accelerator ports. For NIC ports, it uses the DPDK rte_eth_rx_queue or rte_eth_tx_queue primitives.
3 IP reassembly Input packets are either IP fragments or complete IP datagrams. Output packets are complete IP datagrams.
4 IP fragmentation Input packets are jumbo (IP datagrams with length bigger than MTU) or non-jumbo packets. Output packets are non-jumbo packets.
5 Traffic manager Traffic manager attached to a specific NIC output port, performing congestion management and hierarchical scheduling according to pre-defined SLAs.
6 KNI Send/receive packets to/from Linux kernel space.
7 Source Input port used as packet generator. Similar to Linux kernel /dev/zero character device.
8 Sink Output port used to drop all input packets. Similar to Linux kernel /dev/null character device.
Port Interface

Each port is unidirectional, i.e. either input port or output port. Each input/output port is required to implement an abstract interface that defines the initialization and run-time operation of the port. The port abstract interface is described in.

20 Port Abstract Interface
# Port Operation Description
1 Create Create the low-level port object (e.g. queue). Can internally allocate memory.
2 Free Free the resources (e.g. memory) used by the low-level port object.
3 RX Read a burst of input packets. Non-blocking operation. Only defined for input ports.
4 TX Write a burst of input packets. Non-blocking operation. Only defined for output ports.
5 Flush Flush the output buffer. Only defined for output ports.

Table Library Design

Table Types

Table 36.3 is a non-exhaustive list of types of tables that can be implemented with the Packet Framework.

Table Types
# Table Type Description
1 Hash table

Lookup key is n-tuple based.

Typically, the lookup key is hashed to produce a signature that is used to identify a bucket of entries where the lookup key is searched next.

The signature associated with the lookup key of each input packet is either read from the packet descriptor (pre-computed signature) or computed at table lookup time.

The table lookup, add entry and delete entry operations, as well as any other pipeline block that pre-computes the signature all have to use the same hashing algorithm to generate the signature.

Typically used to implement flow classification tables, ARP caches, routing table for tunnelling protocols, etc.

2 Longest Prefix Match (LPM)

Lookup key is the IP address.

Each table entries has an associated IP prefix (IP and depth).

The table lookup operation selects the IP prefix that is matched by the lookup key; in case of multiple matches, the entry with the longest prefix depth wins.

Typically used to implement IP routing tables.

3 Access Control List (ACLs)

Lookup key is 7-tuple of two VLAN/MPLS labels, IP destination address, IP source addresses, L4 protocol, L4 destination port, L4 source port.

Each table entry has an associated ACL and priority. The ACL contains bit masks for the VLAN/MPLS labels, IP prefix for IP destination address, IP prefix for IP source addresses, L4 protocol and bitmask, L4 destination port and bit mask, L4 source port and bit mask.

The table lookup operation selects the ACL that is matched by the lookup key; in case of multiple matches, the entry with the highest priority wins.

Typically used to implement rule databases for firewalls, etc.

4 Pattern matching search

Lookup key is the packet payload.

Table is a database of patterns, with each pattern having a priority assigned.

The table lookup operation selects the patterns that is matched by the input packet; in case of multiple matches, the matching pattern with the highest priority wins.

5 Array Lookup key is the table entry index itself.
Table Interface

Each table is required to implement an abstract interface that defines the initialization and run-time operation of the table. The table abstract interface is described in Table 36.4.

Table Abstract Interface
# Table operation Description
1 Create Create the low-level data structures of the lookup table. Can internally allocate memory.
2 Free Free up all the resources used by the lookup table.
3 Add entry Add new entry to the lookup table.
4 Delete entry Delete specific entry from the lookup table.
5 Lookup

Look up a burst of input packets and return a bit mask specifying the result of the lookup operation for each packet: a set bit signifies lookup hit for the corresponding packet, while a cleared bit a lookup miss.

For each lookup hit packet, the lookup operation also returns a pointer to the table entry that was hit, which contains the actions to be applied on the packet and any associated metadata.

For each lookup miss packet, the actions to be applied on the packet and any associated metadata are specified by the default table entry preconfigured for lookup miss.

Hash Table Design
Hash Table Overview

Hash tables are important because the key lookup operation is optimized for speed: instead of having to linearly search the lookup key through all the keys in the table, the search is limited to only the keys stored in a single table bucket.

Associative Arrays

An associative array is a function that can be specified as a set of (key, value) pairs, with each key from the possible set of input keys present at most once. For a given associative array, the possible operations are:

  1. add (key, value): When no value is currently associated with key, then the (key, value ) association is created. When key is already associated value value0, then the association (key, value0) is removed and association (key, value) is created;
  2. delete key: When no value is currently associated with key, this operation has no effect. When key is already associated value, then association (key, value) is removed;
  3. lookup key: When no value is currently associated with key, then this operation returns void value (lookup miss). When key is associated with value, then this operation returns value. The (key, value) association is not changed.

The matching criterion used to compare the input key against the keys in the associative array is exact match, as the key size (number of bytes) and the key value (array of bytes) have to match exactly for the two keys under comparison.

Hash Function

A hash function deterministically maps data of variable length (key) to data of fixed size (hash value or key signature). Typically, the size of the key is bigger than the size of the key signature. The hash function basically compresses a long key into a short signature. Several keys can share the same signature (collisions).

High quality hash functions have uniform distribution. For large number of keys, when dividing the space of signature values into a fixed number of equal intervals (buckets), it is desirable to have the key signatures evenly distributed across these intervals (uniform distribution), as opposed to most of the signatures going into only a few of the intervals and the rest of the intervals being largely unused (non-uniform distribution).

Hash Table

A hash table is an associative array that uses a hash function for its operation. The reason for using a hash function is to optimize the performance of the lookup operation by minimizing the number of table keys that have to be compared against the input key.

Instead of storing the (key, value) pairs in a single list, the hash table maintains multiple lists (buckets). For any given key, there is a single bucket where that key might exist, and this bucket is uniquely identified based on the key signature. Once the key signature is computed and the hash table bucket identified, the key is either located in this bucket or it is not present in the hash table at all, so the key search can be narrowed down from the full set of keys currently in the table to just the set of keys currently in the identified table bucket.

The performance of the hash table lookup operation is greatly improved, provided that the table keys are evenly distributed among the hash table buckets, which can be achieved by using a hash function with uniform distribution. The rule to map a key to its bucket can simply be to use the key signature (modulo the number of table buckets) as the table bucket ID:

bucket_id = f_hash(key) % n_buckets;

By selecting the number of buckets to be a power of two, the modulo operator can be replaced by a bitwise AND logical operation:

bucket_id = f_hash(key) & (n_buckets - 1);

considering n_bits as the number of bits set in bucket_mask = n_buckets - 1, this means that all the keys that end up in the same hash table bucket have the lower n_bits of their signature identical. In order to reduce the number of keys in the same bucket (collisions), the number of hash table buckets needs to be increased.

In packet processing context, the sequence of operations involved in hash table operations is described in Fig. 36.2:

_images/figure33.png

Sequence of Steps for Hash Table Operations in a Packet Processing Context

Hash Table Use Cases

Flow Classification

Description: The flow classification is executed at least once for each input packet. This operation maps each incoming packet against one of the known traffic flows in the flow database that typically contains millions of flows.

Hash table name: Flow classification table

Number of keys: Millions

Key format: n-tuple of packet fields that uniquely identify a traffic flow/connection. Example: DiffServ 5-tuple of (Source IP address, Destination IP address, L4 protocol, L4 protocol source port, L4 protocol destination port). For IPv4 protocol and L4 protocols like TCP, UDP or SCTP, the size of the DiffServ 5-tuple is 13 bytes, while for IPv6 it is 37 bytes.

Key value (key data): actions and action meta-data describing what processing to be applied for the packets of the current flow. The size of the data associated with each traffic flow can vary from 8 bytes to kilobytes.

Address Resolution Protocol (ARP)

Description: Once a route has been identified for an IP packet (so the output interface and the IP address of the next hop station are known), the MAC address of the next hop station is needed in order to send this packet onto the next leg of the journey towards its destination (as identified by its destination IP address). The MAC address of the next hop station becomes the destination MAC address of the outgoing Ethernet frame.

Hash table name: ARP table

Number of keys: Thousands

Key format: The pair of (Output interface, Next Hop IP address), which is typically 5 bytes for IPv4 and 17 bytes for IPv6.

Key value (key data): MAC address of the next hop station (6 bytes).

Hash Table Types

Table 36.5 lists the hash table configuration parameters shared by all different hash table types.

Configuration Parameters Common for All Hash Table Types
# Parameter Details
1 Key size Measured as number of bytes. All keys have the same size.
2 Key value (key data) size Measured as number of bytes.
3 Number of buckets Needs to be a power of two.
4 Maximum number of keys Needs to be a power of two.
5 Hash function Examples: jhash, CRC hash, etc.
6 Hash function seed Parameter to be passed to the hash function.
7 Key offset Offset of the lookup key byte array within the packet meta-data stored in the packet buffer.
Bucket Full Problem

On initialization, each hash table bucket is allocated space for exactly 4 keys. As keys are added to the table, it can happen that a given bucket already has 4 keys when a new key has to be added to this bucket. The possible options are:

  1. Least Recently Used (LRU) Hash Table. One of the existing keys in the bucket is deleted and the new key is added in its place. The number of keys in each bucket never grows bigger than 4. The logic to pick the key to be dropped from the bucket is LRU. The hash table lookup operation maintains the order in which the keys in the same bucket are hit, so every time a key is hit, it becomes the new Most Recently Used (MRU) key, i.e. the last candidate for drop. When a key is added to the bucket, it also becomes the new MRU key. When a key needs to be picked and dropped, the first candidate for drop, i.e. the current LRU key, is always picked. The LRU logic requires maintaining specific data structures per each bucket.
  2. Extendable Bucket Hash Table. The bucket is extended with space for 4 more keys. This is done by allocating additional memory at table initialization time, which is used to create a pool of free keys (the size of this pool is configurable and always a multiple of 4). On key add operation, the allocation of a group of 4 keys only happens successfully within the limit of free keys, otherwise the key add operation fails. On key delete operation, a group of 4 keys is freed back to the pool of free keys when the key to be deleted is the only key that was used within its group of 4 keys at that time. On key lookup operation, if the current bucket is in extended state and a match is not found in the first group of 4 keys, the search continues beyond the first group of 4 keys, potentially until all keys in this bucket are examined. The extendable bucket logic requires maintaining specific data structures per table and per each bucket.
Configuration Parameters Specific to Extendable Bucket Hash Table
# Parameter Details
1 Number of additional keys Needs to be a power of two, at least equal to 4.
Signature Computation

The possible options for key signature computation are:

  1. Pre-computed key signature. The key lookup operation is split between two CPU cores. The first CPU core (typically the CPU core that performs packet RX) extracts the key from the input packet, computes the key signature and saves both the key and the key signature in the packet buffer as packet meta-data. The second CPU core reads both the key and the key signature from the packet meta-data and performs the bucket search step of the key lookup operation.
  2. Key signature computed on lookup (“do-sig” version). The same CPU core reads the key from the packet meta-data, uses it to compute the key signature and also performs the bucket search step of the key lookup operation.
Configuration Parameters Specific to Pre-computed Key Signature Hash Table
# Parameter Details
1 Signature offset Offset of the pre-computed key signature within the packet meta-data.
Key Size Optimized Hash Tables

For specific key sizes, the data structures and algorithm of key lookup operation can be specially handcrafted for further performance improvements, so following options are possible:

  1. Implementation supporting configurable key size.
  2. Implementation supporting a single key size. Typical key sizes are 8 bytes and 16 bytes.
Bucket Search Logic for Configurable Key Size Hash Tables

The performance of the bucket search logic is one of the main factors influencing the performance of the key lookup operation. The data structures and algorithm are designed to make the best use of Intel CPU architecture resources like: cache memory space, cache memory bandwidth, external memory bandwidth, multiple execution units working in parallel, out of order instruction execution, special CPU instructions, etc.

The bucket search logic handles multiple input packets in parallel. It is built as a pipeline of several stages (3 or 4), with each pipeline stage handling two different packets from the burst of input packets. On each pipeline iteration, the packets are pushed to the next pipeline stage: for the 4-stage pipeline, two packets (that just completed stage 3) exit the pipeline, two packets (that just completed stage 2) are now executing stage 3, two packets (that just completed stage 1) are now executing stage 2, two packets (that just completed stage 0) are now executing stage 1 and two packets (next two packets to read from the burst of input packets) are entering the pipeline to execute stage 0. The pipeline iterations continue until all packets from the burst of input packets execute the last stage of the pipeline.

The bucket search logic is broken into pipeline stages at the boundary of the next memory access. Each pipeline stage uses data structures that are stored (with high probability) into the L1 or L2 cache memory of the current CPU core and breaks just before the next memory access required by the algorithm. The current pipeline stage finalizes by prefetching the data structures required by the next pipeline stage, so given enough time for the prefetch to complete, when the next pipeline stage eventually gets executed for the same packets, it will read the data structures it needs from L1 or L2 cache memory and thus avoid the significant penalty incurred by L2 or L3 cache memory miss.

By prefetching the data structures required by the next pipeline stage in advance (before they are used) and switching to executing another pipeline stage for different packets, the number of L2 or L3 cache memory misses is greatly reduced, hence one of the main reasons for improved performance. This is because the cost of L2/L3 cache memory miss on memory read accesses is high, as usually due to data dependency between instructions, the CPU execution units have to stall until the read operation is completed from L3 cache memory or external DRAM memory. By using prefetch instructions, the latency of memory read accesses is hidden, provided that it is preformed early enough before the respective data structure is actually used.

By splitting the processing into several stages that are executed on different packets (the packets from the input burst are interlaced), enough work is created to allow the prefetch instructions to complete successfully (before the prefetched data structures are actually accessed) and also the data dependency between instructions is loosened. For example, for the 4-stage pipeline, stage 0 is executed on packets 0 and 1 and then, before same packets 0 and 1 are used (i.e. before stage 1 is executed on packets 0 and 1), different packets are used: packets 2 and 3 (executing stage 1), packets 4 and 5 (executing stage 2) and packets 6 and 7 (executing stage 3). By executing useful work while the data structures are brought into the L1 or L2 cache memory, the latency of the read memory accesses is hidden. By increasing the gap between two consecutive accesses to the same data structure, the data dependency between instructions is loosened; this allows making the best use of the super-scalar and out-of-order execution CPU architecture, as the number of CPU core execution units that are active (rather than idle or stalled due to data dependency constraints between instructions) is maximized.

The bucket search logic is also implemented without using any branch instructions. This avoids the important cost associated with flushing the CPU core execution pipeline on every instance of branch misprediction.

Configurable Key Size Hash Table

Fig. 36.3, Table 36.8 and Table 36.9 detail the main data structures used to implement configurable key size hash tables (either LRU or extendable bucket, either with pre-computed signature or “do-sig”).

_images/figure34.png

Data Structures for Configurable Key Size Hash Tables

Main Large Data Structures (Arrays) used for Configurable Key Size Hash Tables
# Array name Number of entries Entry size (bytes) Description
1 Bucket array n_buckets (configurable) 32 Buckets of the hash table.
2 Bucket extensions array n_buckets_ext (configurable) 32 This array is only created for extendable bucket tables.
3 Key array n_keys key_size (configurable) Keys added to the hash table.
4 Data array n_keys entry_size (configurable) Key values (key data) associated with the hash table keys.
Field Description for Bucket Array Entry (Configurable Key Size Hash Tables)
# Field name Field size (bytes) Description
1 Next Ptr/LRU 8

For LRU tables, this fields represents the LRU list for the current bucket stored as array of 4 entries of 2 bytes each. Entry 0 stores the index (0 .. 3) of the MRU key, while entry 3 stores the index of the LRU key.

For extendable bucket tables, this field represents the next pointer (i.e. the pointer to the next group of 4 keys linked to the current bucket). The next pointer is not NULL if the bucket is currently extended or NULL otherwise. To help the branchless implementation, bit 0 (least significant bit) of this field is set to 1 if the next pointer is not NULL and to 0 otherwise.

2 Sig[0 .. 3] 4 x 2

If key X (X = 0 .. 3) is valid, then sig X bits 15 .. 1 store the most significant 15 bits of key X signature and sig X bit 0 is set to 1.

If key X is not valid, then sig X is set to zero.

3 Key Pos [0 .. 3] 4 x 4

If key X is valid (X = 0 .. 3), then Key Pos X represents the index into the key array where key X is stored, as well as the index into the data array where the value associated with key X is stored.

If key X is not valid, then the value of Key Pos X is undefined.

Fig. 36.4 and Table 36.10 detail the bucket search pipeline stages (either LRU or extendable bucket, either with pre-computed signature or “do-sig”). For each pipeline stage, the described operations are applied to each of the two packets handled by that stage.

_images/figure35.png

Bucket Search Pipeline for Key Lookup Operation (Configurable Key Size Hash Tables)

Description of the Bucket Search Pipeline Stages (Configurable Key Size Hash Tables)
# Stage name Description
0 Prefetch packet meta-data

Select next two packets from the burst of input packets.

Prefetch packet meta-data containing the key and key signature.

1 Prefetch table bucket

Read the key signature from the packet meta-data (for extendable bucket hash tables) or read the key from the packet meta-data and compute key signature (for LRU tables).

Identify the bucket ID using the key signature.

Set bit 0 of the signature to 1 (to match only signatures of valid keys from the table).

Prefetch the bucket.

2 Prefetch table key

Read the key signatures from the bucket.

Compare the signature of the input key against the 4 key signatures from the packet. As result, the following is obtained:

match = equal to TRUE if there was at least one signature match and to FALSE in the case of no signature match;

match_many = equal to TRUE is there were more than one signature matches (can be up to 4 signature matches in the worst case scenario) and to FALSE otherwise;

match_pos = the index of the first key that produced signature match (only valid if match is true).

For extendable bucket hash tables only, set match_many to TRUE if next pointer is valid.

Prefetch the bucket key indicated by match_pos (even if match_pos does not point to valid key valid).

3 Prefetch table data

Read the bucket key indicated by match_pos.

Compare the bucket key against the input key. As result, the following is obtained: match_key = equal to TRUE if the two keys match and to FALSE otherwise.

Report input key as lookup hit only when both match and match_key are equal to TRUE and as lookup miss otherwise.

For LRU tables only, use branchless logic to update the bucket LRU list (the current key becomes the new MRU) only on lookup hit.

Prefetch the key value (key data) associated with the current key (to avoid branches, this is done on both lookup hit and miss).

Additional notes:

  1. The pipelined version of the bucket search algorithm is executed only if there are at least 7 packets in the burst of input packets. If there are less than 7 packets in the burst of input packets, a non-optimized implementation of the bucket search algorithm is executed.
  2. Once the pipelined version of the bucket search algorithm has been executed for all the packets in the burst of input packets, the non-optimized implementation of the bucket search algorithm is also executed for any packets that did not produce a lookup hit, but have the match_many flag set. As result of executing the non-optimized version, some of these packets may produce a lookup hit or lookup miss. This does not impact the performance of the key lookup operation, as the probability of matching more than one signature in the same group of 4 keys or of having the bucket in extended state (for extendable bucket hash tables only) is relatively small.

Key Signature Comparison Logic

The key signature comparison logic is described in Table 36.11.

Lookup Tables for Match, Match_Many and Match_Pos
# mask match (1 bit) match_many (1 bit) match_pos (2 bits)
0 0000 0 0 00
1 0001 1 0 00
2 0010 1 0 01
3 0011 1 1 00
4 0100 1 0 10
5 0101 1 1 00
6 0110 1 1 01
7 0111 1 1 00
8 1000 1 0 11
9 1001 1 1 00
10 1010 1 1 01
11 1011 1 1 00
12 1100 1 1 10
13 1101 1 1 00
14 1110 1 1 01
15 1111 1 1 00

The input mask hash bit X (X = 0 .. 3) set to 1 if input signature is equal to bucket signature X and set to 0 otherwise. The outputs match, match_many and match_pos are 1 bit, 1 bit and 2 bits in size respectively and their meaning has been explained above.

As displayed in Table 36.12, the lookup tables for match and match_many can be collapsed into a single 32-bit value and the lookup table for match_pos can be collapsed into a 64-bit value. Given the input mask, the values for match, match_many and match_pos can be obtained by indexing their respective bit array to extract 1 bit, 1 bit and 2 bits respectively with branchless logic.

Collapsed Lookup Tables for Match, Match_Many and Match_Pos
  Bit array Hexadecimal value
match 1111_1111_1111_1110 0xFFFELLU
match_many 1111_1110_1110_1000 0xFEE8LLU
match_pos 0001_0010_0001_0011__0001_0010_0001_0000 0x12131210LLU

The pseudo-code for match, match_many and match_pos is:

match = (0xFFFELLU >> mask) & 1;

match_many = (0xFEE8LLU >> mask) & 1;

match_pos = (0x12131210LLU >> (mask << 1)) & 3;
Single Key Size Hash Tables

Fig. 36.5, Fig. 36.6, Table 36.13 and Table 36.14 detail the main data structures used to implement 8-byte and 16-byte key hash tables (either LRU or extendable bucket, either with pre-computed signature or “do-sig”).

_images/figure37.png

Data Structures for 8-byte Key Hash Tables

_images/figure38.png

Data Structures for 16-byte Key Hash Tables

Main Large Data Structures (Arrays) used for 8-byte and 16-byte Key Size Hash Tables
# Array name Number of entries Entry size (bytes) Description
1 Bucket array n_buckets (configurable)

8-byte key size:

64 + 4 x entry_size

16-byte key size:

128 + 4 x entry_size

Buckets of the hash table.
2 Bucket extensions array n_buckets_ext (configurable)

8-byte key size:

64 + 4 x entry_size

16-byte key size:

128 + 4 x entry_size

This array is only created for extendable bucket tables.
Field Description for Bucket Array Entry (8-byte and 16-byte Key Hash Tables)
# Field name Field size (bytes) Description
1 Valid 8

Bit X (X = 0 .. 3) is set to 1 if key X is valid or to 0 otherwise.

Bit 4 is only used for extendable bucket tables to help with the implementation of the branchless logic. In this case, bit 4 is set to 1 if next pointer is valid (not NULL) or to 0 otherwise.

2 Next Ptr/LRU 8

For LRU tables, this fields represents the LRU list for the current bucket stored as array of 4 entries of 2 bytes each. Entry 0 stores the index (0 .. 3) of the MRU key, while entry 3 stores the index of the LRU key.

For extendable bucket tables, this field represents the next pointer (i.e. the pointer to the next group of 4 keys linked to the current bucket). The next pointer is not NULL if the bucket is currently extended or NULL otherwise.

3 Key [0 .. 3] 4 x key_size Full keys.
4 Data [0 .. 3] 4 x entry_size Full key values (key data) associated with keys 0 .. 3.

and detail the bucket search pipeline used to implement 8-byte and 16-byte key hash tables (either LRU or extendable bucket, either with pre-computed signature or “do-sig”). For each pipeline stage, the described operations are applied to each of the two packets handled by that stage.

_images/figure39.png

Bucket Search Pipeline for Key Lookup Operation (Single Key Size Hash Tables)

Description of the Bucket Search Pipeline Stages (8-byte and 16-byte Key Hash Tables)
# Stage name Description
0 Prefetch packet meta-data
  1. Select next two packets from the burst of input packets.
  2. Prefetch packet meta-data containing the key and key signature.
1 Prefetch table bucket
  1. Read the key signature from the packet meta-data (for extendable bucket hash tables) or read the key from the packet meta-data and compute key signature (for LRU tables).
  2. Identify the bucket ID using the key signature.
  3. Prefetch the bucket.
2 Prefetch table data
  1. Read the bucket.
  2. Compare all 4 bucket keys against the input key.
  3. Report input key as lookup hit only when a match is identified (more than one key match is not possible)
  4. For LRU tables only, use branchless logic to update the bucket LRU list (the current key becomes the new MRU) only on lookup hit.
  5. Prefetch the key value (key data) associated with the matched key (to avoid branches, this is done on both lookup hit and miss).

Additional notes:

  1. The pipelined version of the bucket search algorithm is executed only if there are at least 5 packets in the burst of input packets. If there are less than 5 packets in the burst of input packets, a non-optimized implementation of the bucket search algorithm is executed.
  2. For extendable bucket hash tables only, once the pipelined version of the bucket search algorithm has been executed for all the packets in the burst of input packets, the non-optimized implementation of the bucket search algorithm is also executed for any packets that did not produce a lookup hit, but have the bucket in extended state. As result of executing the non-optimized version, some of these packets may produce a lookup hit or lookup miss. This does not impact the performance of the key lookup operation, as the probability of having the bucket in extended state is relatively small.

Pipeline Library Design

A pipeline is defined by:

  1. The set of input ports;
  2. The set of output ports;
  3. The set of tables;
  4. The set of actions.

The input ports are connected with the output ports through tree-like topologies of interconnected tables. The table entries contain the actions defining the operations to be executed on the input packets and the packet flow within the pipeline.

Connectivity of Ports and Tables

To avoid any dependencies on the order in which pipeline elements are created, the connectivity of pipeline elements is defined after all the pipeline input ports, output ports and tables have been created.

General connectivity rules:

  1. Each input port is connected to a single table. No input port should be left unconnected;
  2. The table connectivity to other tables or to output ports is regulated by the next hop actions of each table entry and the default table entry. The table connectivity is fluid, as the table entries and the default table entry can be updated during run-time.
    • A table can have multiple entries (including the default entry) connected to the same output port. A table can have different entries connected to different output ports. Different tables can have entries (including default table entry) connected to the same output port.
    • A table can have multiple entries (including the default entry) connected to another table, in which case all these entries have to point to the same table. This constraint is enforced by the API and prevents tree-like topologies from being created (allowing table chaining only), with the purpose of simplifying the implementation of the pipeline run-time execution engine.
Port Actions
Port Action Handler

An action handler can be assigned to each input/output port to define actions to be executed on each input packet that is received by the port. Defining the action handler for a specific input/output port is optional (i.e. the action handler can be disabled).

For input ports, the action handler is executed after RX function. For output ports, the action handler is executed before the TX function.

The action handler can decide to drop packets.

Table Actions
Table Action Handler

An action handler to be executed on each input packet can be assigned to each table. Defining the action handler for a specific table is optional (i.e. the action handler can be disabled).

The action handler is executed after the table lookup operation is performed and the table entry associated with each input packet is identified. The action handler can only handle the user-defined actions, while the reserved actions (e.g. the next hop actions) are handled by the Packet Framework. The action handler can decide to drop the input packet.

Reserved Actions

The reserved actions are handled directly by the Packet Framework without the user being able to change their meaning through the table action handler configuration. A special category of the reserved actions is represented by the next hop actions, which regulate the packet flow between input ports, tables and output ports through the pipeline. Table 36.16 lists the next hop actions.

Next Hop Actions (Reserved)
# Next hop action Description
1 Drop Drop the current packet.
2 Send to output port Send the current packet to specified output port. The output port ID is metadata stored in the same table entry.
3 Send to table Send the current packet to specified table. The table ID is metadata stored in the same table entry.
User Actions

For each table, the meaning of user actions is defined through the configuration of the table action handler. Different tables can be configured with different action handlers, therefore the meaning of the user actions and their associated meta-data is private to each table. Within the same table, all the table entries (including the table default entry) share the same definition for the user actions and their associated meta-data, with each table entry having its own set of enabled user actions and its own copy of the action meta-data. Table 36.17 contains a non-exhaustive list of user action examples.

User Action Examples
# User action Description
1 Metering Per flow traffic metering using the srTCM and trTCM algorithms.
2 Statistics Update the statistics counters maintained per flow.
3 App ID Per flow state machine fed by variable length sequence of packets at the flow initialization with the purpose of identifying the traffic type and application.
4 Push/pop labels Push/pop VLAN/MPLS labels to/from the current packet.
5 Network Address Translation (NAT) Translate between the internal (LAN) and external (WAN) IP destination/source address and/or L4 protocol destination/source port.
6 TTL update Decrement IP TTL and, in case of IPv4 packets, update the IP checksum.

Multicore Scaling

A complex application is typically split across multiple cores, with cores communicating through SW queues. There is usually a performance limit on the number of table lookups and actions that can be fitted on the same CPU core due to HW constraints like: available CPU cycles, cache memory size, cache transfer BW, memory transfer BW, etc.

As the application is split across multiple CPU cores, the Packet Framework facilitates the creation of several pipelines, the assignment of each such pipeline to a different CPU core and the interconnection of all CPU core-level pipelines into a single application-level complex pipeline. For example, if CPU core A is assigned to run pipeline P1 and CPU core B pipeline P2, then the interconnection of P1 with P2 could be achieved by having the same set of SW queues act like output ports for P1 and input ports for P2.

This approach enables the application development using the pipeline, run-to-completion (clustered) or hybrid (mixed) models.

It is allowed for the same core to run several pipelines, but it is not allowed for several cores to run the same pipeline.

Shared Data Structures

The threads performing table lookup are actually table writers rather than just readers. Even if the specific table lookup algorithm is thread-safe for multiple readers (e. g. read-only access of the search algorithm data structures is enough to conduct the lookup operation), once the table entry for the current packet is identified, the thread is typically expected to update the action meta-data stored in the table entry (e.g. increment the counter tracking the number of packets that hit this table entry), and thus modify the table entry. During the time this thread is accessing this table entry (either writing or reading; duration is application specific), for data consistency reasons, no other threads (threads performing table lookup or entry add/delete operations) are allowed to modify this table entry.

Mechanisms to share the same table between multiple threads:

  1. Multiple writer threads. Threads need to use synchronization primitives like semaphores (distinct semaphore per table entry) or atomic instructions. The cost of semaphores is usually high, even when the semaphore is free. The cost of atomic instructions is normally higher than the cost of regular instructions.
  2. Multiple writer threads, with single thread performing table lookup operations and multiple threads performing table entry add/delete operations. The threads performing table entry add/delete operations send table update requests to the reader (typically through message passing queues), which does the actual table updates and then sends the response back to the request initiator.
  3. Single writer thread performing table entry add/delete operations and multiple reader threads that perform table lookup operations with read-only access to the table entries. The reader threads use the main table copy while the writer is updating the mirror copy. Once the writer update is done, the writer can signal to the readers and busy wait until all readers swaps between the mirror copy (which now becomes the main copy) and the mirror copy (which now becomes the main copy).

Interfacing with Accelerators

The presence of accelerators is usually detected during the initialization phase by inspecting the HW devices that are part of the system (e.g. by PCI bus enumeration). Typical devices with acceleration capabilities are:

  • Inline accelerators: NICs, switches, FPGAs, etc;
  • Look-aside accelerators: chipsets, FPGAs, etc.

Usually, to support a specific functional block, specific implementation of Packet Framework tables and/or ports and/or actions has to be provided for each accelerator, with all the implementations sharing the same API: pure SW implementation (no acceleration), implementation using accelerator A, implementation using accelerator B, etc. The selection between these implementations could be done at build time or at run-time (recommended), based on which accelerators are present in the system, with no application changes required.

Vhost Library

The vhost library implements a user space virtio net server allowing the user to manipulate the virtio ring directly. In another words, it allows the user to fetch/put packets from/to the VM virtio net device. To achieve this, a vhost library should be able to:

  • Access the guest memory:

    For QEMU, this is done by using the -object memory-backend-file,share=on,... option. Which means QEMU will create a file to serve as the guest RAM. The share=on option allows another process to map that file, which means it can access the guest RAM.

  • Know all the necessary information about the vring:

    Information such as where the available ring is stored. Vhost defines some messages (passed through a Unix domain socket file) to tell the backend all the information it needs to know how to manipulate the vring.

Vhost API Overview

The following is an overview of some key Vhost API functions:

  • rte_vhost_driver_register(path, flags)

    This function registers a vhost driver into the system. path specifies the Unix domain socket file path.

    Currently supported flags are:

    • RTE_VHOST_USER_CLIENT

      DPDK vhost-user will act as the client when this flag is given. See below for an explanation.

    • RTE_VHOST_USER_NO_RECONNECT

      When DPDK vhost-user acts as the client it will keep trying to reconnect to the server (QEMU) until it succeeds. This is useful in two cases:

      • When QEMU is not started yet.
      • When QEMU restarts (for example due to a guest OS reboot).

      This reconnect option is enabled by default. However, it can be turned off by setting this flag.

    • RTE_VHOST_USER_DEQUEUE_ZERO_COPY

      Dequeue zero copy will be enabled when this flag is set. It is disabled by default.

      There are some truths (including limitations) you might want to know while setting this flag:

      • zero copy is not good for small packets (typically for packet size below 512).

      • zero copy is really good for VM2VM case. For iperf between two VMs, the boost could be above 70% (when TSO is enableld).

      • for VM2NIC case, the nb_tx_desc has to be small enough: <= 64 if virtio indirect feature is not enabled and <= 128 if it is enabled.

        This is because when dequeue zero copy is enabled, guest Tx used vring will be updated only when corresponding mbuf is freed. Thus, the nb_tx_desc has to be small enough so that the PMD driver will run out of available Tx descriptors and free mbufs timely. Otherwise, guest Tx vring would be starved.

      • Guest memory should be backended with huge pages to achieve better performance. Using 1G page size is the best.

        When dequeue zero copy is enabled, the guest phys address and host phys address mapping has to be established. Using non-huge pages means far more page segments. To make it simple, DPDK vhost does a linear search of those segments, thus the fewer the segments, the quicker we will get the mapping. NOTE: we may speed it by using tree searching in future.

    • RTE_VHOST_USER_IOMMU_SUPPORT

      IOMMU support will be enabled when this flag is set. It is disabled by default.

      Enabling this flag makes possible to use guest vIOMMU to protect vhost from accessing memory the virtio device isn’t allowed to, when the feature is negotiated and an IOMMU device is declared.

      However, this feature enables vhost-user’s reply-ack protocol feature, which implementation is buggy in Qemu v2.7.0-v2.9.0 when doing multiqueue. Enabling this flag with these Qemu version results in Qemu being blocked when multiple queue pairs are declared.

  • rte_vhost_driver_set_features(path, features)

    This function sets the feature bits the vhost-user driver supports. The vhost-user driver could be vhost-user net, yet it could be something else, say, vhost-user SCSI.

  • rte_vhost_driver_callback_register(path, vhost_device_ops)

    This function registers a set of callbacks, to let DPDK applications take the appropriate action when some events happen. The following events are currently supported:

    • new_device(int vid)

      This callback is invoked when a virtio device becomes ready. vid is the vhost device ID.

    • destroy_device(int vid)

      This callback is invoked when a virtio device is paused or shut down.

    • vring_state_changed(int vid, uint16_t queue_id, int enable)

      This callback is invoked when a specific queue’s state is changed, for example to enabled or disabled.

    • features_changed(int vid, uint64_t features)

      This callback is invoked when the features is changed. For example, VHOST_F_LOG_ALL will be set/cleared at the start/end of live migration, respectively.

    • new_connection(int vid)

      This callback is invoked on new vhost-user socket connection. If DPDK acts as the server the device should not be deleted before destroy_connection callback is received.

    • destroy_connection(int vid)

      This callback is invoked when vhost-user socket connection is closed. It indicates that device with id vid is no longer in use and can be safely deleted.

  • rte_vhost_driver_disable/enable_features(path, features))

    This function disables/enables some features. For example, it can be used to disable mergeable buffers and TSO features, which both are enabled by default.

  • rte_vhost_driver_start(path)

    This function triggers the vhost-user negotiation. It should be invoked at the end of initializing a vhost-user driver.

  • rte_vhost_enqueue_burst(vid, queue_id, pkts, count)

    Transmits (enqueues) count packets from host to guest.

  • rte_vhost_dequeue_burst(vid, queue_id, mbuf_pool, pkts, count)

    Receives (dequeues) count packets from guest, and stored them at pkts.

Vhost-user Implementations

Vhost-user uses Unix domain sockets for passing messages. This means the DPDK vhost-user implementation has two options:

  • DPDK vhost-user acts as the server.

    DPDK will create a Unix domain socket server file and listen for connections from the frontend.

    Note, this is the default mode, and the only mode before DPDK v16.07.

  • DPDK vhost-user acts as the client.

    Unlike the server mode, this mode doesn’t create the socket file; it just tries to connect to the server (which responses to create the file instead).

    When the DPDK vhost-user application restarts, DPDK vhost-user will try to connect to the server again. This is how the “reconnect” feature works.

    Note

    • The “reconnect” feature requires QEMU v2.7 (or above).
    • The vhost supported features must be exactly the same before and after the restart. For example, if TSO is disabled and then enabled, nothing will work and issues undefined might happen.

No matter which mode is used, once a connection is established, DPDK vhost-user will start receiving and processing vhost messages from QEMU.

For messages with a file descriptor, the file descriptor can be used directly in the vhost process as it is already installed by the Unix domain socket.

The supported vhost messages are:

  • VHOST_SET_MEM_TABLE
  • VHOST_SET_VRING_KICK
  • VHOST_SET_VRING_CALL
  • VHOST_SET_LOG_FD
  • VHOST_SET_VRING_ERR

For VHOST_SET_MEM_TABLE message, QEMU will send information for each memory region and its file descriptor in the ancillary data of the message. The file descriptor is used to map that region.

VHOST_SET_VRING_KICK is used as the signal to put the vhost device into the data plane, and VHOST_GET_VRING_BASE is used as the signal to remove the vhost device from the data plane.

When the socket connection is closed, vhost will destroy the device.

Vhost supported vSwitch reference

For more vhost details and how to support vhost in vSwitch, please refer to the vhost example in the DPDK Sample Applications Guide.

Metrics Library

The Metrics library implements a mechanism by which producers can publish numeric information for later querying by consumers. In practice producers will typically be other libraries or primary processes, whereas consumers will typically be applications.

Metrics themselves are statistics that are not generated by PMDs. Metric information is populated using a push model, where producers update the values contained within the metric library by calling an update function on the relevant metrics. Consumers receive metric information by querying the central metric data, which is held in shared memory.

For each metric, a separate value is maintained for each port id, and when publishing metric values the producers need to specify which port is being updated. In addition there is a special id RTE_METRICS_GLOBAL that is intended for global statistics that are not associated with any individual device. Since the metrics library is self-contained, the only restriction on port numbers is that they are less than RTE_MAX_ETHPORTS - there is no requirement for the ports to actually exist.

Initialising the library

Before the library can be used, it has to be initialized by calling rte_metrics_init() which sets up the metric store in shared memory. This is where producers will publish metric information to, and where consumers will query it from.

rte_metrics_init(rte_socket_id());

This function must be called from a primary process, but otherwise producers and consumers can be in either primary or secondary processes.

Registering metrics

Metrics must first be registered, which is the way producers declare the names of the metrics they will be publishing. Registration can either be done individually, or a set of metrics can be registered as a group. Individual registration is done using rte_metrics_reg_name():

id_1 = rte_metrics_reg_name("mean_bits_in");
id_2 = rte_metrics_reg_name("mean_bits_out");
id_3 = rte_metrics_reg_name("peak_bits_in");
id_4 = rte_metrics_reg_name("peak_bits_out");

or alternatively, a set of metrics can be registered together using rte_metrics_reg_names():

const char * const names[] = {
    "mean_bits_in", "mean_bits_out",
    "peak_bits_in", "peak_bits_out",
};
id_set = rte_metrics_reg_names(&names[0], 4);

If the return value is negative, it means registration failed. Otherwise the return value is the key for the metric, which is used when updating values. A table mapping together these key values and the metrics’ names can be obtained using rte_metrics_get_names().

Updating metric values

Once registered, producers can update the metric for a given port using the rte_metrics_update_value() function. This uses the metric key that is returned when registering the metric, and can also be looked up using rte_metrics_get_names().

rte_metrics_update_value(port_id, id_1, values[0]);
rte_metrics_update_value(port_id, id_2, values[1]);
rte_metrics_update_value(port_id, id_3, values[2]);
rte_metrics_update_value(port_id, id_4, values[3]);

if metrics were registered as a single set, they can either be updated individually using rte_metrics_update_value(), or updated together using the rte_metrics_update_values() function:

rte_metrics_update_value(port_id, id_set, values[0]);
rte_metrics_update_value(port_id, id_set + 1, values[1]);
rte_metrics_update_value(port_id, id_set + 2, values[2]);
rte_metrics_update_value(port_id, id_set + 3, values[3]);

rte_metrics_update_values(port_id, id_set, values, 4);

Note that rte_metrics_update_values() cannot be used to update metric values from multiple sets, as there is no guarantee two sets registered one after the other have contiguous id values.

Querying metrics

Consumers can obtain metric values by querying the metrics library using the rte_metrics_get_values() function that return an array of struct rte_metric_value. Each entry within this array contains a metric value and its associated key. A key-name mapping can be obtained using the rte_metrics_get_names() function that returns an array of struct rte_metric_name that is indexed by the key. The following will print out all metrics for a given port:

void print_metrics() {
    struct rte_metric_value *metrics;
    struct rte_metric_name *names;
    int len;

    len = rte_metrics_get_names(NULL, 0);
    if (len < 0) {
        printf("Cannot get metrics count\n");
        return;
    }
    if (len == 0) {
        printf("No metrics to display (none have been registered)\n");
        return;
    }
    metrics = malloc(sizeof(struct rte_metric_value) * len);
    names =  malloc(sizeof(struct rte_metric_name) * len);
    if (metrics == NULL || names == NULL) {
        printf("Cannot allocate memory\n");
        free(metrics);
        free(names);
        return;
    }
    ret = rte_metrics_get_values(port_id, metrics, len);
    if (ret < 0 || ret > len) {
        printf("Cannot get metrics values\n");
        free(metrics);
        free(names);
        return;
    }
    printf("Metrics for port %i:\n", port_id);
    for (i = 0; i < len; i++)
        printf("  %s: %"PRIu64"\n",
            names[metrics[i].key].name, metrics[i].value);
    free(metrics);
    free(names);
}

Bit-rate statistics library

The bit-rate library calculates the exponentially-weighted moving average and peak bit-rates for each active port (i.e. network device). These statistics are reported via the metrics library using the following names:

  • mean_bits_in: Average inbound bit-rate
  • mean_bits_out: Average outbound bit-rate
  • ewma_bits_in: Average inbound bit-rate (EWMA smoothed)
  • ewma_bits_out: Average outbound bit-rate (EWMA smoothed)
  • peak_bits_in: Peak inbound bit-rate
  • peak_bits_out: Peak outbound bit-rate

Once initialised and clocked at the appropriate frequency, these statistics can be obtained by querying the metrics library.

Initialization

Before the library can be used, it has to be initialised by calling rte_stats_bitrate_create(), which will return a bit-rate calculation object. Since the bit-rate library uses the metrics library to report the calculated statistics, the bit-rate library then needs to register the calculated statistics with the metrics library. This is done using the helper function rte_stats_bitrate_reg().

struct rte_stats_bitrates *bitrate_data;

bitrate_data = rte_stats_bitrate_create();
if (bitrate_data == NULL)
    rte_exit(EXIT_FAILURE, "Could not allocate bit-rate data.\n");
rte_stats_bitrate_reg(bitrate_data);
Controlling the sampling rate

Since the library works by periodic sampling but does not use an internal thread, the application has to periodically call rte_stats_bitrate_calc(). The frequency at which this function is called should be the intended sampling rate required for the calculated statistics. For instance if per-second statistics are desired, this function should be called once a second.

tics_datum = rte_rdtsc();
tics_per_1sec = rte_get_timer_hz();

while( 1 ) {
    /* ... */
    tics_current = rte_rdtsc();
    if (tics_current - tics_datum >= tics_per_1sec) {
        /* Periodic bitrate calculation */
        for (idx_port = 0; idx_port < cnt_ports; idx_port++)
                rte_stats_bitrate_calc(bitrate_data, idx_port);
            tics_datum = tics_current;
        }
    /* ... */
}

Latency statistics library

The latency statistics library calculates the latency of packet processing by a DPDK application, reporting the minimum, average, and maximum nano-seconds that packet processing takes, as well as the jitter in processing delay. These statistics are then reported via the metrics library using the following names:

  • min_latency_ns: Minimum processing latency (nano-seconds)
  • avg_latency_ns: Average processing latency (nano-seconds)
  • mac_latency_ns: Maximum processing latency (nano-seconds)
  • jitter_ns: Variance in processing latency (nano-seconds)

Once initialised and clocked at the appropriate frequency, these statistics can be obtained by querying the metrics library.

Initialization

Before the library can be used, it has to be initialised by calling rte_latencystats_init().

lcoreid_t latencystats_lcore_id = -1;

int ret = rte_latencystats_init(1, NULL);
if (ret)
    rte_exit(EXIT_FAILURE, "Could not allocate latency data.\n");
Triggering statistic updates

The rte_latencystats_update() function needs to be called periodically so that latency statistics can be updated.

if (latencystats_lcore_id == rte_lcore_id())
    rte_latencystats_update();
Library shutdown

When finished, rte_latencystats_uninit() needs to be called to de-initialise the latency library.

rte_latencystats_uninit();

Port Hotplug Framework

The Port Hotplug Framework provides DPDK applications with the ability to attach and detach ports at runtime. Because the framework depends on PMD implementation, the ports that PMDs cannot handle are out of scope of this framework. Furthermore, after detaching a port from a DPDK application, the framework doesn’t provide a way for removing the devices from the system. For the ports backed by a physical NIC, the kernel will need to support PCI Hotplug feature.

Overview

The basic requirements of the Port Hotplug Framework are:

  • DPDK applications that use the Port Hotplug Framework must manage their own ports.

    The Port Hotplug Framework is implemented to allow DPDK applications to manage ports. For example, when DPDK applications call the port attach function, the attached port number is returned. DPDK applications can also detach the port by port number.

  • Kernel support is needed for attaching or detaching physical device ports.

    To attach new physical device ports, the device will be recognized by userspace driver I/O framework in kernel at first. Then DPDK applications can call the Port Hotplug functions to attach the ports. For detaching, steps are vice versa.

  • Before detaching, they must be stopped and closed.

    DPDK applications must call “rte_eth_dev_stop()” and “rte_eth_dev_close()” APIs before detaching ports. These functions will start finalization sequence of the PMDs.

  • The framework doesn’t affect legacy DPDK applications behavior.

    If the Port Hotplug functions aren’t called, all legacy DPDK apps can still work without modifications.

Port Hotplug API overview

  • Attaching a port

    “rte_eth_dev_attach()” API attaches a port to DPDK application, and returns the attached port number. Before calling the API, the device should be recognized by an userspace driver I/O framework. The API receives a pci address like “0000:01:00.0” or a virtual device name like “net_pcap0,iface=eth0”. In the case of virtual device name, the format is the same as the general “–vdev” option of DPDK.

  • Detaching a port

    “rte_eth_dev_detach()” API detaches a port from DPDK application, and returns a pci address of the detached device or a virtual device name of the device.

Reference

“testpmd” supports the Port Hotplug Framework.

Limitations

  • The Port Hotplug APIs are not thread safe.
  • The framework can only be enabled with Linux. BSD is not supported.
  • Not all PMDs support detaching feature. The underlying bus must support hot-unplug. If not supported, the function rte_eth_dev_detach() will return negative ENOTSUP.

Part 2: Development Environment

Source Organization

This section describes the organization of sources in the DPDK framework.

Makefiles and Config

Note

In the following descriptions, RTE_SDK is the environment variable that points to the base directory into which the tarball was extracted. See Useful Variables Provided by the Build System for descriptions of other variables.

Makefiles that are provided by the DPDK libraries and applications are located in $(RTE_SDK)/mk.

Config templates are located in $(RTE_SDK)/config. The templates describe the options that are enabled for each target. The config file also contains items that can be enabled and disabled for many of the DPDK libraries, including debug options. The user should look at the config file and become familiar with these options. The config file is also used to create a header file, which will be located in the new build directory.

Libraries

Libraries are located in subdirectories of $(RTE_SDK)/lib. By convention a library refers to any code that provides an API to an application. Typically, it generates an archive file (.a), but a kernel module would also go in the same directory.

The lib directory contains:

lib
+-- librte_cmdline      # Command line interface helper
+-- librte_distributor  # Packet distributor
+-- librte_eal          # Environment abstraction layer
+-- librte_ether        # Generic interface to poll mode driver
+-- librte_hash         # Hash library
+-- librte_ip_frag      # IP fragmentation library
+-- librte_kni          # Kernel NIC interface
+-- librte_kvargs       # Argument parsing library
+-- librte_lpm          # Longest prefix match library
+-- librte_mbuf         # Packet and control mbuf manipulation
+-- librte_mempool      # Memory pool manager (fixed sized objects)
+-- librte_meter        # QoS metering library
+-- librte_net          # Various IP-related headers
+-- librte_power        # Power management library
+-- librte_ring         # Software rings (act as lockless FIFOs)
+-- librte_sched        # QoS scheduler and dropper library
+-- librte_timer        # Timer library

Drivers

Drivers are special libraries which provide poll-mode driver implementations for devices: either hardware devices or pseudo/virtual devices. They are contained in the drivers subdirectory, classified by type, and each compiles to a library with the format librte_pmd_X.a where X is the driver name.

The drivers directory has a net subdirectory which contains:

drivers/net
+-- af_packet          # Poll mode driver based on Linux af_packet
+-- bonding            # Bonding poll mode driver
+-- cxgbe              # Chelsio Terminator 10GbE/40GbE poll mode driver
+-- e1000              # 1GbE poll mode drivers (igb and em)
+-- enic               # Cisco VIC Ethernet NIC Poll-mode Driver
+-- fm10k              # Host interface PMD driver for FM10000 Series
+-- i40e               # 40GbE poll mode driver
+-- ixgbe              # 10GbE poll mode driver
+-- mlx4               # Mellanox ConnectX-3 poll mode driver
+-- null               # NULL poll mode driver for testing
+-- pcap               # PCAP poll mode driver
+-- ring               # Ring poll mode driver
+-- szedata2           # SZEDATA2 poll mode driver
+-- virtio             # Virtio poll mode driver
+-- vmxnet3            # VMXNET3 poll mode driver

Note

Several of the driver/net directories contain a base sub-directory. The base directory generally contains code the shouldn’t be modified directly by the user. Any enhancements should be done via the X_osdep.c and/or X_osdep.h files in that directory. Refer to the local README in the base directories for driver specific instructions.

Applications

Applications are source files that contain a main() function. They are located in the $(RTE_SDK)/app and $(RTE_SDK)/examples directories.

The app directory contains sample applications that are used to test DPDK (such as autotests) or the Poll Mode Drivers (test-pmd):

app
+-- chkincs            # Test program to check include dependencies
+-- cmdline_test       # Test the commandline library
+-- test               # Autotests to validate DPDK features
+-- test-acl           # Test the ACL library
+-- test-pipeline      # Test the IP Pipeline framework
+-- test-pmd           # Test and benchmark poll mode drivers

The examples directory contains sample applications that show how libraries can be used:

examples
+-- cmdline            # Example of using the cmdline library
+-- exception_path     # Sending packets to and from Linux TAP device
+-- helloworld         # Basic Hello World example
+-- ip_reassembly      # Example showing IP reassembly
+-- ip_fragmentation   # Example showing IPv4 fragmentation
+-- ipv4_multicast     # Example showing IPv4 multicast
+-- kni                # Kernel NIC Interface (KNI) example
+-- l2fwd              # L2 forwarding with and without SR-IOV
+-- l3fwd              # L3 forwarding example
+-- l3fwd-power        # L3 forwarding example with power management
+-- l3fwd-vf           # L3 forwarding example with SR-IOV
+-- link_status_interrupt # Link status change interrupt example
+-- load_balancer      # Load balancing across multiple cores/sockets
+-- multi_process      # Example apps using multiple DPDK processes
+-- qos_meter          # QoS metering example
+-- qos_sched          # QoS scheduler and dropper example
+-- timer              # Example of using librte_timer library
+-- vmdq_dcb           # Example of VMDQ and DCB receiving
+-- vmdq               # Example of VMDQ receiving
+-- vhost              # Example of userspace vhost and switch

Note

The actual examples directory may contain additional sample applications to those shown above. Check the latest DPDK source files for details.

Development Kit Build System

The DPDK requires a build system for compilation activities and so on. This section describes the constraints and the mechanisms used in the DPDK framework.

There are two use-cases for the framework:

  • Compilation of the DPDK libraries and sample applications; the framework generates specific binary libraries, include files and sample applications
  • Compilation of an external application or library, using an installed binary DPDK

Building the Development Kit Binary

The following provides details on how to build the DPDK binary.

Build Directory Concept

After installation, a build directory structure is created. Each build directory contains include files, libraries, and applications.

A build directory is specific to a configuration that includes architecture + execution environment + toolchain. It is possible to have several build directories sharing the same sources with different configurations.

For instance, to create a new build directory called my_sdk_build_dir using the default configuration template config/defconfig_x86_64-linuxapp, we use:

cd ${RTE_SDK}
make config T=x86_64-native-linuxapp-gcc O=my_sdk_build_dir

This creates a new my_sdk_build_dir directory. After that, we can compile by doing:

cd my_sdk_build_dir
make

which is equivalent to:

make O=my_sdk_build_dir

The content of the my_sdk_build_dir is then:

-- .config                         # used configuration

-- Makefile                        # wrapper that calls head Makefile
                                   # with $PWD as build directory


    -- build                              #All temporary files used during build
    +--app                                # process, including . o, .d, and .cmd files.
        |  +-- test                       # For libraries, we have the .a file.
        |  +-- test.o                     # For applications, we have the elf file.
        |  `-- ...
        +-- lib
            +-- librte_eal
            |   `-- ...
            +-- librte_mempool
            |  +--  mempool-file1.o
            |  +--  .mempool-file1.o.cmd
            |  +--  .mempool-file1.o.d
            |  +--   mempool-file2.o
            |  +--  .mempool-file2.o.cmd
            |  +--  .mempool-file2.o.d
            |  `--  mempool.a
            `-- ...

-- include                # All include files installed by libraries
    +-- librte_mempool.h  # and applications are located in this
    +-- rte_eal.h         # directory. The installed files can depend
    +-- rte_spinlock.h    # on configuration if needed (environment,
    +-- rte_atomic.h      # architecture, ..)
    `-- \*.h ...

-- lib                    # all compiled libraries are copied in this
    +-- librte_eal.a      # directory
    +-- librte_mempool.a
    `-- \*.a ...

-- app                    # All compiled applications are installed
+ --test                  # here. It includes the binary in elf format

Refer to Development Kit Root Makefile Help for details about make commands that can be used from the root of DPDK.

Building External Applications

Since DPDK is in essence a development kit, the first objective of end users will be to create an application using this SDK. To compile an application, the user must set the RTE_SDK and RTE_TARGET environment variables.

export RTE_SDK=/opt/DPDK
export RTE_TARGET=x86_64-native-linuxapp-gcc
cd /path/to/my_app

For a new application, the user must create their own Makefile that includes some .mk files, such as ${RTE_SDK}/mk/rte.vars.mk, and ${RTE_SDK}/mk/ rte.app.mk. This is described in Building Your Own Application.

Depending on the chosen target (architecture, machine, executive environment, toolchain) defined in the Makefile or as an environment variable, the applications and libraries will compile using the appropriate .h files and will link with the appropriate .a files. These files are located in ${RTE_SDK}/arch-machine-execenv-toolchain, which is referenced internally by ${RTE_BIN_SDK}.

To compile their application, the user just has to call make. The compilation result will be located in /path/to/my_app/build directory.

Sample applications are provided in the examples directory.

Makefile Description

General Rules For DPDK Makefiles

In the DPDK, Makefiles always follow the same scheme:

  1. Include $(RTE_SDK)/mk/rte.vars.mk at the beginning.

  2. Define specific variables for RTE build system.

  3. Include a specific $(RTE_SDK)/mk/rte.XYZ.mk, where XYZ can be app, lib, extapp, extlib, obj, gnuconfigure, and so on, depending on what kind of object you want to build. See Makefile Types below.

  4. Include user-defined rules and variables.

    The following is a very simple example of an external application Makefile:

    include $(RTE_SDK)/mk/rte.vars.mk
    
    # binary name
    APP = helloworld
    
    # all source are stored in SRCS-y
    SRCS-y := main.c
    
    CFLAGS += -O3
    CFLAGS += $(WERROR_FLAGS)
    
    include $(RTE_SDK)/mk/rte.extapp.mk
    
Makefile Types

Depending on the .mk file which is included at the end of the user Makefile, the Makefile will have a different role. Note that it is not possible to build a library and an application in the same Makefile. For that, the user must create two separate Makefiles, possibly in two different directories.

In any case, the rte.vars.mk file must be included in the user Makefile as soon as possible.

Application

These Makefiles generate a binary application.

  • rte.app.mk: Application in the development kit framework
  • rte.extapp.mk: External application
  • rte.hostapp.mk: prerequisite tool to build dpdk
Library

Generate a .a library.

  • rte.lib.mk: Library in the development kit framework
  • rte.extlib.mk: external library
  • rte.hostlib.mk: host library in the development kit framework
Install
  • rte.install.mk: Does not build anything, it is only used to create links or copy files to the installation directory. This is useful for including files in the development kit framework.
Kernel Module
  • rte.module.mk: Build a kernel module in the development kit framework.
Objects
  • rte.obj.mk: Object aggregation (merge several .o in one) in the development kit framework.
  • rte.extobj.mk: Object aggregation (merge several .o in one) outside the development kit framework.
Misc
  • rte.doc.mk: Documentation in the development kit framework
  • rte.gnuconfigure.mk: Build an application that is configure-based.
  • rte.subdir.mk: Build several directories in the development kit framework.
Internally Generated Build Tools

app/dpdk-pmdinfogen

dpdk-pmdinfogen scans an object (.o) file for various well known symbol names. These well known symbol names are defined by various macros and used to export important information about hardware support and usage for pmd files. For instance the macro:

RTE_PMD_REGISTER_PCI(name, drv)

Creates the following symbol:

static char this_pmd_name0[] __attribute__((used)) = "<name>";

Which dpdk-pmdinfogen scans for. Using this information other relevant bits of data can be exported from the object file and used to produce a hardware support description, that dpdk-pmdinfogen then encodes into a json formatted string in the following format:

static char <name_pmd_string>="PMD_INFO_STRING=\"{'name' : '<name>', ...}\"";

These strings can then be searched for by external tools to determine the hardware support of a given library or application.

Useful Variables Provided by the Build System
  • RTE_SDK: The absolute path to the DPDK sources. When compiling the development kit, this variable is automatically set by the framework. It has to be defined by the user as an environment variable if compiling an external application.
  • RTE_SRCDIR: The path to the root of the sources. When compiling the development kit, RTE_SRCDIR = RTE_SDK. When compiling an external application, the variable points to the root of external application sources.
  • RTE_OUTPUT: The path to which output files are written. Typically, it is $(RTE_SRCDIR)/build, but it can be overridden by the O= option in the make command line.
  • RTE_TARGET: A string identifying the target for which we are building. The format is arch-machine-execenv-toolchain. When compiling the SDK, the target is deduced by the build system from the configuration (.config). When building an external application, it must be specified by the user in the Makefile or as an environment variable.
  • RTE_SDK_BIN: References $(RTE_SDK)/$(RTE_TARGET).
  • RTE_ARCH: Defines the architecture (i686, x86_64). It is the same value as CONFIG_RTE_ARCH but without the double-quotes around the string.
  • RTE_MACHINE: Defines the machine. It is the same value as CONFIG_RTE_MACHINE but without the double-quotes around the string.
  • RTE_TOOLCHAIN: Defines the toolchain (gcc , icc). It is the same value as CONFIG_RTE_TOOLCHAIN but without the double-quotes around the string.
  • RTE_EXEC_ENV: Defines the executive environment (linuxapp). It is the same value as CONFIG_RTE_EXEC_ENV but without the double-quotes around the string.
  • RTE_KERNELDIR: This variable contains the absolute path to the kernel sources that will be used to compile the kernel modules. The kernel headers must be the same as the ones that will be used on the target machine (the machine that will run the application). By default, the variable is set to /lib/modules/$(shell uname -r)/build, which is correct when the target machine is also the build machine.
  • RTE_DEVEL_BUILD: Stricter options (stop on warning). It defaults to y in a git tree.
Variables that Can be Set/Overridden in a Makefile Only
  • VPATH: The path list that the build system will search for sources. By default, RTE_SRCDIR will be included in VPATH.
  • CFLAGS: Flags to use for C compilation. The user should use += to append data in this variable.
  • LDFLAGS: Flags to use for linking. The user should use += to append data in this variable.
  • ASFLAGS: Flags to use for assembly. The user should use += to append data in this variable.
  • CPPFLAGS: Flags to use to give flags to C preprocessor (only useful when assembling .S files). The user should use += to append data in this variable.
  • LDLIBS: In an application, the list of libraries to link with (for example, -L /path/to/libfoo -lfoo ). The user should use += to append data in this variable.
  • SRC-y: A list of source files (.c, .S, or .o if the source is a binary) in case of application, library or object Makefiles. The sources must be available from VPATH.
  • INSTALL-y-$(INSTPATH): A list of files to be installed in $(INSTPATH). The files must be available from VPATH and will be copied in $(RTE_OUTPUT)/$(INSTPATH). Can be used in almost any RTE Makefile.
  • SYMLINK-y-$(INSTPATH): A list of files to be installed in $(INSTPATH). The files must be available from VPATH and will be linked (symbolically) in $(RTE_OUTPUT)/$(INSTPATH). This variable can be used in almost any DPDK Makefile.
  • PREBUILD: A list of prerequisite actions to be taken before building. The user should use += to append data in this variable.
  • POSTBUILD: A list of actions to be taken after the main build. The user should use += to append data in this variable.
  • PREINSTALL: A list of prerequisite actions to be taken before installing. The user should use += to append data in this variable.
  • POSTINSTALL: A list of actions to be taken after installing. The user should use += to append data in this variable.
  • PRECLEAN: A list of prerequisite actions to be taken before cleaning. The user should use += to append data in this variable.
  • POSTCLEAN: A list of actions to be taken after cleaning. The user should use += to append data in this variable.
  • DEPDIRS-$(DIR): Only used in the development kit framework to specify if the build of the current directory depends on build of another one. This is needed to support parallel builds correctly.
Variables that can be Set/Overridden by the User on the Command Line Only

Some variables can be used to configure the build system behavior. They are documented in Development Kit Root Makefile Help and External Application/Library Makefile Help

  • WERROR_CFLAGS: By default, this is set to a specific value that depends on the compiler. Users are encouraged to use this variable as follows:

    CFLAGS += $(WERROR_CFLAGS)

This avoids the use of different cases depending on the compiler (icc or gcc). Also, this variable can be overridden from the command line, which allows bypassing of the flags for testing purposes.

Variables that Can be Set/Overridden by the User in a Makefile or Command Line
  • CFLAGS_my_file.o: Specific flags to add for C compilation of my_file.c.
  • LDFLAGS_my_app: Specific flags to add when linking my_app.
  • EXTRA_CFLAGS: The content of this variable is appended after CFLAGS when compiling.
  • EXTRA_LDFLAGS: The content of this variable is appended after LDFLAGS when linking.
  • EXTRA_LDLIBS: The content of this variable is appended after LDLIBS when linking.
  • EXTRA_ASFLAGS: The content of this variable is appended after ASFLAGS when assembling.
  • EXTRA_CPPFLAGS: The content of this variable is appended after CPPFLAGS when using a C preprocessor on assembly files.

Development Kit Root Makefile Help

The DPDK provides a root level Makefile with targets for configuration, building, cleaning, testing, installation and others. These targets are explained in the following sections.

Configuration Targets

The configuration target requires the name of the target, which is specified using T=mytarget and it is mandatory. The list of available targets are in $(RTE_SDK)/config (remove the defconfig _ prefix).

Configuration targets also support the specification of the name of the output directory, using O=mybuilddir. This is an optional parameter, the default output directory is build.

  • Config

    This will create a build directory, and generates a configuration from a template. A Makefile is also created in the new build directory.

    Example:

    make config O=mybuild T=x86_64-native-linuxapp-gcc
    

Build Targets

Build targets support the optional specification of the name of the output directory, using O=mybuilddir. The default output directory is build.

  • all, build or just make

    Build the DPDK in the output directory previously created by a make config.

    Example:

    make O=mybuild
    
  • clean

    Clean all objects created using make build.

    Example:

    make clean O=mybuild
    
  • %_sub

    Build a subdirectory only, without managing dependencies on other directories.

    Example:

    make lib/librte_eal_sub O=mybuild
    
  • %_clean

    Clean a subdirectory only.

    Example:

    make lib/librte_eal_clean O=mybuild
    

Install Targets

Test Targets

  • test

    Launch automatic tests for a build directory specified using O=mybuilddir. It is optional, the default output directory is build.

    Example:

    make test O=mybuild
    

Documentation Targets

  • doc

    Generate the documentation (API and guides).

  • doc-api-html

    Generate the Doxygen API documentation in html.

  • doc-guides-html

    Generate the guides documentation in html.

  • doc-guides-pdf

    Generate the guides documentation in pdf.

Misc Targets

  • help

    Show a quick help.

Other Useful Command-line Variables

The following variables can be specified on the command line:

  • V=

    Enable verbose build (show full compilation command line, and some intermediate commands).

  • D=

    Enable dependency debugging. This provides some useful information about why a target is built or not.

  • EXTRA_CFLAGS=, EXTRA_LDFLAGS=, EXTRA_LDLIBS=, EXTRA_ASFLAGS=, EXTRA_CPPFLAGS=

    Append specific compilation, link or asm flags.

  • CROSS=

    Specify a cross toolchain header that will prefix all gcc/binutils applications. This only works when using gcc.

Make in a Build Directory

All targets described above are called from the SDK root $(RTE_SDK). It is possible to run the same Makefile targets inside the build directory. For instance, the following command:

cd $(RTE_SDK)
make config O=mybuild T=x86_64-native-linuxapp-gcc
make O=mybuild

is equivalent to:

cd $(RTE_SDK)
make config O=mybuild T=x86_64-native-linuxapp-gcc
cd mybuild

# no need to specify O= now
make

Compiling for Debug

To compile the DPDK and sample applications with debugging information included and the optimization level set to 0, the EXTRA_CFLAGS environment variable should be set before compiling as follows:

export EXTRA_CFLAGS='-O0 -g'

Extending the DPDK

This chapter describes how a developer can extend the DPDK to provide a new library, a new target, or support a new target.

Example: Adding a New Library libfoo

To add a new library to the DPDK, proceed as follows:

  1. Add a new configuration option:

    for f in config/\*; do \
        echo CONFIG_RTE_LIBFOO=y >> $f; done
    
  2. Create a new directory with sources:

    mkdir ${RTE_SDK}/lib/libfoo
    touch ${RTE_SDK}/lib/libfoo/foo.c
    touch ${RTE_SDK}/lib/libfoo/foo.h
    
  3. Add a foo() function in libfoo.

    Definition is in foo.c:

    void foo(void)
    {
    }
    

    Declaration is in foo.h:

    extern void foo(void);
    
  4. Update lib/Makefile:

    vi ${RTE_SDK}/lib/Makefile
    # add:
    # DIRS-$(CONFIG_RTE_LIBFOO) += libfoo
    
  5. Create a new Makefile for this library, for example, derived from mempool Makefile:

    cp ${RTE_SDK}/lib/librte_mempool/Makefile ${RTE_SDK}/lib/libfoo/
    
    vi ${RTE_SDK}/lib/libfoo/Makefile
    # replace:
    # librte_mempool -> libfoo
    # rte_mempool -> foo
    
  6. Update mk/DPDK.app.mk, and add -lfoo in LDLIBS variable when the option is enabled. This will automatically add this flag when linking a DPDK application.

  7. Build the DPDK with the new library (we only show a specific target here):

    cd ${RTE_SDK}
    make config T=x86_64-native-linuxapp-gcc
    make
    
  8. Check that the library is installed:

    ls build/lib
    ls build/include
    
Example: Using libfoo in the Test Application

The test application is used to validate all functionality of the DPDK. Once you have added a library, a new test case should be added in the test application.

  • A new test_foo.c file should be added, that includes foo.h and calls the foo() function from test_foo(). When the test passes, the test_foo() function should return 0.

  • Makefile, test.h and commands.c must be updated also, to handle the new test case.

  • Test report generation: autotest.py is a script that is used to generate the test report that is available in the ${RTE_SDK}/doc/rst/test_report/autotests directory. This script must be updated also. If libfoo is in a new test family, the links in ${RTE_SDK}/doc/rst/test_report/test_report.rst must be updated.

  • Build the DPDK with the updated test application (we only show a specific target here):

    cd ${RTE_SDK}
    make config T=x86_64-native-linuxapp-gcc
    make
    

Building Your Own Application

Compiling a Sample Application in the Development Kit Directory

When compiling a sample application (for example, hello world), the following variables must be exported: RTE_SDK and RTE_TARGET.

~/DPDK$ cd examples/helloworld/
~/DPDK/examples/helloworld$ export RTE_SDK=/home/user/DPDK
~/DPDK/examples/helloworld$ export RTE_TARGET=x86_64-native-linuxapp-gcc
~/DPDK/examples/helloworld$ make
    CC main.o
    LD helloworld
    INSTALL-APP helloworld
    INSTALL-MAP helloworld.map

The binary is generated in the build directory by default:

~/DPDK/examples/helloworld$ ls build/app
helloworld helloworld.map

Build Your Own Application Outside the Development Kit

The sample application (Hello World) can be duplicated in a new directory as a starting point for your development:

~$ cp -r DPDK/examples/helloworld my_rte_app
~$ cd my_rte_app/
~/my_rte_app$ export RTE_SDK=/home/user/DPDK
~/my_rte_app$ export RTE_TARGET=x86_64-native-linuxapp-gcc
~/my_rte_app$ make
    CC main.o
    LD helloworld
    INSTALL-APP helloworld
    INSTALL-MAP helloworld.map

Customizing Makefiles

Application Makefile

The default makefile provided with the Hello World sample application is a good starting point. It includes:

  • $(RTE_SDK)/mk/rte.vars.mk at the beginning
  • $(RTE_SDK)/mk/rte.extapp.mk at the end

The user must define several variables:

  • APP: Contains the name of the application.
  • SRCS-y: List of source files (*.c, *.S).
Library Makefile

It is also possible to build a library in the same way:

  • Include $(RTE_SDK)/mk/rte.vars.mk at the beginning.
  • Include $(RTE_SDK)/mk/rte.extlib.mk at the end.

The only difference is that APP should be replaced by LIB, which contains the name of the library. For example, libfoo.a.

Customize Makefile Actions

Some variables can be defined to customize Makefile actions. The most common are listed below. Refer to Makefile Description section in Development Kit Build System

chapter for details.

  • VPATH: The path list where the build system will search for sources. By default, RTE_SRCDIR will be included in VPATH.
  • CFLAGS_my_file.o: The specific flags to add for C compilation of my_file.c.
  • CFLAGS: The flags to use for C compilation.
  • LDFLAGS: The flags to use for linking.
  • CPPFLAGS: The flags to use to provide flags to the C preprocessor (only useful when assembling .S files)
  • LDLIBS: A list of libraries to link with (for example, -L /path/to/libfoo - lfoo)

External Application/Library Makefile help

External applications or libraries should include specific Makefiles from RTE_SDK, located in mk directory. These Makefiles are:

  • ${RTE_SDK}/mk/rte.extapp.mk: Build an application
  • ${RTE_SDK}/mk/rte.extlib.mk: Build a static library
  • ${RTE_SDK}/mk/rte.extobj.mk: Build objects (.o)

Prerequisites

The following variables must be defined:

  • ${RTE_SDK}: Points to the root directory of the DPDK.
  • ${RTE_TARGET}: Reference the target to be used for compilation (for example, x86_64-native-linuxapp-gcc).

Build Targets

Build targets support the specification of the name of the output directory, using O=mybuilddir. This is optional; the default output directory is build.

  • all, “nothing” (meaning just make)

    Build the application or the library in the specified output directory.

    Example:

    make O=mybuild
    
  • clean

    Clean all objects created using make build.

    Example:

    make clean O=mybuild
    

Help Targets

  • help

    Show this help.

Other Useful Command-line Variables

The following variables can be specified at the command line:

  • S=

    Specify the directory in which the sources are located. By default, it is the current directory.

  • M=

    Specify the Makefile to call once the output directory is created. By default, it uses $(S)/Makefile.

  • V=

    Enable verbose build (show full compilation command line and some intermediate commands).

  • D=

    Enable dependency debugging. This provides some useful information about why a target must be rebuilt or not.

  • EXTRA_CFLAGS=, EXTRA_LDFLAGS=, EXTRA_ASFLAGS=, EXTRA_CPPFLAGS=

    Append specific compilation, link or asm flags.

  • CROSS=

    Specify a cross-toolchain header that will prefix all gcc/binutils applications. This only works when using gcc.

Make from Another Directory

It is possible to run the Makefile from another directory, by specifying the output and the source dir. For example:

export RTE_SDK=/path/to/DPDK
export RTE_TARGET=x86_64-native-linuxapp-icc
make -f /path/to/my_app/Makefile S=/path/to/my_app O=/path/to/build_dir

Part 3: Performance Optimization

Performance Optimization Guidelines

Introduction

The following sections describe optimizations used in DPDK and optimizations that should be considered for new applications.

They also highlight the performance-impacting coding techniques that should, and should not be, used when developing an application using the DPDK.

And finally, they give an introduction to application profiling using a Performance Analyzer from Intel to optimize the software.

Writing Efficient Code

This chapter provides some tips for developing efficient code using the DPDK. For additional and more general information, please refer to the Intel® 64 and IA-32 Architectures Optimization Reference Manual which is a valuable reference to writing efficient code.

Memory

This section describes some key memory considerations when developing applications in the DPDK environment.

Memory Copy: Do not Use libc in the Data Plane

Many libc functions are available in the DPDK, via the Linux* application environment. This can ease the porting of applications and the development of the configuration plane. However, many of these functions are not designed for performance. Functions such as memcpy() or strcpy() should not be used in the data plane. To copy small structures, the preference is for a simpler technique that can be optimized by the compiler. Refer to the VTune™ Performance Analyzer Essentials publication from Intel Press for recommendations.

For specific functions that are called often, it is also a good idea to provide a self-made optimized function, which should be declared as static inline.

The DPDK API provides an optimized rte_memcpy() function.

Memory Allocation

Other functions of libc, such as malloc(), provide a flexible way to allocate and free memory. In some cases, using dynamic allocation is necessary, but it is really not advised to use malloc-like functions in the data plane because managing a fragmented heap can be costly and the allocator may not be optimized for parallel allocation.

If you really need dynamic allocation in the data plane, it is better to use a memory pool of fixed-size objects. This API is provided by librte_mempool. This data structure provides several services that increase performance, such as memory alignment of objects, lockless access to objects, NUMA awareness, bulk get/put and per-lcore cache. The rte_malloc () function uses a similar concept to mempools.

Concurrent Access to the Same Memory Area

Read-Write (RW) access operations by several lcores to the same memory area can generate a lot of data cache misses, which are very costly. It is often possible to use per-lcore variables, for example, in the case of statistics. There are at least two solutions for this:

  • Use RTE_PER_LCORE variables. Note that in this case, data on lcore X is not available to lcore Y.
  • Use a table of structures (one per lcore). In this case, each structure must be cache-aligned.

Read-mostly variables can be shared among lcores without performance losses if there are no RW variables in the same cache line.

NUMA

On a NUMA system, it is preferable to access local memory since remote memory access is slower. In the DPDK, the memzone, ring, rte_malloc and mempool APIs provide a way to create a pool on a specific socket.

Sometimes, it can be a good idea to duplicate data to optimize speed. For read-mostly variables that are often accessed, it should not be a problem to keep them in one socket only, since data will be present in cache.

Distribution Across Memory Channels

Modern memory controllers have several memory channels that can load or store data in parallel. Depending on the memory controller and its configuration, the number of channels and the way the memory is distributed across the channels varies. Each channel has a bandwidth limit, meaning that if all memory access operations are done on the first channel only, there is a potential bottleneck.

By default, the Mempool Library spreads the addresses of objects among memory channels.

Locking memory pages

The underlying operating system is allowed to load/unload memory pages at its own discretion. These page loads could impact the performance, as the process is on hold when the kernel fetches them.

To avoid these you could pre-load, and lock them into memory with the mlockall() call.

if (mlockall(MCL_CURRENT | MCL_FUTURE)) {
    RTE_LOG(NOTICE, USER1, "mlockall() failed with error \"%s\"\n",
            strerror(errno));
}

Communication Between lcores

To provide a message-based communication between lcores, it is advised to use the DPDK ring API, which provides a lockless ring implementation.

The ring supports bulk and burst access, meaning that it is possible to read several elements from the ring with only one costly atomic operation (see Ring Library). Performance is greatly improved when using bulk access operations.

The code algorithm that dequeues messages may be something similar to the following:

 #define MAX_BULK 32

 while (1) {
     /* Process as many elements as can be dequeued. */
     count = rte_ring_dequeue_burst(ring, obj_table, MAX_BULK, NULL);
     if (unlikely(count == 0))
         continue;

     my_process_bulk(obj_table, count);
}

PMD Driver

The DPDK Poll Mode Driver (PMD) is also able to work in bulk/burst mode, allowing the factorization of some code for each call in the send or receive function.

Avoid partial writes. When PCI devices write to system memory through DMA, it costs less if the write operation is on a full cache line as opposed to part of it. In the PMD code, actions have been taken to avoid partial writes as much as possible.

Lower Packet Latency

Traditionally, there is a trade-off between throughput and latency. An application can be tuned to achieve a high throughput, but the end-to-end latency of an average packet will typically increase as a result. Similarly, the application can be tuned to have, on average, a low end-to-end latency, at the cost of lower throughput.

In order to achieve higher throughput, the DPDK attempts to aggregate the cost of processing each packet individually by processing packets in bursts.

Using the testpmd application as an example, the burst size can be set on the command line to a value of 16 (also the default value). This allows the application to request 16 packets at a time from the PMD. The testpmd application then immediately attempts to transmit all the packets that were received, in this case, all 16 packets.

The packets are not transmitted until the tail pointer is updated on the corresponding TX queue of the network port. This behavior is desirable when tuning for high throughput because the cost of tail pointer updates to both the RX and TX queues can be spread across 16 packets, effectively hiding the relatively slow MMIO cost of writing to the PCIe* device. However, this is not very desirable when tuning for low latency because the first packet that was received must also wait for another 15 packets to be received. It cannot be transmitted until the other 15 packets have also been processed because the NIC will not know to transmit the packets until the TX tail pointer has been updated, which is not done until all 16 packets have been processed for transmission.

To consistently achieve low latency, even under heavy system load, the application developer should avoid processing packets in bunches. The testpmd application can be configured from the command line to use a burst value of 1. This will allow a single packet to be processed at a time, providing lower latency, but with the added cost of lower throughput.

Locks and Atomic Operations

Atomic operations imply a lock prefix before the instruction, causing the processor’s LOCK# signal to be asserted during execution of the following instruction. This has a big impact on performance in a multicore environment.

Performance can be improved by avoiding lock mechanisms in the data plane. It can often be replaced by other solutions like per-lcore variables. Also, some locking techniques are more efficient than others. For instance, the Read-Copy-Update (RCU) algorithm can frequently replace simple rwlocks.

Coding Considerations

Inline Functions

Small functions can be declared as static inline in the header file. This avoids the cost of a call instruction (and the associated context saving). However, this technique is not always efficient; it depends on many factors including the compiler.

Branch Prediction

The Intel® C/C++ Compiler (icc)/gcc built-in helper functions likely() and unlikely() allow the developer to indicate if a code branch is likely to be taken or not. For instance:

if (likely(x > 1))
    do_stuff();

Setting the Target CPU Type

The DPDK supports CPU microarchitecture-specific optimizations by means of CONFIG_RTE_MACHINE option in the DPDK configuration file. The degree of optimization depends on the compiler’s ability to optimize for a specific microarchitecture, therefore it is preferable to use the latest compiler versions whenever possible.

If the compiler version does not support the specific feature set (for example, the Intel® AVX instruction set), the build process gracefully degrades to whatever latest feature set is supported by the compiler.

Since the build and runtime targets may not be the same, the resulting binary also contains a platform check that runs before the main() function and checks if the current machine is suitable for running the binary.

Along with compiler optimizations, a set of preprocessor defines are automatically added to the build process (regardless of the compiler version). These defines correspond to the instruction sets that the target CPU should be able to support. For example, a binary compiled for any SSE4.2-capable processor will have RTE_MACHINE_CPUFLAG_SSE4_2 defined, thus enabling compile-time code path selection for different platforms.

Profile Your Application

The following sections describe methods of profiling DPDK applications on different architectures.

Profiling on x86

Intel processors provide performance counters to monitor events. Some tools provided by Intel, such as Intel® VTune™ Amplifier, can be used to profile and benchmark an application. See the VTune Performance Analyzer Essentials publication from Intel Press for more information.

For a DPDK application, this can be done in a Linux* application environment only.

The main situations that should be monitored through event counters are:

  • Cache misses
  • Branch mis-predicts
  • DTLB misses
  • Long latency instructions and exceptions

Refer to the Intel Performance Analysis Guide for details about application profiling.

Empty cycles tracing

Iterations that yielded no RX packets (empty cycles, wasted iterations) can be analyzed using VTune Amplifier. This profiling employs the Instrumentation and Tracing Technology (ITT) API feature of VTune Amplifier and requires only reconfiguring the DPDK library, no changes in a DPDK application are needed.

To trace wasted iterations on RX queues, first reconfigure DPDK with CONFIG_RTE_ETHDEV_RXTX_CALLBACKS and CONFIG_RTE_ETHDEV_PROFILE_ITT_WASTED_RX_ITERATIONS enabled.

Then rebuild DPDK, specifying paths to the ITT header and library, which can be found in any VTune Amplifier distribution in the include and lib directories respectively:

make EXTRA_CFLAGS=-I<path to ittnotify.h> \
     EXTRA_LDLIBS="-L<path to libittnotify.a> -littnotify"

Finally, to see wasted iterations in your performance analysis results, select the “Analyze user tasks, events, and counters” checkbox in the “Analysis Type” tab when configuring analysis via VTune Amplifier GUI. Alternatively, when running VTune Amplifier via command line, specify -knob enable-user-tasks=true option.

Collected regions of wasted iterations will be marked on VTune Amplifier’s timeline as ITT tasks. These ITT tasks have predefined names, containing Ethernet device and RX queue identifiers.

Profiling on ARM64

Using Linux perf

The ARM64 architecture provide performance counters to monitor events. The Linux perf tool can be used to profile and benchmark an application. In addition to the standard events, perf can be used to profile arm64 specific PMU (Performance Monitor Unit) events through raw events (-e -rXX).

For more derails refer to the ARM64 specific PMU events enumeration.

High-resolution cycle counter

The default cntvct_el0 based rte_rdtsc() provides a portable means to get a wall clock counter in user space. Typically it runs at <= 100MHz.

The alternative method to enable rte_rdtsc() for a high resolution wall clock counter is through the armv8 PMU subsystem. The PMU cycle counter runs at CPU frequency. However, access to the PMU cycle counter from user space is not enabled by default in the arm64 linux kernel. It is possible to enable cycle counter for user space access by configuring the PMU from the privileged mode (kernel space).

By default the rte_rdtsc() implementation uses a portable cntvct_el0 scheme. Application can choose the PMU based implementation with CONFIG_RTE_ARM_EAL_RDTSC_USE_PMU.

The example below shows the steps to configure the PMU based cycle counter on an armv8 machine.

git clone https://github.com/jerinjacobk/armv8_pmu_cycle_counter_el0
cd armv8_pmu_cycle_counter_el0
make
sudo insmod pmu_el0_cycle_counter.ko
cd $DPDK_DIR
make config T=arm64-armv8a-linuxapp-gcc
echo "CONFIG_RTE_ARM_EAL_RDTSC_USE_PMU=y" >> build/.config
make

Warning

The PMU based scheme is useful for high accuracy performance profiling with rte_rdtsc(). However, this method can not be used in conjunction with Linux userspace profiling tools like perf as this scheme alters the PMU registers state.

Glossary

ACL
Access Control List
API
Application Programming Interface
ASLR
Linux* kernel Address-Space Layout Randomization
BSD
Berkeley Software Distribution
Clr
Clear
CIDR
Classless Inter-Domain Routing
Control Plane
The control plane is concerned with the routing of packets and with providing a start or end point.
Core
A core may include several lcores or threads if the processor supports hyperthreading.
Core Components
A set of libraries provided by the DPDK, including eal, ring, mempool, mbuf, timers, and so on.
CPU
Central Processing Unit
CRC
Cyclic Redundancy Check
ctrlmbuf
An mbuf carrying control data.
Data Plane
In contrast to the control plane, the data plane in a network architecture are the layers involved when forwarding packets. These layers must be highly optimized to achieve good performance.
DIMM
Dual In-line Memory Module
Doxygen
A documentation generator used in the DPDK to generate the API reference.
DPDK
Data Plane Development Kit
DRAM
Dynamic Random Access Memory
EAL
The Environment Abstraction Layer (EAL) provides a generic interface that hides the environment specifics from the applications and libraries. The services expected from the EAL are: development kit loading and launching, core affinity/ assignment procedures, system memory allocation/description, PCI bus access, inter-partition communication.
FIFO
First In First Out
FPGA
Field Programmable Gate Array
GbE
Gigabit Ethernet
HW
Hardware
HPET
High Precision Event Timer; a hardware timer that provides a precise time reference on x86 platforms.
ID
Identifier
IOCTL
Input/Output Control
I/O
Input/Output
IP
Internet Protocol
IPv4
Internet Protocol version 4
IPv6
Internet Protocol version 6
lcore
A logical execution unit of the processor, sometimes called a hardware thread.
KNI
Kernel Network Interface
L1
Layer 1
L2
Layer 2
L3
Layer 3
L4
Layer 4
LAN
Local Area Network
LPM
Longest Prefix Match
master lcore
The execution unit that executes the main() function and that launches other lcores.
mbuf
An mbuf is a data structure used internally to carry messages (mainly network packets). The name is derived from BSD stacks. To understand the concepts of packet buffers or mbuf, refer to TCP/IP Illustrated, Volume 2: The Implementation.
MESI
Modified Exclusive Shared Invalid (CPU cache coherency protocol)
MTU
Maximum Transfer Unit
NIC
Network Interface Card
OOO
Out Of Order (execution of instructions within the CPU pipeline)
NUMA
Non-uniform Memory Access
PCI
Peripheral Connect Interface
PHY
An abbreviation for the physical layer of the OSI model.
pktmbuf
An mbuf carrying a network packet.
PMD
Poll Mode Driver
QoS
Quality of Service
RCU
Read-Copy-Update algorithm, an alternative to simple rwlocks.
Rd
Read
RED
Random Early Detection
RSS
Receive Side Scaling
RTE
Run Time Environment. Provides a fast and simple framework for fast packet processing, in a lightweight environment as a Linux* application and using Poll Mode Drivers (PMDs) to increase speed.
Rx
Reception
Slave lcore
Any lcore that is not the master lcore.
Socket
A physical CPU, that includes several cores.
SLA
Service Level Agreement
srTCM
Single Rate Three Color Marking
SRTD
Scheduler Round Trip Delay
SW
Software
Target
In the DPDK, the target is a combination of architecture, machine, executive environment and toolchain. For example: i686-native-linuxapp-gcc.
TCP
Transmission Control Protocol
TC
Traffic Class
TLB
Translation Lookaside Buffer
TLS
Thread Local Storage
trTCM
Two Rate Three Color Marking
TSC
Time Stamp Counter
Tx
Transmission
TUN/TAP
TUN and TAP are virtual network kernel devices.
VLAN
Virtual Local Area Network
Wr
Write
WRED
Weighted Random Early Detection
WRR
Weighted Round Robin

Figures

Fig. 2.1 Core Components Architecture

Fig. 3.1 EAL Initialization in a Linux Application Environment

Fig. 3.2 Example of a malloc heap and malloc elements within the malloc library

Fig. 5.2 Ring Structure

Fig. 5.3 Enqueue first step

Fig. 5.4 Enqueue second step

Fig. 5.5 Enqueue last step

Fig. 5.6 Dequeue last step

Fig. 5.7 Dequeue second step

Fig. 5.8 Dequeue last step

Fig. 5.9 Multiple producer enqueue first step

Fig. 5.10 Multiple producer enqueue second step

Fig. 5.11 Multiple producer enqueue third step

Fig. 5.12 Multiple producer enqueue fourth step

Fig. 5.13 Multiple producer enqueue last step

Fig. 5.14 Modulo 32-bit indexes - Example 1

Fig. 5.15 Modulo 32-bit indexes - Example 2

Fig. 6.1 Two Channels and Quad-ranked DIMM Example

Fig. 6.2 Three Channels and Two Dual-ranked DIMM Example

Fig. 6.3 A mempool in Memory with its Associated Ring

Fig. 7.1 An mbuf with One Segment

Fig. 7.2 An mbuf with Three Segments

Fig. 28.5 Memory Sharing in the DPDK Multi-process Sample Application

Fig. 29.1 Components of a DPDK KNI Application

Fig. 29.2 Packet Flow via mbufs in the DPDK KNI

Fig. 33.1 Complex Packet Processing Pipeline with QoS Support

Fig. 33.2 Hierarchical Scheduler Block Internal Diagram

Fig. 33.3 Scheduling Hierarchy per Port

Fig. 33.4 Internal Data Structures per Port

Fig. 33.5 Prefetch Pipeline for the Hierarchical Scheduler Enqueue Operation

Fig. 33.6 Pipe Prefetch State Machine for the Hierarchical Scheduler Dequeue Operation

Fig. 33.7 High-level Block Diagram of the DPDK Dropper

Fig. 33.8 Flow Through the Dropper

Fig. 33.9 Example Data Flow Through Dropper

Fig. 33.10 Packet Drop Probability for a Given RED Configuration

Fig. 33.11 Initial Drop Probability (pb), Actual Drop probability (pa) Computed Using a Factor 1 (Blue Curve) and a Factor 2 (Red Curve)

Fig. 36.1 Example of Packet Processing Pipeline where Input Ports 0 and 1 are Connected with Output Ports 0, 1 and 2 through Tables 0 and 1

Fig. 36.2 Sequence of Steps for Hash Table Operations in a Packet Processing Context

Fig. 36.3 Data Structures for Configurable Key Size Hash Tables

Fig. 36.4 Bucket Search Pipeline for Key Lookup Operation (Configurable Key Size Hash Tables)

Fig. 36.5 Data Structures for 8-byte Key Hash Tables

Fig. 36.6 Data Structures for 16-byte Key Hash Tables

Fig. 36.7 Bucket Search Pipeline for Key Lookup Operation (Single Key Size Hash Tables)

Fig. 17.3 Load Balancing Using Front End Node

Fig. 17.4 Consistent Hashing

Fig. 17.5 Table Based Flow Distribution

Fig. 17.6 Searching for Perfect Hash Function

Fig. 17.7 Divide and Conquer for Millions of Keys

Fig. 17.8 EFD as a Flow-Level Load Balancer

Fig. 17.9 Group Assignment

Fig. 17.10 Perfect Hash Search - Assigned Keys & Target Value

Fig. 17.11 Perfect Hash Search - Satisfy Target Values

Fig. 17.12 Finding Hash Index for Conflict Free lookup_table

Fig. 17.13 EFD Lookup Operation

Fig. 18.4 Example Usages of Membership Library

Fig. 18.5 Bloom Filter False Positive Probability

Fig. 18.6 Detecting Routing Loops Using BF

Fig. 18.7 Vector Bloom Filter (vBF) Overview

Fig. 18.8 vBF for Flow Scheduling to Worker Thread

Fig. 18.9 Using HTSS for Attack Signature Matching

Fig. 18.10 Using HTSS with False Negatives for Wild Card Classification

Tables

Table 33.1 Packet Processing Pipeline Implementing QoS

Table 33.2 Infrastructure Blocks Used by the Packet Processing Pipeline

Table 33.3 Port Scheduling Hierarchy

Table 33.4 Scheduler Internal Data Structures per Port

Table 33.5 Ethernet Frame Overhead Fields

Table 33.6 Token Bucket Generic Operations

Table 33.7 Token Bucket Generic Parameters

Table 33.8 Token Bucket Persistent Data Structure

Table 33.9 Token Bucket Operations

Table 33.10 Subport/Pipe Traffic Class Upper Limit Enforcement Persistent Data Structure

Table 33.11 Subport/Pipe Traffic Class Upper Limit Enforcement Operations

Table 33.12 Weighted Round Robin (WRR)

Table 33.13 Subport Traffic Class Oversubscription

Table 33.14 Watermark Propagation from Subport Level to Member Pipes at the Beginning of Each Traffic Class Upper Limit Enforcement Period

Table 33.15 Watermark Calculation

Table 33.16 RED Configuration Parameters

Table 33.17 Relative Performance of Alternative Approaches

Table 33.18 RED Configuration Corresponding to RED Configuration File

Table 36.1 Port Types

Table 36.2 20 Port Abstract Interface

Table 36.3 Table Types

Table 36.5 Configuration Parameters Common for All Hash Table Types

Table 36.6 Configuration Parameters Specific to Extendable Bucket Hash Table

Table 36.7 Configuration Parameters Specific to Pre-computed Key Signature Hash Table

Table 36.8 Main Large Data Structures (Arrays) used for Configurable Key Size Hash Tables

Table 36.9 Field Description for Bucket Array Entry (Configurable Key Size Hash Tables)

Table 36.10 Description of the Bucket Search Pipeline Stages (Configurable Key Size Hash Tables)

Table 36.11 Lookup Tables for Match, Match_Many and Match_Pos

Table 36.12 Collapsed Lookup Tables for Match, Match_Many and Match_Pos

Table 36.13 Main Large Data Structures (Arrays) used for 8-byte and 16-byte Key Size Hash Tables

Table 36.14 Field Description for Bucket Array Entry (8-byte and 16-byte Key Hash Tables)

Table 36.15 Description of the Bucket Search Pipeline Stages (8-byte and 16-byte Key Hash Tables)

Table 36.16 Next Hop Actions (Reserved)

Table 36.17 User Action Examples

Table 16.1 Entry distribution measured with an example table with 1024 random entries using jhash algorithm

Table 16.2 Entry distribution measured with an example table with 1 million random entries using jhash algorithm

HowTo Guides

Live Migration of VM with SR-IOV VF

Overview

It is not possible to migrate a Virtual Machine which has an SR-IOV Virtual Function (VF).

To get around this problem the bonding PMD is used.

The following sections show an example of how to do this.

Test Setup

A bonded device is created in the VM. The virtio and VF PMD’s are added as slaves to the bonded device. The VF is set as the primary slave of the bonded device.

A bridge must be set up on the Host connecting the tap device, which is the backend of the Virtio device and the Physical Function (PF) device.

To test the Live Migration two servers with identical operating systems installed are used. KVM and Qemu 2.3 is also required on the servers.

In this example, the servers have Niantic and or Fortville NIC’s installed. The NIC’s on both servers are connected to a switch which is also connected to the traffic generator.

The switch is configured to broadcast traffic on all the NIC ports. A Sample switch configuration can be found in this section.

The host is running the Kernel PF driver (ixgbe or i40e).

The ip address of host_server_1 is 10.237.212.46

The ip address of host_server_2 is 10.237.212.131

_images/lm_bond_virtio_sriov.svg

Live Migration steps

The sample scripts mentioned in the steps below can be found in the Sample host scripts and Sample VM scripts sections.

On host_server_1: Terminal 1
cd /root/dpdk/host_scripts
./setup_vf_on_212_46.sh

For Fortville NIC

./vm_virtio_vf_i40e_212_46.sh

For Niantic NIC

./vm_virtio_vf_one_212_46.sh
On host_server_1: Terminal 2
cd /root/dpdk/host_scripts
./setup_bridge_on_212_46.sh
./connect_to_qemu_mon_on_host.sh
(qemu)
On host_server_1: Terminal 1

In VM on host_server_1:

cd /root/dpdk/vm_scripts
./setup_dpdk_in_vm.sh
./run_testpmd_bonding_in_vm.sh

testpmd> show port info all

The mac_addr command only works with kernel PF for Niantic

testpmd> mac_addr add port 1 vf 0 AA:BB:CC:DD:EE:FF

The syntax of the testpmd command is:

Create bonded device (mode) (socket).

Mode 1 is active backup.

Virtio is port 0 (P0).

VF is port 1 (P1).

Bonding is port 2 (P2).

testpmd> create bonded device 1 0
Created new bonded device net_bond_testpmd_0 on (port 2).
testpmd> add bonding slave 0 2
testpmd> add bonding slave 1 2
testpmd> show bonding config 2

The syntax of the testpmd command is:

set bonding primary (slave id) (port id)

Set primary to P1 before starting bonding port.

testpmd> set bonding primary 1 2
testpmd> show bonding config 2
testpmd> port start 2
Port 2: 02:09:C0:68:99:A5
Checking link statuses...
Port 0 Link Up - speed 10000 Mbps - full-duplex
Port 1 Link Up - speed 10000 Mbps - full-duplex
Port 2 Link Up - speed 10000 Mbps - full-duplex

testpmd> show bonding config 2

Primary is now P1. There are 2 active slaves.

Use P2 only for forwarding.

testpmd> set portlist 2
testpmd> show config fwd
testpmd> set fwd mac
testpmd> start
testpmd> show bonding config 2

Primary is now P1. There are 2 active slaves.

testpmd> show port stats all

VF traffic is seen at P1 and P2.

testpmd> clear port stats all
testpmd> set bonding primary 0 2
testpmd> remove bonding slave 1 2
testpmd> show bonding config 2

Primary is now P0. There is 1 active slave.

testpmd> clear port stats all
testpmd> show port stats all

No VF traffic is seen at P0 and P2, VF MAC address still present.

testpmd> port stop 1
testpmd> port close 1

Port close should remove VF MAC address, it does not remove perm_addr.

The mac_addr command only works with the kernel PF for Niantic.

testpmd> mac_addr remove 1 AA:BB:CC:DD:EE:FF
testpmd> port detach 1
Port '0000:00:04.0' is detached. Now total ports is 2
testpmd> show port stats all

No VF traffic is seen at P0 and P2.

On host_server_1: Terminal 2
(qemu) device_del vf1
On host_server_1: Terminal 1

In VM on host_server_1:

testpmd> show bonding config 2

Primary is now P0. There is 1 active slave.

testpmd> show port info all
testpmd> show port stats all
On host_server_2: Terminal 1
cd /root/dpdk/host_scripts
./setup_vf_on_212_131.sh
./vm_virtio_one_migrate.sh
On host_server_2: Terminal 2
./setup_bridge_on_212_131.sh
./connect_to_qemu_mon_on_host.sh
(qemu) info status
VM status: paused (inmigrate)
(qemu)
On host_server_1: Terminal 2

Check that the switch is up before migrating.

(qemu) migrate tcp:10.237.212.131:5555
(qemu) info status
VM status: paused (postmigrate)

For the Niantic NIC.

(qemu) info migrate
capabilities: xbzrle: off rdma-pin-all: off auto-converge: off zero-blocks: off
Migration status: completed
total time: 11834 milliseconds
downtime: 18 milliseconds
setup: 3 milliseconds
transferred ram: 389137 kbytes
throughput: 269.49 mbps
remaining ram: 0 kbytes
total ram: 1590088 kbytes
duplicate: 301620 pages
skipped: 0 pages
normal: 96433 pages
normal bytes: 385732 kbytes
dirty sync count: 2
(qemu) quit

For the Fortville NIC.

(qemu) info migrate
capabilities: xbzrle: off rdma-pin-all: off auto-converge: off zero-blocks: off
Migration status: completed
total time: 11619 milliseconds
downtime: 5 milliseconds
setup: 7 milliseconds
transferred ram: 379699 kbytes
throughput: 267.82 mbps
remaining ram: 0 kbytes
total ram: 1590088 kbytes
duplicate: 303985 pages
skipped: 0 pages
normal: 94073 pages
normal bytes: 376292 kbytes
dirty sync count: 2
(qemu) quit
On host_server_2: Terminal 1

In VM on host_server_2:

Hit Enter key. This brings the user to the testpmd prompt.
testpmd>
On host_server_2: Terminal 2
(qemu) info status
VM status: running

For the Niantic NIC.

(qemu) device_add pci-assign,host=06:10.0,id=vf1

For the Fortville NIC.

(qemu) device_add pci-assign,host=03:02.0,id=vf1
On host_server_2: Terminal 1

In VM on host_server_2:

testomd> show port info all
testpmd> show port stats all
testpmd> show bonding config 2
testpmd> port attach 0000:00:04.0
Port 1 is attached.
Now total ports is 3
Done

testpmd> port start 1

The mac_addr command only works with the Kernel PF for Niantic.

testpmd> mac_addr add port 1 vf 0 AA:BB:CC:DD:EE:FF
testpmd> show port stats all.
testpmd> show config fwd
testpmd> show bonding config 2
testpmd> add bonding slave 1 2
testpmd> set bonding primary 1 2
testpmd> show bonding config 2
testpmd> show port stats all

VF traffic is seen at P1 (VF) and P2 (Bonded device).

testpmd> remove bonding slave 0 2
testpmd> show bonding config 2
testpmd> port stop 0
testpmd> port close 0
testpmd> port detach 0
Port '0000:00:03.0' is detached. Now total ports is 2

testpmd> show port info all
testpmd> show config fwd
testpmd> show port stats all

VF traffic is seen at P1 (VF) and P2 (Bonded device).

Sample host scripts

setup_vf_on_212_46.sh

Set up Virtual Functions on host_server_1

#!/bin/sh
# This script is run on the host 10.237.212.46 to setup the VF

# set up Niantic VF
cat /sys/bus/pci/devices/0000\:09\:00.0/sriov_numvfs
echo 1 > /sys/bus/pci/devices/0000\:09\:00.0/sriov_numvfs
cat /sys/bus/pci/devices/0000\:09\:00.0/sriov_numvfs
rmmod ixgbevf

# set up Fortville VF
cat /sys/bus/pci/devices/0000\:02\:00.0/sriov_numvfs
echo 1 > /sys/bus/pci/devices/0000\:02\:00.0/sriov_numvfs
cat /sys/bus/pci/devices/0000\:02\:00.0/sriov_numvfs
rmmod i40evf
vm_virtio_vf_one_212_46.sh

Setup Virtual Machine on host_server_1

#!/bin/sh

# Path to KVM tool
KVM_PATH="/usr/bin/qemu-system-x86_64"

# Guest Disk image
DISK_IMG="/home/username/disk_image/virt1_sml.disk"

# Number of guest cpus
VCPUS_NR="4"

# Memory
MEM=1536

taskset -c 1-5 $KVM_PATH \
 -enable-kvm \
 -m $MEM \
 -smp $VCPUS_NR \
 -cpu host \
 -name VM1 \
 -no-reboot \
 -net none \
 -vnc none -nographic \
 -hda $DISK_IMG \
 -netdev type=tap,id=net1,script=no,downscript=no,ifname=tap1 \
 -device virtio-net-pci,netdev=net1,mac=CC:BB:BB:BB:BB:BB \
 -device pci-assign,host=09:10.0,id=vf1 \
 -monitor telnet::3333,server,nowait
setup_bridge_on_212_46.sh

Setup bridge on host_server_1

#!/bin/sh
# This script is run on the host 10.237.212.46 to setup the bridge
# for the Tap device and the PF device.
# This enables traffic to go from the PF to the Tap to the Virtio PMD in the VM.

# ens3f0 is the Niantic NIC
# ens6f0 is the Fortville NIC

ifconfig ens3f0 down
ifconfig tap1 down
ifconfig ens6f0 down
ifconfig virbr0 down

brctl show virbr0
brctl addif virbr0 ens3f0
brctl addif virbr0 ens6f0
brctl addif virbr0 tap1
brctl show virbr0

ifconfig ens3f0 up
ifconfig tap1 up
ifconfig ens6f0 up
ifconfig virbr0 up
connect_to_qemu_mon_on_host.sh
#!/bin/sh
# This script is run on both hosts when the VM is up,
# to connect to the Qemu Monitor.

telnet 0 3333
setup_vf_on_212_131.sh

Set up Virtual Functions on host_server_2

#!/bin/sh
# This script is run on the host 10.237.212.131 to setup the VF

# set up Niantic VF
cat /sys/bus/pci/devices/0000\:06\:00.0/sriov_numvfs
echo 1 > /sys/bus/pci/devices/0000\:06\:00.0/sriov_numvfs
cat /sys/bus/pci/devices/0000\:06\:00.0/sriov_numvfs
rmmod ixgbevf

# set up Fortville VF
cat /sys/bus/pci/devices/0000\:03\:00.0/sriov_numvfs
echo 1 > /sys/bus/pci/devices/0000\:03\:00.0/sriov_numvfs
cat /sys/bus/pci/devices/0000\:03\:00.0/sriov_numvfs
rmmod i40evf
vm_virtio_one_migrate.sh

Setup Virtual Machine on host_server_2

#!/bin/sh
# Start the VM on host_server_2 with the same parameters except without the VF
# parameters, as the VM on host_server_1, in migration-listen mode
# (-incoming tcp:0:5555)

# Path to KVM tool
KVM_PATH="/usr/bin/qemu-system-x86_64"

# Guest Disk image
DISK_IMG="/home/username/disk_image/virt1_sml.disk"

# Number of guest cpus
VCPUS_NR="4"

# Memory
MEM=1536

taskset -c 1-5 $KVM_PATH \
 -enable-kvm \
 -m $MEM \
 -smp $VCPUS_NR \
 -cpu host \
 -name VM1 \
 -no-reboot \
 -net none \
 -vnc none -nographic \
 -hda $DISK_IMG \
 -netdev type=tap,id=net1,script=no,downscript=no,ifname=tap1 \
 -device virtio-net-pci,netdev=net1,mac=CC:BB:BB:BB:BB:BB \
 -incoming tcp:0:5555 \
 -monitor telnet::3333,server,nowait
setup_bridge_on_212_131.sh

Setup bridge on host_server_2

#!/bin/sh
# This script is run on the host to setup the bridge
# for the Tap device and the PF device.
# This enables traffic to go from the PF to the Tap to the Virtio PMD in the VM.

# ens4f0 is the Niantic NIC
# ens5f0 is the Fortville NIC

ifconfig ens4f0 down
ifconfig tap1 down
ifconfig ens5f0 down
ifconfig virbr0 down

brctl show virbr0
brctl addif virbr0 ens4f0
brctl addif virbr0 ens5f0
brctl addif virbr0 tap1
brctl show virbr0

ifconfig ens4f0 up
ifconfig tap1 up
ifconfig ens5f0 up
ifconfig virbr0 up

Sample VM scripts

setup_dpdk_in_vm.sh

Set up DPDK in the Virtual Machine

#!/bin/sh
# this script matches the vm_virtio_vf_one script
# virtio port is 03
# vf port is 04

cat  /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
echo 1024 > /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
cat  /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages

ifconfig -a
/root/dpdk/usertools/dpdk-devbind.py --status

rmmod virtio-pci ixgbevf

modprobe uio
insmod /root/dpdk/x86_64-default-linuxapp-gcc/kmod/igb_uio.ko

/root/dpdk/usertools/dpdk-devbind.py -b igb_uio 0000:00:03.0
/root/dpdk/usertools/dpdk-devbind.py -b igb_uio 0000:00:04.0

/root/dpdk/usertools/dpdk-devbind.py --status
run_testpmd_bonding_in_vm.sh

Run testpmd in the Virtual Machine.

#!/bin/sh
# Run testpmd in the VM

# The test system has 8 cpus (0-7), use cpus 2-7 for VM
# Use taskset -pc <core number> <thread_id>

# use for bonding of virtio and vf tests in VM

/root/dpdk/x86_64-default-linuxapp-gcc/app/testpmd \
-l 0-3 -n 4 --socket-mem 350 --  --i --port-topology=chained

Sample switch configuration

The Intel switch is used to connect the traffic generator to the NIC’s on host_server_1 and host_server_2.

In order to run the switch configuration two console windows are required.

Log in as root in both windows.

TestPointShared, run_switch.sh and load /root/switch_config must be executed in the sequence below.

On Switch: Terminal 1

run TestPointShared

/usr/bin/TestPointShared
On Switch: Terminal 2

execute run_switch.sh

/root/run_switch.sh
On Switch: Terminal 1

load switch configuration

load /root/switch_config
Sample switch configuration script

The /root/switch_config script:

# TestPoint History
show port 1,5,9,13,17,21,25
set port 1,5,9,13,17,21,25 up
show port 1,5,9,13,17,21,25
del acl 1
create acl 1
create acl-port-set
create acl-port-set
add port port-set 1 0
add port port-set 5,9,13,17,21,25 1
create acl-rule 1 1
add acl-rule condition 1 1 port-set 1
add acl-rule action 1 1 redirect 1
apply acl
create vlan 1000
add vlan port 1000 1,5,9,13,17,21,25
set vlan tagging 1000 1,5,9,13,17,21,25 tag
set switch config flood_ucast fwd
show port stats all 1,5,9,13,17,21,25

Live Migration of VM with Virtio on host running vhost_user

Overview

Live Migration of a VM with DPDK Virtio PMD on a host which is running the Vhost sample application (vhost-switch) and using the DPDK PMD (ixgbe or i40e).

The Vhost sample application uses VMDQ so SRIOV must be disabled on the NIC’s.

The following sections show an example of how to do this migration.

Test Setup

To test the Live Migration two servers with identical operating systems installed are used. KVM and QEMU is also required on the servers.

QEMU 2.5 is required for Live Migration of a VM with vhost_user running on the hosts.

In this example, the servers have Niantic and or Fortville NIC’s installed. The NIC’s on both servers are connected to a switch which is also connected to the traffic generator.

The switch is configured to broadcast traffic on all the NIC ports.

The ip address of host_server_1 is 10.237.212.46

The ip address of host_server_2 is 10.237.212.131

_images/lm_vhost_user.svg

Live Migration steps

The sample scripts mentioned in the steps below can be found in the Sample host scripts and Sample VM scripts sections.

On host_server_1: Terminal 1

Setup DPDK on host_server_1

cd /root/dpdk/host_scripts
./setup_dpdk_on_host.sh
On host_server_1: Terminal 2

Bind the Niantic or Fortville NIC to igb_uio on host_server_1.

For Fortville NIC.

cd /root/dpdk/usertools
./dpdk-devbind.py -b igb_uio 0000:02:00.0

For Niantic NIC.

cd /root/dpdk/usertools
./dpdk-devbind.py -b igb_uio 0000:09:00.0
On host_server_1: Terminal 3

For Fortville and Niantic NIC’s reset SRIOV and run the vhost_user sample application (vhost-switch) on host_server_1.

cd /root/dpdk/host_scripts
./reset_vf_on_212_46.sh
./run_vhost_switch_on_host.sh
On host_server_1: Terminal 1

Start the VM on host_server_1

./vm_virtio_vhost_user.sh
On host_server_1: Terminal 4

Connect to the QEMU monitor on host_server_1.

cd /root/dpdk/host_scripts
./connect_to_qemu_mon_on_host.sh
(qemu)
On host_server_1: Terminal 1

In VM on host_server_1:

Setup DPDK in the VM and run testpmd in the VM.

cd /root/dpdk/vm_scripts
./setup_dpdk_in_vm.sh
./run_testpmd_in_vm.sh

testpmd> show port info all
testpmd> set fwd mac retry
testpmd> start tx_first
testpmd> show port stats all

Virtio traffic is seen at P1 and P2.

On host_server_2: Terminal 1

Set up DPDK on the host_server_2.

cd /root/dpdk/host_scripts
./setup_dpdk_on_host.sh
On host_server_2: Terminal 2

Bind the Niantic or Fortville NIC to igb_uio on host_server_2.

For Fortville NIC.

cd /root/dpdk/usertools
./dpdk-devbind.py -b igb_uio 0000:03:00.0

For Niantic NIC.

cd /root/dpdk/usertools
./dpdk-devbind.py -b igb_uio 0000:06:00.0
On host_server_2: Terminal 3

For Fortville and Niantic NIC’s reset SRIOV, and run the vhost_user sample application on host_server_2.

cd /root/dpdk/host_scripts
./reset_vf_on_212_131.sh
./run_vhost_switch_on_host.sh
On host_server_2: Terminal 1

Start the VM on host_server_2.

./vm_virtio_vhost_user_migrate.sh
On host_server_2: Terminal 4

Connect to the QEMU monitor on host_server_2.

cd /root/dpdk/host_scripts
./connect_to_qemu_mon_on_host.sh
(qemu) info status
VM status: paused (inmigrate)
(qemu)
On host_server_1: Terminal 4

Check that switch is up before migrating the VM.

(qemu) migrate tcp:10.237.212.131:5555
(qemu) info status
VM status: paused (postmigrate)

(qemu) info migrate
capabilities: xbzrle: off rdma-pin-all: off auto-converge: off zero-blocks: off
Migration status: completed
total time: 11619 milliseconds
downtime: 5 milliseconds
setup: 7 milliseconds
transferred ram: 379699 kbytes
throughput: 267.82 mbps
remaining ram: 0 kbytes
total ram: 1590088 kbytes
duplicate: 303985 pages
skipped: 0 pages
normal: 94073 pages
normal bytes: 376292 kbytes
dirty sync count: 2
(qemu) quit
On host_server_2: Terminal 1

In VM on host_server_2:

Hit Enter key. This brings the user to the testpmd prompt.
testpmd>
On host_server_2: Terminal 4

In QEMU monitor on host_server_2

(qemu) info status
VM status: running
On host_server_2: Terminal 1

In VM on host_server_2:

testomd> show port info all
testpmd> show port stats all

Virtio traffic is seen at P0 and P1.

Sample host scripts

reset_vf_on_212_46.sh
#!/bin/sh
# This script is run on the host 10.237.212.46 to reset SRIOV

# BDF for Fortville NIC is 0000:02:00.0
cat /sys/bus/pci/devices/0000\:02\:00.0/max_vfs
echo 0 > /sys/bus/pci/devices/0000\:02\:00.0/max_vfs
cat /sys/bus/pci/devices/0000\:02\:00.0/max_vfs

# BDF for Niantic NIC is 0000:09:00.0
cat /sys/bus/pci/devices/0000\:09\:00.0/max_vfs
echo 0 > /sys/bus/pci/devices/0000\:09\:00.0/max_vfs
cat /sys/bus/pci/devices/0000\:09\:00.0/max_vfs
vm_virtio_vhost_user.sh
#/bin/sh
# Script for use with vhost_user sample application
# The host system has 8 cpu's (0-7)

# Path to KVM tool
KVM_PATH="/usr/bin/qemu-system-x86_64"

# Guest Disk image
DISK_IMG="/home/user/disk_image/virt1_sml.disk"

# Number of guest cpus
VCPUS_NR="6"

# Memory
MEM=1024

VIRTIO_OPTIONS="csum=off,gso=off,guest_tso4=off,guest_tso6=off,guest_ecn=off"

# Socket Path
SOCKET_PATH="/root/dpdk/host_scripts/usvhost"

taskset -c 2-7 $KVM_PATH \
 -enable-kvm \
 -m $MEM \
 -smp $VCPUS_NR \
 -object memory-backend-file,id=mem,size=1024M,mem-path=/mnt/huge,share=on \
 -numa node,memdev=mem,nodeid=0 \
 -cpu host \
 -name VM1 \
 -no-reboot \
 -net none \
 -vnc none \
 -nographic \
 -hda $DISK_IMG \
 -chardev socket,id=chr0,path=$SOCKET_PATH \
 -netdev type=vhost-user,id=net1,chardev=chr0,vhostforce \
 -device virtio-net-pci,netdev=net1,mac=CC:BB:BB:BB:BB:BB,$VIRTIO_OPTIONS \
 -chardev socket,id=chr1,path=$SOCKET_PATH \
 -netdev type=vhost-user,id=net2,chardev=chr1,vhostforce \
 -device virtio-net-pci,netdev=net2,mac=DD:BB:BB:BB:BB:BB,$VIRTIO_OPTIONS \
 -monitor telnet::3333,server,nowait
connect_to_qemu_mon_on_host.sh
#!/bin/sh
# This script is run on both hosts when the VM is up,
# to connect to the Qemu Monitor.

telnet 0 3333
reset_vf_on_212_131.sh
#!/bin/sh
# This script is run on the host 10.237.212.131 to reset SRIOV

# BDF for Ninatic NIC is 0000:06:00.0
cat /sys/bus/pci/devices/0000\:06\:00.0/max_vfs
echo 0 > /sys/bus/pci/devices/0000\:06\:00.0/max_vfs
cat /sys/bus/pci/devices/0000\:06\:00.0/max_vfs

# BDF for Fortville NIC is 0000:03:00.0
cat /sys/bus/pci/devices/0000\:03\:00.0/max_vfs
echo 0 > /sys/bus/pci/devices/0000\:03\:00.0/max_vfs
cat /sys/bus/pci/devices/0000\:03\:00.0/max_vfs
vm_virtio_vhost_user_migrate.sh
#/bin/sh
# Script for use with vhost user sample application
# The host system has 8 cpu's (0-7)

# Path to KVM tool
KVM_PATH="/usr/bin/qemu-system-x86_64"

# Guest Disk image
DISK_IMG="/home/user/disk_image/virt1_sml.disk"

# Number of guest cpus
VCPUS_NR="6"

# Memory
MEM=1024

VIRTIO_OPTIONS="csum=off,gso=off,guest_tso4=off,guest_tso6=off,guest_ecn=off"

# Socket Path
SOCKET_PATH="/root/dpdk/host_scripts/usvhost"

taskset -c 2-7 $KVM_PATH \
 -enable-kvm \
 -m $MEM \
 -smp $VCPUS_NR \
 -object memory-backend-file,id=mem,size=1024M,mem-path=/mnt/huge,share=on \
 -numa node,memdev=mem,nodeid=0 \
 -cpu host \
 -name VM1 \
 -no-reboot \
 -net none \
 -vnc none \
 -nographic \
 -hda $DISK_IMG \
 -chardev socket,id=chr0,path=$SOCKET_PATH \
 -netdev type=vhost-user,id=net1,chardev=chr0,vhostforce \
 -device virtio-net-pci,netdev=net1,mac=CC:BB:BB:BB:BB:BB,$VIRTIO_OPTIONS \
 -chardev socket,id=chr1,path=$SOCKET_PATH \
 -netdev type=vhost-user,id=net2,chardev=chr1,vhostforce \
 -device virtio-net-pci,netdev=net2,mac=DD:BB:BB:BB:BB:BB,$VIRTIO_OPTIONS \
 -incoming tcp:0:5555 \
 -monitor telnet::3333,server,nowait

Sample VM scripts

setup_dpdk_virtio_in_vm.sh
#!/bin/sh
# this script matches the vm_virtio_vhost_user script
# virtio port is 03
# virtio port is 04

cat  /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
echo 1024 > /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
cat  /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages

ifconfig -a
/root/dpdk/usertools/dpdk-devbind.py --status

rmmod virtio-pci

modprobe uio
insmod /root/dpdk/x86_64-default-linuxapp-gcc/kmod/igb_uio.ko

/root/dpdk/usertools/dpdk-devbind.py -b igb_uio 0000:00:03.0
/root/dpdk/usertools/dpdk-devbind.py -b igb_uio 0000:00:04.0

/root/dpdk/usertools/dpdk-devbind.py --status
run_testpmd_in_vm.sh
#!/bin/sh
# Run testpmd for use with vhost_user sample app.
# test system has 8 cpus (0-7), use cpus 2-7 for VM

/root/dpdk/x86_64-default-linuxapp-gcc/app/testpmd \
-l 0-5 -n 4 --socket-mem 350 -- --burst=64 --i --disable-hw-vlan-filter

Flow Bifurcation How-to Guide

Flow Bifurcation is a mechanism which uses hardware capable Ethernet devices to split traffic between Linux user space and kernel space. Since it is a hardware assisted feature this approach can provide line rate processing capability. Other than KNI, the software is just required to enable device configuration, there is no need to take care of the packet movement during the traffic split. This can yield better performance with less CPU overhead.

The Flow Bifurcation splits the incoming data traffic to user space applications (such as DPDK applications) and/or kernel space programs (such as the Linux kernel stack). It can direct some traffic, for example data plane traffic, to DPDK, while directing some other traffic, for example control plane traffic, to the traditional Linux networking stack.

There are a number of technical options to achieve this. A typical example is to combine the technology of SR-IOV and packet classification filtering.

SR-IOV is a PCI standard that allows the same physical adapter to be split as multiple virtual functions. Each virtual function (VF) has separated queues with physical functions (PF). The network adapter will direct traffic to a virtual function with a matching destination MAC address. In a sense, SR-IOV has the capability for queue division.

Packet classification filtering is a hardware capability available on most network adapters. Filters can be configured to direct specific flows to a given receive queue by hardware. Different NICs may have different filter types to direct flows to a Virtual Function or a queue that belong to it.

In this way the Linux networking stack can receive specific traffic through the kernel driver while a DPDK application can receive specific traffic bypassing the Linux kernel by using drivers like VFIO or the DPDK igb_uio module.

_images/flow_bifurcation_overview.svg

Flow Bifurcation Overview

Using Flow Bifurcation on IXGBE in Linux

On Intel 82599 10 Gigabit Ethernet Controller series NICs Flow Bifurcation can be achieved by SR-IOV and Intel Flow Director technologies. Traffic can be directed to queues by the Flow Director capability, typically by matching 5-tuple of UDP/TCP packets.

The typical procedure to achieve this is as follows:

  1. Boot the system without iommu, or with iommu=pt.

  2. Create Virtual Functions:

    echo 2 > /sys/bus/pci/devices/0000:01:00.0/sriov_numvfs
    
  3. Enable and set flow filters:

    ethtool -K eth1 ntuple on
    ethtool -N eth1 flow-type udp4 src-ip 192.0.2.2 dst-ip 198.51.100.2 \
            action $queue_index_in_VF0
    ethtool -N eth1 flow-type udp4 src-ip 198.51.100.2 dst-ip 192.0.2.2 \
            action $queue_index_in_VF1
    

    Where:

    • $queue_index_in_VFn: Bits 39:32 of the variable defines VF id + 1; the lower 32 bits indicates the queue index of the VF. Thus:
      • $queue_index_in_VF0 = (0x1 & 0xFF) << 32 + [queue index].
      • $queue_index_in_VF1 = (0x2 & 0xFF) << 32 + [queue index].
    _images/ixgbe_bifu_queue_idx.svg
  4. Compile the DPDK application and insert igb_uio or probe the vfio-pci kernel modules as normal.

  5. Bind the virtual functions:

    modprobe vfio-pci
    dpdk-devbind.py -b vfio-pci 01:10.0
    dpdk-devbind.py -b vfio-pci 01:10.1
    
  6. Run a DPDK application on the VFs:

    testpmd -l 0-7 -n 4 -- -i -w 01:10.0 -w 01:10.1 --forward-mode=mac
    

In this example, traffic matching the rules will go through the VF by matching the filter rule. All other traffic, not matching the rules, will go through the default queue or scaling on queues in the PF. That is to say UDP packets with the specified IP source and destination addresses will go through the DPDK application. All other traffic, with different hosts or different protocols, will go through the Linux networking stack.

Note

Using Flow Bifurcation on I40E in Linux

On Intel X710/XL710 series Ethernet Controllers Flow Bifurcation can be achieved by SR-IOV, Cloud Filter and L3 VEB switch. The traffic can be directed to queues by the Cloud Filter and L3 VEB switch’s matching rule.

  • L3 VEB filters work for non-tunneled packets. It can direct a packet just by the Destination IP address to a queue in a VF.

  • Cloud filters work for the following types of tunneled packets.

    • Inner mac.
    • Inner mac + VNI.
    • Outer mac + Inner mac + VNI.
    • Inner mac + Inner vlan + VNI.
    • Inner mac + Inner vlan.

The typical procedure to achieve this is as follows:

  1. Boot the system without iommu, or with iommu=pt.

  2. Build and insert the i40e.ko module.

  3. Create Virtual Functions:

    echo 2 > /sys/bus/pci/devices/0000:01:00.0/sriov_numvfs
    
  4. Add udp port offload to the NIC if using cloud filter:

    ip li add vxlan0 type vxlan id 42 group 239.1.1.1 local 10.16.43.214 dev <name>
    ifconfig vxlan0 up
    ip -d li show vxlan0
    

    Note

    Output such as add vxlan port 8472, index 0 success should be found in the system log.

  5. Examples of enabling and setting flow filters:

    • L3 VEB filter, for a route whose destination IP is 192.168.50.108 to VF 0’s queue 2.

      ethtool -N <dev_name> flow-type ip4 dst-ip 192.168.50.108 \
              user-def 0xffffffff00000000 action 2 loc 8
      
    • Inner mac, for a route whose inner destination mac is 0:0:0:0:9:0 to PF’s queue 6.

      ethtool -N <dev_name> flow-type ether dst 00:00:00:00:00:00 \
              m ff:ff:ff:ff:ff:ff src 00:00:00:00:09:00 m 00:00:00:00:00:00 \
              user-def 0xffffffff00000003 action 6 loc 1
      
    • Inner mac + VNI, for a route whose inner destination mac is 0:0:0:0:9:0 and VNI is 8 to PF’s queue 4.

      ethtool -N <dev_name> flow-type ether dst 00:00:00:00:00:00 \
              m ff:ff:ff:ff:ff:ff src 00:00:00:00:09:00 m 00:00:00:00:00:00 \
              user-def 0x800000003 action 4 loc 4
      
    • Outer mac + Inner mac + VNI, for a route whose outer mac is 68:05:ca:24:03:8b, inner destination mac is c2:1a:e1:53:bc:57, and VNI is 8 to PF’s queue 2.

      ethtool -N <dev_name> flow-type ether dst 68:05:ca:24:03:8b \
              m 00:00:00:00:00:00 src c2:1a:e1:53:bc:57 m 00:00:00:00:00:00 \
              user-def 0x800000003 action 2 loc 2
      
    • Inner mac + Inner vlan + VNI, for a route whose inner destination mac is 00:00:00:00:20:00, inner vlan is 10, and VNI is 8 to VF 0’s queue 1.

      ethtool -N <dev_name> flow-type ether dst 00:00:00:00:01:00 \
              m ff:ff:ff:ff:ff:ff src 00:00:00:00:20:00 m 00:00:00:00:00:00 \
              vlan 10 user-def 0x800000000 action 1 loc 5
      
    • Inner mac + Inner vlan, for a route whose inner destination mac is 00:00:00:00:20:00, and inner vlan is 10 to VF 0’s queue 1.

      ethtool -N <dev_name> flow-type ether dst 00:00:00:00:01:00 \
              m ff:ff:ff:ff:ff:ff src 00:00:00:00:20:00 m 00:00:00:00:00:00 \
              vlan 10 user-def 0xffffffff00000000 action 1 loc 5
      

    Note

    • If the upper 32 bits of ‘user-def’ are 0xffffffff, then the filter can be used for programming an L3 VEB filter, otherwise the upper 32 bits of ‘user-def’ can carry the tenant ID/VNI if specified/required.
    • Cloud filters can be defined with inner mac, outer mac, inner ip, inner vlan and VNI as part of the cloud tuple. It is always the destination (not source) mac/ip that these filters use. For all these examples dst and src mac address fields are overloaded dst == outer, src == inner.
    • The filter will direct a packet matching the rule to a vf id specified in the lower 32 bit of user-def to the queue specified by ‘action’.
    • If the vf id specified by the lower 32 bit of user-def is greater than or equal to max_vfs, then the filter is for the PF queues.
  6. Compile the DPDK application and insert igb_uio or probe the vfio-pci kernel modules as normal.

  7. Bind the virtual function:

    modprobe vfio-pci
    dpdk-devbind.py -b vfio-pci 01:10.0
    dpdk-devbind.py -b vfio-pci 01:10.1
    
  8. run DPDK application on VFs:

    testpmd -l 0-7 -n 4 -- -i -w 01:10.0 -w 01:10.1 --forward-mode=mac
    

Note

  • The above steps work on the i40e Linux kernel driver v1.5.16.
  • The Ethtool version used in this example is 3.18. The mask ff means ‘not involved’, while 00 or no mask means ‘involved’.
  • For more details of the configuration, refer to the cloud filter test plan

Generic flow API - examples

This document demonstrates some concrete examples for programming flow rules with the rte_flow APIs.

  • Detail of the rte_flow APIs can be found in the following link: Generic flow API .
  • Details of the TestPMD commands to set the flow rules can be found in the following link: TestPMD Flow rules

Simple IPv4 drop

Description

In this example we will create a simple rule that drops packets whose IPv4 destination equals 192.168.3.2. This code is equivalent to the following testpmd command (wrapped for clarity):

tpmd> flow create 0 ingress pattern eth / vlan /
                  ipv4 dst is 192.168.3.2 / end actions drop / end
Code
/* create the attribute structure */
struct rte_flow_attr attr = {.ingress = 1};
struct rte_flow_item pattern[MAX_PATTERN_IN_FLOW];
struct rte_flow_action actions[MAX_ACTIONS_IN_FLOW];
struct rte_flow_item_etc eth;
struct rte_flow_item_vlan vlan;
struct rte_flow_item_ipv4 ipv4;
struct rte_flow *flow;
struct rte_flow_error error;

/* setting the eth to pass all packets */
pattern[0].type = RTE_FLOW_ITEM_TYPE_ETH;
pattern[0].spec = &eth;

/* set the vlan to pas all packets */
pattern[1] = RTE_FLOW_ITEM_TYPE_VLAN;
pattern[1].spec = &vlan;

/* set the dst ipv4 packet to the required value */
ipv4.hdr.dst_addr = htonl(0xc0a80302);
pattern[2].type = RTE_FLOW_ITEM_TYPE_IPV4;
pattern[2].spec = &ipv4;

/* end the pattern array */
pattern[3].type = RTE_FLOW_ITEM)TYPE_END;

/* create the drop action */
actions[0].type = RTE_FLOW_ACTION_TYPE_DROP;
actions[1].type = RTE_FLOW_ACTION_TYPE_END;

/* validate and create the flow rule */
if (!rte_flow_validate(port_id, &attr, pattern, actions, &error)
    flow = rte_flow_create(port_id, &attr, pattern, actions, &error)
Output

Terminal 1: running sample app with the flow rule disabled:

./filter-program disable
[waiting for packets]

Terminal 2: running scapy:

$scapy
welcome to Scapy
>> sendp(Ether()/Dot1Q()/IP(src='176.80.50.4', dst='192.168.3.1'), \
         iface='some interface', count=1)
>> sendp(Ether()/Dot1Q()/IP(src='176.80.50.5', dst='192.168.3.2'), \
         iface='some interface', count=1)

Terminal 1: output log:

received packet with src ip = 176.80.50.4
received packet with src ip = 176.80.50.5

Terminal 1: running sample the app flow rule enabled:

./filter-program enabled
[waiting for packets]

Terminal 2: running scapy:

$scapy
welcome to Scapy
>> sendp(Ether()/Dot1Q()/IP(src='176.80.50.4', dst='192.168.3.1'),  \
         iface='some interface', count=1)
>> sendp(Ether()/Dot1Q()/IP(src='176.80.50.5', dst ='192.168.3.2'), \
         iface='some interface', count=1)

Terminal 1: output log:

received packet with src ip = 176.80.50.4

Range IPv4 drop

Description

In this example we will create a simple rule that drops packets whose IPv4 destination is in the range 192.168.3.0 to 192.168.3.255. This is done using a mask.

This code is equivalent to the following testpmd command (wrapped for clarity):

tpmd> flow create 0 ingress pattern eth / vlan /
                  ipv4 dst spec 192.168.3.0 dst mask 255.255.255.0 /
                  end actions drop / end
Code
struct rte_flow_attr attr = {.ingress = 1};
struct rte_flow_item pattern[MAX_PATTERN_IN_FLOW];
struct rte_flow_action actions[MAX_ACTIONS_IN_FLOW];
struct rte_flow_item_etc eth;
struct rte_flow_item_vlan vlan;
struct rte_flow_item_ipv4 ipv4;
struct rte_flow_item_ipv4 ipv4_mask;
struct rte_flow *flow;
struct rte_flow_error error;

/* setting the eth to pass all packets */
pattern[0].type = RTE_FLOW_ITEM_TYPE_ETH;
pattern[0].spec = &eth;

/* set the vlan to pas all packets */
pattern[1] = RTE_FLOW_ITEM_TYPE_VLAN;
pattern[1].spec = &vlan;

/* set the dst ipv4 packet to the required value */
ipv4.hdr.dst_addr = htonl(0xc0a80300);
ipv4_mask.hdr.dst_addr = htonl(0xffffff00);
pattern[2].type = RTE_FLOW_ITEM_TYPE_IPV4;
pattern[2].spec = &ipv4;
pattern[2].mask = &ipv4_mask;

/* end the pattern array */
pattern[3].type = RTE_FLOW_ITEM)TYPE_END;

/* create the drop action */
actions[0].type = RTE_FLOW_ACTION_TYPE_DROP;
actions[1].type = RTE_FLOW_ACTION_TYPE_END;

/* validate and create the flow rule */
if (!rte_flow_validate(port_id, &attr, pattern, actions, &error)
    flow = rte_flow_create(port_id, &attr, pattern, actions, &error)
Output

Terminal 1: running sample app flow rule disabled:

./filter-program disable
[waiting for packets]

Terminal 2: running scapy:

$scapy
welcome to Scapy
>> sendp(Ether()/Dot1Q()/IP(src='176.80.50.4', dst='192.168.3.1'), \
         iface='some interface', count=1)
>> sendp(Ether()/Dot1Q()/IP(src='176.80.50.5', dst='192.168.3.2'), \
         iface='some interface', count=1)
>> sendp(Ether()/Dot1Q()/IP(src='176.80.50.6', dst='192.168.5.2'), \
         iface='some interface', count=1)

Terminal 1: output log:

received packet with src ip = 176.80.50.4
received packet with src ip = 176.80.50.5
received packet with src ip = 176.80.50.6

Terminal 1: running sample app flow rule enabled:

./filter-program enabled
[waiting for packets]

Terminal 2: running scapy:

$scapy
welcome to Scapy
>> sendp(Ether()/Dot1Q()/IP(src='176.80.50.4', dst='192.168.3.1'), \
         iface='some interface', count=1)
>> sendp(Ether()/Dot1Q()/IP(src='176.80.50.5', dst='192.168.3.2'), \
         iface='some interface', count=1)
>> sendp(Ether()/Dot1Q()/IP(src='176.80.50.6', dst='192.168.5.2'), \
         iface='some interface', count=1)

Terminal 1: output log:

received packet with src ip = 176.80.50.6

Send vlan to queue

Description

In this example we will create a rule that routes all vlan id 123 to queue 3.

This code is equivalent to the following testpmd command (wrapped for clarity):

tpmd> flow create 0 ingress pattern eth / vlan vid spec 123 /
                  end actions queue index 3 / end
Code
struct rte_flow_attr attr = {.ingress = 1};
struct rte_flow_item pattern[MAX_PATTERN_IN_FLOW];
struct rte_flow_action actions[MAX_ACTIONS_IN_FLOW];
struct rte_flow_item_etc eth;
struct rte_flow_item_vlan vlan;
struct rte_flow_action_queue queue = { .index = 3 };
struct rte_flow *flow;
struct rte_flow_error error;

/* setting the eth to pass all packets */
pattern[0].type = RTE_FLOW_ITEM_TYPE_ETH;
pattern[0].spec = &eth;

/* set the vlan to pas all packets */
vlan.vid = 123;
pattern[1] = RTE_FLOW_ITEM_TYPE_VLAN;
pattern[1].spec = &vlan;

/* end the pattern array */
pattern[2].type = RTE_FLOW_ITEM)TYPE_END;

/* create the drop action */
actions[0].type = RTE_FLOW_ACTION_TYPE_QUEUE;
actions[0].conf = &queue
actions[1].type = RTE_FLOW_ACTION_TYPE_END;

/* validate and create the flow rule */
if (!rte_flow_validate(port_id, &attr, pattern, actions, &error)
    flow = rte_flow_create(port_id, &attr, pattern, actions, &error)
Output

Terminal 1: running sample app flow rule disabled:

./filter-program disable
[waiting for packets]

Terminal 2: running scapy:

$scapy
welcome to Scapy
>> sendp(Ether()/Dot1Q(vlan=123)/IP(src='176.80.50.4', dst='192.168.3.1'), \
         iface='some interface', count=1)
>> sendp(Ether()/Dot1Q(vlan=50)/IP(src='176.80.50.5', dst='192.168.3.2'),  \
         iface='some interface', count=1)
>> sendp(Ether()/Dot1Q(vlan=123)/IP(src='176.80.50.6', dst='192.168.5.2'), \
         iface='some interface', count=1)

Terminal 1: output log:

received packet with src ip = 176.80.50.4 sent to queue 2
received packet with src ip = 176.80.50.5 sent to queue 1
received packet with src ip = 176.80.50.6 sent to queue 0

Terminal 1: running sample app flow rule enabled:

./filter-program enabled
[waiting for packets]

Terminal 2: running scapy:

$scapy
welcome to Scapy
>> sendp(Ether()/Dot1Q(vlan=123)/IP(src='176.80.50.4', dst='192.168.3.1'), \
         iface='some interface', count=1)
>> sendp(Ether()/Dot1Q(vlan=50)/IP(src='176.80.50.5', dst='192.168.3.2'),  \
         iface='some interface', count=1)
>> sendp(Ether()/Dot1Q(vlan=123)/IP(src='176.80.50.6', dst='192.168.5.2'), \
         iface='some interface', count=1)

Terminal 1: output log:

received packet with src ip = 176.80.50.4 sent to queue 3
received packet with src ip = 176.80.50.5 sent to queue 1
received packet with src ip = 176.80.50.6 sent to queue 3

PVP reference benchmark setup using testpmd

This guide lists the steps required to setup a PVP benchmark using testpmd as a simple forwarder between NICs and Vhost interfaces. The goal of this setup is to have a reference PVP benchmark without using external vSwitches (OVS, VPP, ...) to make it easier to obtain reproducible results and to facilitate continuous integration testing.

The guide covers two ways of launching the VM, either by directly calling the QEMU command line, or by relying on libvirt. It has been tested with DPDK v16.11 using RHEL7 for both host and guest.

Setup overview

_images/pvp_2nics.svg

PVP setup using 2 NICs

In this diagram, each red arrow represents one logical core. This use-case requires 6 dedicated logical cores. A forwarding configuration with a single NIC is also possible, requiring 3 logical cores.

Host setup

In this setup, we isolate 6 cores (from CPU2 to CPU7) on the same NUMA node. Two cores are assigned to the VM vCPUs running testpmd and four are assigned to testpmd on the host.

Host tuning
  1. On BIOS, disable turbo-boost and hyper-threads.

  2. Append these options to Kernel command line:

    intel_pstate=disable mce=ignore_ce default_hugepagesz=1G hugepagesz=1G hugepages=6 isolcpus=2-7 rcu_nocbs=2-7 nohz_full=2-7 iommu=pt intel_iommu=on
    
  3. Disable hyper-threads at runtime if necessary or if BIOS is not accessible:

    cat /sys/devices/system/cpu/cpu*[0-9]/topology/thread_siblings_list \
        | sort | uniq \
        | awk -F, '{system("echo 0 > /sys/devices/system/cpu/cpu"$2"/online")}'
    
  4. Disable NMIs:

    echo 0 > /proc/sys/kernel/nmi_watchdog
    
  5. Exclude isolated CPUs from the writeback cpumask:

    echo ffffff03 > /sys/bus/workqueue/devices/writeback/cpumask
    
  6. Isolate CPUs from IRQs:

    clear_mask=0xfc #Isolate CPU2 to CPU7 from IRQs
    for i in /proc/irq/*/smp_affinity
    do
      echo "obase=16;$(( 0x$(cat $i) & ~$clear_mask ))" | bc > $i
    done
    
Qemu build

Build Qemu:

git clone git://git.qemu.org/qemu.git
cd qemu
mkdir bin
cd bin
../configure --target-list=x86_64-softmmu
make
DPDK build

Build DPDK:

git clone git://dpdk.org/dpdk
cd dpdk
export RTE_SDK=$PWD
make install T=x86_64-native-linuxapp-gcc DESTDIR=install
Testpmd launch
  1. Assign NICs to DPDK:

    modprobe vfio-pci
    $RTE_SDK/install/sbin/dpdk-devbind -b vfio-pci 0000:11:00.0 0000:11:00.1
    

    Note

    The Sandy Bridge family seems to have some IOMMU limitations giving poor performance results. To achieve good performance on these machines consider using UIO instead.

  2. Launch the testpmd application:

    $RTE_SDK/install/bin/testpmd -l 0,2,3,4,5 --socket-mem=1024 -n 4 \
        --vdev 'net_vhost0,iface=/tmp/vhost-user1' \
        --vdev 'net_vhost1,iface=/tmp/vhost-user2' -- \
        --portmask=f --disable-hw-vlan -i --rxq=1 --txq=1 \
        --nb-cores=4 --forward-mode=io
    

    With this command, isolated CPUs 2 to 5 will be used as lcores for PMD threads.

  3. In testpmd interactive mode, set the portlist to obtain the correct port chaining:

    set portlist 0,2,1,3
    start
    
VM launch

The VM may be launched either by calling QEMU directly, or by using libvirt.

Qemu way

Launch QEMU with two Virtio-net devices paired to the vhost-user sockets created by testpmd. Below example uses default Virtio-net options, but options may be specified, for example to disable mergeable buffers or indirect descriptors.

<QEMU path>/bin/x86_64-softmmu/qemu-system-x86_64 \
    -enable-kvm -cpu host -m 3072 -smp 3 \
    -chardev socket,id=char0,path=/tmp/vhost-user1 \
    -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \
    -device virtio-net-pci,netdev=mynet1,mac=52:54:00:02:d9:01,addr=0x10 \
    -chardev socket,id=char1,path=/tmp/vhost-user2 \
    -netdev type=vhost-user,id=mynet2,chardev=char1,vhostforce \
    -device virtio-net-pci,netdev=mynet2,mac=52:54:00:02:d9:02,addr=0x11 \
    -object memory-backend-file,id=mem,size=3072M,mem-path=/dev/hugepages,share=on \
    -numa node,memdev=mem -mem-prealloc \
    -net user,hostfwd=tcp::1002$1-:22 -net nic \
    -qmp unix:/tmp/qmp.socket,server,nowait \
    -monitor stdio <vm_image>.qcow2

You can use this qmp-vcpu-pin script to pin vCPUs.

It can be used as follows, for example to pin 3 vCPUs to CPUs 1, 6 and 7, where isolated CPUs 6 and 7 will be used as lcores for Virtio PMDs:

export PYTHONPATH=$PYTHONPATH:<QEMU path>/scripts/qmp
./qmp-vcpu-pin -s /tmp/qmp.socket 1 6 7
Libvirt way

Some initial steps are required for libvirt to be able to connect to testpmd’s sockets.

First, SELinux policy needs to be set to permissive, since testpmd is generally run as root (note, as reboot is required):

cat /etc/selinux/config

# This file controls the state of SELinux on the system.
# SELINUX= can take one of these three values:
#     enforcing  - SELinux security policy is enforced.
#     permissive - SELinux prints warnings instead of enforcing.
#     disabled   - No SELinux policy is loaded.
SELINUX=permissive

# SELINUXTYPE= can take one of three two values:
#     targeted - Targeted processes are protected,
#     minimum  - Modification of targeted policy.
#                Only selected processes are protected.
#     mls      - Multi Level Security protection.
SELINUXTYPE=targeted

Also, Qemu needs to be run as root, which has to be specified in /etc/libvirt/qemu.conf:

user = "root"

Once the domain created, the following snippet is an extract of he most important information (hugepages, vCPU pinning, Virtio PCI devices):

<domain type='kvm'>
  <memory unit='KiB'>3145728</memory>
  <currentMemory unit='KiB'>3145728</currentMemory>
  <memoryBacking>
    <hugepages>
      <page size='1048576' unit='KiB' nodeset='0'/>
    </hugepages>
    <locked/>
  </memoryBacking>
  <vcpu placement='static'>3</vcpu>
  <cputune>
    <vcpupin vcpu='0' cpuset='1'/>
    <vcpupin vcpu='1' cpuset='6'/>
    <vcpupin vcpu='2' cpuset='7'/>
    <emulatorpin cpuset='0'/>
  </cputune>
  <numatune>
    <memory mode='strict' nodeset='0'/>
  </numatune>
  <os>
    <type arch='x86_64' machine='pc-i440fx-rhel7.0.0'>hvm</type>
    <boot dev='hd'/>
  </os>
  <cpu mode='host-passthrough'>
    <topology sockets='1' cores='3' threads='1'/>
    <numa>
      <cell id='0' cpus='0-2' memory='3145728' unit='KiB' memAccess='shared'/>
    </numa>
  </cpu>
  <devices>
    <interface type='vhostuser'>
      <mac address='56:48:4f:53:54:01'/>
      <source type='unix' path='/tmp/vhost-user1' mode='client'/>
      <model type='virtio'/>
      <driver name='vhost' rx_queue_size='256' />
      <address type='pci' domain='0x0000' bus='0x00' slot='0x10' function='0x0'/>
    </interface>
    <interface type='vhostuser'>
      <mac address='56:48:4f:53:54:02'/>
      <source type='unix' path='/tmp/vhost-user2' mode='client'/>
      <model type='virtio'/>
      <driver name='vhost' rx_queue_size='256' />
      <address type='pci' domain='0x0000' bus='0x00' slot='0x11' function='0x0'/>
    </interface>
  </devices>
</domain>

Guest setup

Guest tuning
  1. Append these options to the Kernel command line:

    default_hugepagesz=1G hugepagesz=1G hugepages=1 intel_iommu=on iommu=pt isolcpus=1,2 rcu_nocbs=1,2 nohz_full=1,2
    
  2. Disable NMIs:

    echo 0 > /proc/sys/kernel/nmi_watchdog
    
  3. Exclude isolated CPU1 and CPU2 from the writeback cpumask:

    echo 1 > /sys/bus/workqueue/devices/writeback/cpumask
    
  4. Isolate CPUs from IRQs:

    clear_mask=0x6 #Isolate CPU1 and CPU2 from IRQs
    for i in /proc/irq/*/smp_affinity
    do
      echo "obase=16;$(( 0x$(cat $i) & ~$clear_mask ))" | bc > $i
    done
    
DPDK build

Build DPDK:

git clone git://dpdk.org/dpdk
cd dpdk
export RTE_SDK=$PWD
make install T=x86_64-native-linuxapp-gcc DESTDIR=install
Testpmd launch

Probe vfio module without iommu:

modprobe -r vfio_iommu_type1
modprobe -r vfio
modprobe  vfio enable_unsafe_noiommu_mode=1
cat /sys/module/vfio/parameters/enable_unsafe_noiommu_mode
modprobe vfio-pci

Bind the virtio-net devices to DPDK:

$RTE_SDK/usertools/dpdk-devbind.py -b vfio-pci 0000:00:10.0 0000:00:11.0

Start testpmd:

$RTE_SDK/install/bin/testpmd -l 0,1,2 --socket-mem 1024 -n 4 \
    --proc-type auto --file-prefix pg -- \
    --portmask=3 --forward-mode=macswap --port-topology=chained \
    --disable-hw-vlan --disable-rss -i --rxq=1 --txq=1 \
    --rxd=256 --txd=256 --nb-cores=2 --auto-start

Results template

Below template should be used when sharing results:

Traffic Generator: <Test equipment (e.g. IXIA, Moongen, ...)>
Acceptable Loss: <n>%
Validation run time: <n>min
Host DPDK version/commit: <version, SHA-1>
Guest DPDK version/commit: <version, SHA-1>
Patches applied: <link to patchwork>
QEMU version/commit: <version>
Virtio features: <features (e.g. mrg_rxbuf='off', leave empty if default)>
CPU: <CPU model>, <CPU frequency>
NIC: <NIC model>
Result: <n> Mpps

VF daemon (VFd)

VFd (the VF daemon) is a mechanism which can be used to configure features on a VF (SR-IOV Virtual Function) without direct access to the PF (SR-IOV Physical Function). VFd is an EXPERIMENTAL feature which can only be used in the scenario of DPDK PF with a DPDK VF. If the PF port is driven by the Linux kernel driver then the VFd feature will not work. Currently VFd is only supported by the ixgbe and i40e drivers.

In general VF features cannot be configured directly by an end user application since they are under the control of the PF. The normal approach to configuring a feature on a VF is that an application would call the APIs provided by the VF driver. If the required feature cannot be configured by the VF directly (the most common case) the VF sends a message to the PF through the mailbox on ixgbe and i40e. This means that the availability of the feature depends on whether the appropriate mailbox messages are defined.

DPDK leverages the mailbox interface defined by the Linux kernel driver so that compatibility with the kernel driver can be guaranteed. The downside of this approach is that the availability of messages supported by the kernel become a limitation when the user wants to configure features on the VF.

VFd is a new method of controlling the features on a VF. The VF driver doesn’t talk directly to the PF driver when configuring a feature on the VF. When a VF application (i.e., an application using the VF ports) wants to enable a VF feature, it can send a message to the PF application (i.e., the application using the PF port, which can be the same as the VF application). The PF application will configure the feature for the VF. Obviously, the PF application can also configure the VF features without a request from the VF application.

_images/vf_daemon_overview.svg

VF daemon (VFd) Overview

Compared with the traditional approach the VFd moves the negotiation between VF and PF from the driver level to application level. So the application should define how the negotiation between the VF and PF works, or even if the control should be limited to the PF.

It is the application’s responsibility to use VFd. Consider for example a KVM migration, the VF application may transfer from one VM to another. It is recommended in this case that the PF control the VF features without participation from the VF. Then the VF application has no capability to configure the features. So the user doesn’t need to define the interface between the VF application and the PF application. The service provider should take the control of all the features.

The following sections describe the VFd functionality.

Note

Although VFd is supported by both ixgbe and i40e, please be aware that since the hardware capability is different, the functions supported by ixgbe and i40e are not the same.

Preparing

VFd only can be used in the scenario of DPDK PF + DPDK VF. Users should bind the PF port to igb_uio, then create the VFs based on the DPDK PF host.

The typical procedure to achieve this is as follows:

  1. Boot the system without iommu, or with iommu=pt.

  2. Bind the PF port to igb_uio, for example:

    dpdk-devbind.py -b igb_uio 01:00.0
    
  3. Create a Virtual Function:

    echo 1 > /sys/bus/pci/devices/0000:01:00.0/max_vfs
    
  4. Start a VM with the new VF port bypassed to it.

  5. Run a DPDK application on the PF in the host:

    testpmd -l 0-7 -n 4 -- -i --txqflags=0
    
  6. Bind the VF port to igb_uio in the VM:

    dpdk-devbind.py -b igb_uio 03:00.0
    
  7. Run a DPDK application on the VF in the VM:

    testpmd -l 0-7 -n 4 -- -i --txqflags=0
    

Common functions of IXGBE and I40E

The following sections show how to enable PF/VF functionality based on the above testpmd setup.

TX loopback

Run a testpmd runtime command on the PF to set TX loopback:

set tx loopback 0 on|off

This sets whether the PF port and all the VF ports that belong to it are allowed to send the packets to other virtual ports.

Although it is a VFd function, it is the global setting for the whole physical port. When using this function, the PF and all the VFs TX loopback will be enabled/disabled.

VF MAC address setting

Run a testpmd runtime command on the PF to set the MAC address for a VF port:

set vf mac addr 0 0 A0:36:9F:7B:C3:51

This testpmd runtime command will change the MAC address of the VF port to this new address. If any other addresses are set before, they will be overwritten.

VF MAC anti-spoofing

Run a testpmd runtime command on the PF to enable/disable the MAC anti-spoofing for a VF port:

set vf mac antispoof 0 0 on|off

When enabling the MAC anti-spoofing, the port will not forward packets whose source MAC address is not the same as the port.

VF VLAN anti-spoofing

Run a testpmd runtime command on the PF to enable/disable the VLAN anti-spoofing for a VF port:

set vf vlan antispoof 0 0 on|off

When enabling the VLAN anti-spoofing, the port will not send packets whose VLAN ID does not belong to VLAN IDs that this port can receive.

VF VLAN insertion

Run a testpmd runtime command on the PF to set the VLAN insertion for a VF port:

set vf vlan insert 0 0 1

When using this testpmd runtime command, an assigned VLAN ID can be inserted to the transmitted packets by the hardware.

The assigned VLAN ID can be 0. It means disabling the VLAN insertion.

VF VLAN stripping

Run a testpmd runtime command on the PF to enable/disable the VLAN stripping for a VF port:

set vf vlan stripq 0 0 on|off

This testpmd runtime command is used to enable/disable the RX VLAN stripping for a specific VF port.

VF VLAN filtering

Run a testpmd runtime command on the PF to set the VLAN filtering for a VF port:

rx_vlan add 1 port 0 vf 1
rx_vlan rm  1 port 0 vf 1

These two testpmd runtime commands can be used to add or remove the VLAN filter for several VF ports. When the VLAN filters are added only the packets that have the assigned VLAN IDs can be received. Other packets will be dropped by hardware.

The IXGBE specific VFd functions

The functions in this section are specific to the ixgbe driver.

All queues drop

Run a testpmd runtime command on the PF to enable/disable the all queues drop:

set all queues drop on|off

This is a global setting for the PF and all the VF ports of the physical port.

Enabling the all queues drop feature means that when there is no available descriptor for the received packets they are dropped. The all queues drop feature should be enabled in SR-IOV mode to avoid one queue blocking others.

VF packet drop

Run a testpmd runtime command on the PF to enable/disable the packet drop for a specific VF:

set vf split drop 0 0 on|off

This is a similar function as all queues drop. The difference is that this function is per VF setting and the previous function is a global setting.

VF rate limit

Run a testpmd runtime command on the PF to all queues’ rate limit for a specific VF:

set port 0 vf 0 rate 10 queue_mask 1

This is a function to set the rate limit for all the queues in the queue_mask bitmap. It is not used to set the summary of the rate limit. The rate limit of every queue will be set equally to the assigned rate limit.

VF RX enabling

Run a testpmd runtime command on the PF to enable/disable packet receiving for a specific VF:

set port 0 vf 0 rx on|off

This function can be used to stop/start packet receiving on a VF.

VF TX enabling

Run a testpmd runtime command on the PF to enable/disable packet transmitting for a specific VF:

set port 0 vf 0 tx on|off

This function can be used to stop/start packet transmitting on a VF.

VF RX mode setting

Run a testpmd runtime command on the PF to set the RX mode for a specific VF:

set port 0 vf 0 rxmode AUPE|ROPE|BAM|MPE on|off

This function can be used to enable/disable some RX modes on the VF, including:

  • If it accept untagged packets.
  • If it accepts packets matching the MAC filters.
  • If it accept MAC broadcast packets,
  • If it enables MAC multicast promiscuous mode.

The I40E specific VFd functions

The functions in this section are specific to the i40e driver.

VF statistics

This provides an API to get the a specific VF’s statistic from PF.

VF statistics resetting

This provides an API to rest the a specific VF’s statistic from PF.

VF MAC broadcast setting

Run a testpmd runtime command on the PF to enable/disable MAC broadcast packet receiving for a specific VF:

set vf broadcast 0 0 on|off
VF MAC multicast promiscuous mode

Run a testpmd runtime command on the PF to enable/disable MAC multicast promiscuous mode for a specific VF:

set vf allmulti 0 0 on|off
VF MAC unicast promiscuous mode

Run a testpmd runtime command on the PF to enable/disable MAC unicast promiscuous mode for a specific VF:

set vf promisc 0 0 on|off
VF max bandwidth

Run a testpmd runtime command on the PF to set the TX maximum bandwidth for a specific VF:

set vf tx max-bandwidth 0 0 2000

The maximum bandwidth is an absolute value in Mbps.

VF TC bandwidth allocation

Run a testpmd runtime command on the PF to set the TCs (traffic class) TX bandwidth allocation for a specific VF:

set vf tc tx min-bandwidth 0 0 (20,20,20,40)

The allocated bandwidth should be set for all the TCs. The allocated bandwidth is a relative value as a percentage. The sum of all the bandwidth should be 100.

VF TC max bandwidth

Run a testpmd runtime command on the PF to set the TCs TX maximum bandwidth for a specific VF:

set vf tc tx max-bandwidth 0 0 0 10000

The maximum bandwidth is an absolute value in Mbps.

TC strict priority scheduling

Run a testpmd runtime command on the PF to enable/disable several TCs TX strict priority scheduling:

set tx strict-link-priority 0 0x3

The 0 in the TC bitmap means disabling the strict priority scheduling for this TC. To enable use a value of 1.

Virtio_user for Container Networking

Container becomes more and more popular for strengths, like low overhead, fast boot-up time, and easy to deploy, etc. How to use DPDK to accelerate container networking becomes a common question for users. There are two use models of running DPDK inside containers, as shown in Fig. 7.3.

_images/use_models_for_running_dpdk_in_containers.svg

Use models of running DPDK inside container

This page will only cover aggregation model.

Overview

The virtual device, virtio-user, with unmodified vhost-user backend, is designed for high performance user space container networking or inter-process communication (IPC).

The overview of accelerating container networking by virtio-user is shown in Fig. 7.4.

_images/virtio_user_for_container_networking.svg

Overview of accelerating container networking by virtio-user

Different virtio PCI devices we usually use as a para-virtualization I/O in the context of QEMU/VM, the basic idea here is to present a kind of virtual devices, which can be attached and initialized by DPDK. The device emulation layer by QEMU in VM’s context is saved by just registering a new kind of virtual device in DPDK’s ether layer. And to minimize the change, we reuse already-existing virtio PMD code (driver/net/virtio/).

Virtio, in essence, is a shm-based solution to transmit/receive packets. How is memory shared? In VM’s case, qemu always shares the whole physical layout of VM to vhost backend. But it’s not feasible for a container, as a process, to share all virtual memory regions to backend. So only those virtual memory regions (aka, hugepages initialized in DPDK) are sent to backend. It restricts that only addresses in these areas can be used to transmit or receive packets.

Sample Usage

Here we use Docker as container engine. It also applies to LXC, Rocket with some minor changes.

  1. Compile DPDK.

    make install RTE_SDK=`pwd` T=x86_64-native-linuxapp-gcc
    
  2. Write a Dockerfile like below.

    cat <<EOT >> Dockerfile
    FROM ubuntu:latest
    WORKDIR /usr/src/dpdk
    COPY . /usr/src/dpdk
    ENV PATH "$PATH:/usr/src/dpdk/x86_64-native-linuxapp-gcc/app/"
    EOT
    
  3. Build a Docker image.

    docker build -t dpdk-app-testpmd .
    
  4. Start a testpmd on the host with a vhost-user port.

    $(testpmd) -l 0-1 -n 4 --socket-mem 1024,1024 \
        --vdev 'eth_vhost0,iface=/tmp/sock0' \
        --file-prefix=host --no-pci -- -i
    
  5. Start a container instance with a virtio-user port.

    docker run -i -t -v /tmp/sock0:/var/run/usvhost \
        -v /dev/hugepages:/dev/hugepages \
        dpdk-app-testpmd testpmd -l 6-7 -n 4 -m 1024 --no-pci \
        --vdev=virtio_user0,path=/var/run/usvhost \
        --file-prefix=container \
        -- -i --txqflags=0xf00 --disable-hw-vlan
    

Note: If we run all above setup on the host, it’s a shm-based IPC.

Limitations

We have below limitations in this solution:
  • Cannot work with –huge-unlink option. As we need to reopen the hugepage file to share with vhost backend.
  • Cannot work with –no-huge option. Currently, DPDK uses anonymous mapping under this option which cannot be reopened to share with vhost backend.
  • Cannot work when there are more than VHOST_MEMORY_MAX_NREGIONS(8) hugepages. In another word, do not use 2MB hugepage so far.
  • Applications should not use file name like HUGEFILE_FMT (“%smap_%d”). That will bring confusion when sharing hugepage files with backend by name.
  • Root privilege is a must. DPDK resolves physical addresses of hugepages which seems not necessary, and some discussions are going on to remove this restriction.

Virtio_user as Exceptional Path

The virtual device, virtio-user, was originally introduced with vhost-user backend, as a high performance solution for IPC (Inter-Process Communication) and user space container networking.

Virtio_user with vhost-kernel backend is a solution for exceptional path, such as KNI which exchanges packets with kernel networking stack. This solution is very promising in:

  • Maintenance

    All kernel modules needed by this solution, vhost and vhost-net (kernel), are upstreamed and extensively used kernel module.

  • Features

    vhost-net is born to be a networking solution, which has lots of networking related featuers, like multi queue, tso, multi-seg mbuf, etc.

  • Performance

    similar to KNI, this solution would use one or more kthreads to send/receive packets to/from user space DPDK applications, which has little impact on user space polling thread (except that it might enter into kernel space to wake up those kthreads if necessary).

The overview of an application using virtio-user as exceptional path is shown in Fig. 8.1.

_images/virtio_user_as_exceptional_path.svg

Overview of a DPDK app using virtio-user as excpetional path

Sample Usage

As a prerequisite, the vhost/vhost-net kernel CONFIG should be chosen before compiling the kernel and those kernel modules should be inserted.

  1. Compile DPDK and bind a physical NIC to igb_uio/uio_pci_generic/vfio-pci.

    This physical NIC is for communicating with outside.

  2. Run testpmd.

    $(testpmd) -l 2-3 -n 4 \
            --vdev=virtio_user0,path=/dev/vhost-net,queue_size=1024 \
            -- -i --txqflags=0x0 --disable-hw-vlan --enable-lro \
            --enable-rx-cksum --rxd=1024 --txd=1024
    

    This command runs testpmd with two ports, one physical NIC to communicate with outside, and one virtio-user to communicate with kernel.

  • --enable-lro

    This is used to negotiate VIRTIO_NET_F_GUEST_TSO4 and VIRTIO_NET_F_GUEST_TSO6 feature so that large packets from kernel can be transmitted to DPDK application and further TSOed by physical NIC.

  • --enable-rx-cksum

    This is used to negotiate VIRTIO_NET_F_GUEST_CSUM so that packets from kernel can be deemed as valid Rx checksumed.

  • queue_size

    256 by default. To avoid shortage of descriptors, we can increase it to 1024.

  • queues

    Number of multi-queues. Each qeueue will be served by a kthread. For example:

    $(testpmd) -l 2-3 -n 4 \
            --vdev=virtio_user0,path=/dev/vhost-net,queues=2,queue_size=1024 \
            -- -i --txqflags=0x0 --disable-hw-vlan --enable-lro \
            --enable-rx-cksum --txq=2 --rxq=2 --rxd=1024 \
            --txd=1024
    
  1. Start testpmd:

    (testpmd) start
    
  2. Configure IP address and start tap:

    ifconfig tap0 1.1.1.1/24 up
    

Note

The tap device will be named tap0, tap1, etc, by kernel.

Then, all traffic from physical NIC can be forwarded into kernel stack, and all traffic on the tap0 can be sent out from physical NIC.

Limitations

This solution is only available on Linux systems.

DPDK pdump Library and pdump Tool

This document describes how the Data Plane Development Kit (DPDK) Packet Capture Framework is used for capturing packets on DPDK ports. It is intended for users of DPDK who want to know more about the Packet Capture feature and for those who want to monitor traffic on DPDK-controlled devices.

The DPDK packet capture framework was introduced in DPDK v16.07. The DPDK packet capture framework consists of the DPDK pdump library and DPDK pdump tool.

Introduction

The librte_pdump library provides the APIs required to allow users to initialize the packet capture framework and to enable or disable packet capture. The library works on a client/server model and its usage is recommended for debugging purposes.

The dpdk-pdump tool is developed based on the librte_pdump library. It runs as a DPDK secondary process and is capable of enabling or disabling packet capture on DPDK ports. The dpdk-pdump tool provides command-line options with which users can request enabling or disabling of the packet capture on DPDK ports.

The application which initializes the packet capture framework will act as a server and the application that enables or disables the packet capture will act as a client. The server sends the Rx and Tx packets from the DPDK ports to the client.

In DPDK the testpmd application can be used to initialize the packet capture framework and act as a server, and the dpdk-pdump tool acts as a client. To view Rx or Tx packets of testpmd, the application should be launched first, and then the dpdk-pdump tool. Packets from testpmd will be sent to the tool, which then sends them on to the Pcap PMD device and that device writes them to the Pcap file or to an external interface depending on the command-line option used.

Some things to note:

  • The dpdk-pdump tool can only be used in conjunction with a primary application which has the packet capture framework initialized already. In dpdk, only testpmd is modified to initialize packet capture framework, other applications remain untouched. So, if the dpdk-pdump tool has to be used with any application other than the testpmd, the user needs to explicitly modify that application to call the packet capture framework initialization code. Refer to the app/test-pmd/testpmd.c code and look for pdump keyword to see how this is done.
  • The dpdk-pdump tool depends on the libpcap based PMD which is disabled by default in the build configuration files, owing to an external dependency on the libpcap development files. Once the libpcap development files are installed, the libpcap based PMD can be enabled by setting CONFIG_RTE_LIBRTE_PMD_PCAP=y and recompiling the DPDK.

Test Environment

The overview of using the Packet Capture Framework and the dpdk-pdump tool for packet capturing on the DPDK port in Fig. 9.1.

_images/packet_capture_framework.svg

Packet capturing on a DPDK port using the dpdk-pdump tool.

Configuration

Modify the DPDK primary application to initialize the packet capture framework as mentioned in the above notes and enable the following config options and build DPDK:

CONFIG_RTE_LIBRTE_PMD_PCAP=y
CONFIG_RTE_LIBRTE_PDUMP=y

Running the Application

The following steps demonstrate how to run the dpdk-pdump tool to capture Rx side packets on dpdk_port0 in Fig. 9.1 and inspect them using tcpdump.

  1. Launch testpmd as the primary application:

    sudo ./app/testpmd -c 0xf0 -n 4 -- -i --port-topology=chained
    
  2. Launch the pdump tool as follows:

    sudo ./build/app/dpdk-pdump -- \
         --pdump 'port=0,queue=*,rx-dev=/tmp/capture.pcap'
    
  3. Send traffic to dpdk_port0 from traffic generator. Inspect packets captured in the file capture.pcap using a tool that can interpret Pcap files, for example tcpdump:

    $tcpdump -nr /tmp/capture.pcap
    reading from file /tmp/capture.pcap, link-type EN10MB (Ethernet)
    11:11:36.891404 IP 4.4.4.4.whois++ > 3.3.3.3.whois++: UDP, length 18
    11:11:36.891442 IP 4.4.4.4.whois++ > 3.3.3.3.whois++: UDP, length 18
    11:11:36.891445 IP 4.4.4.4.whois++ > 3.3.3.3.whois++: UDP, length 18
    

DPDK Tools User Guides

dpdk-procinfo Application

The dpdk-procinfo application is a Data Plane Development Kit (DPDK) application that runs as a DPDK secondary process and is capable of retrieving port statistics, resetting port statistics and printing DPDK memory information. This application extends the original functionality that was supported by dump_cfg.

Running the Application

The application has a number of command line options:

./$(RTE_TARGET)/app/dpdk-procinfo -- -m | [-p PORTMASK] [--stats | --xstats |
--stats-reset | --xstats-reset]
Parameters

-p PORTMASK: Hexadecimal bitmask of ports to configure.

–stats The stats parameter controls the printing of generic port statistics. If no port mask is specified stats are printed for all DPDK ports.

–xstats The xstats parameter controls the printing of extended port statistics. If no port mask is specified xstats are printed for all DPDK ports.

–stats-reset The stats-reset parameter controls the resetting of generic port statistics. If no port mask is specified, the generic stats are reset for all DPDK ports.

–xstats-reset The xstats-reset parameter controls the resetting of extended port statistics. If no port mask is specified xstats are reset for all DPDK ports.

-m: Print DPDK memory information.

dpdk-pdump Application

The dpdk-pdump tool is a Data Plane Development Kit (DPDK) tool that runs as a DPDK secondary process and is capable of enabling packet capture on dpdk ports.

Note

  • The dpdk-pdump tool can only be used in conjunction with a primary application which has the packet capture framework initialized already. In dpdk, only the testpmd is modified to initialize packet capture framework, other applications remain untouched. So, if the dpdk-pdump tool has to be used with any application other than the testpmd, user needs to explicitly modify that application to call packet capture framework initialization code. Refer app/test-pmd/testpmd.c code to see how this is done.
  • The dpdk-pdump tool depends on libpcap based PMD which is disabled by default in the build configuration files, owing to an external dependency on the libpcap development files which must be installed on the board. Once the libpcap development files are installed, the libpcap based PMD can be enabled by setting CONFIG_RTE_LIBRTE_PMD_PCAP=y and recompiling the DPDK.

Running the Application

The tool has a number of command line options:

./build/app/dpdk-pdump --
                       --pdump '(port=<port id> | device_id=<pci id or vdev name>),
                                (queue=<queue_id>),
                                (rx-dev=<iface or pcap file> |
                                 tx-dev=<iface or pcap file>),
                                [ring-size=<ring size>],
                                [mbuf-size=<mbuf data size>],
                                [total-num-mbufs=<number of mbufs>]'
                       [--server-socket-path=<server socket dir>]
                       [--client-socket-path=<client socket dir>]

The --pdump command line option is mandatory and it takes various sub arguments which are described in below section.

Note

  • Parameters inside the parentheses represents mandatory parameters.
  • Parameters inside the square brackets represents optional parameters.
  • Multiple instances of --pdump can be passed to capture packets on different port and queue combinations.

The --server-socket-path command line option is optional. This represents the server socket directory. If no value is passed default values are used i.e. /var/run/.dpdk/ for root users and ~/.dpdk/ for non root users.

The --client-socket-path command line option is optional. This represents the client socket directory. If no value is passed default values are used i.e. /var/run/.dpdk/ for root users and ~/.dpdk/ for non root users.

The --pdump parameters

port: Port id of the eth device on which packets should be captured.

device_id: PCI address (or) name of the eth device on which packets should be captured.

Note

  • As of now the dpdk-pdump tool cannot capture the packets of virtual devices in the primary process due to a bug in the ethdev library. Due to this bug, in a multi process context, when the primary and secondary have different ports set, then the secondary process (here the dpdk-pdump tool) overwrites the rte_eth_devices[] entries of the primary process.

queue: Queue id of the eth device on which packets should be captured. The user can pass a queue value of * to enable packet capture on all queues of the eth device.

rx-dev: Can be either a pcap file name or any Linux iface.

tx-dev: Can be either a pcap file name or any Linux iface.

Note

  • To receive ingress packets only, rx-dev should be passed.
  • To receive egress packets only, tx-dev should be passed.
  • To receive ingress and egress packets separately rx-dev and tx-dev should both be passed with the different file names or the Linux iface names.
  • To receive ingress and egress packets together, rx-dev and tx-dev should both be passed with the same file name or the same Linux iface name.

ring-size: Size of the ring. This value is used internally for ring creation. The ring will be used to enqueue the packets from the primary application to the secondary. This is an optional parameter with default size 16384.

mbuf-size: Size of the mbuf data. This is used internally for mempool creation. Ideally this value must be same as the primary application’s mempool’s mbuf data size which is used for packet RX. This is an optional parameter with default size 2176.

total-num-mbufs: Total number mbufs in mempool. This is used internally for mempool creation. This is an optional parameter with default value 65535.

Example

$ sudo ./build/app/dpdk-pdump -- --pdump 'port=0,queue=*,rx-dev=/tmp/rx.pcap'

dpdk-pmdinfo Application

The dpdk-pmdinfo tool is a Data Plane Development Kit (DPDK) utility that can dump a PMDs hardware support info.

Running the Application

The tool has a number of command line options:

dpdk-pmdinfo [-hrtp] [-d <pci id file] <elf-file>

-h, --help            Show a short help message and exit
-r, --raw             Dump as raw json strings
-d FILE, --pcidb=FILE Specify a pci database to get vendor names from
-t, --table           Output information on hw support as a hex table
-p, --plugindir       Scan dpdk for autoload plugins

Note

  • Parameters inside the square brackets represents optional parameters.

dpdk-devbind Application

The dpdk-devbind tool is a Data Plane Development Kit (DPDK) utility that helps binding and unbinding devices from specific drivers. As well as checking their status in that regard.

Running the Application

The tool has a number of command line options:

dpdk-devbind [options] DEVICE1 DEVICE2 ....

OPTIONS

  • --help, --usage

    Display usage information and quit

  • -s, --status

    Print the current status of all known network interfaces. For each device, it displays the PCI domain, bus, slot and function, along with a text description of the device. Depending upon whether the device is being used by a kernel driver, the igb_uio driver, or no driver, other relevant information will be displayed: - the Linux interface name e.g. if=eth0 - the driver being used e.g. drv=igb_uio - any suitable drivers not currently using that device e.g. unused=igb_uio NOTE: if this flag is passed along with a bind/unbind option, the status display will always occur after the other operations have taken place.

  • -b driver, --bind=driver

    Select the driver to use or “none” to unbind the device

  • -u, --unbind

    Unbind a device (Equivalent to -b none)

  • --force

    By default, devices which are used by Linux - as indicated by having routes in the routing table - cannot be modified. Using the --force flag overrides this behavior, allowing active links to be forcibly unbound. WARNING: This can lead to loss of network connection and should be used with caution.

Warning

Due to the way VFIO works, there are certain limitations to which devices can be used with VFIO. Mainly it comes down to how IOMMU groups work. Any Virtual Function device can be used with VFIO on its own, but physical devices will require either all ports bound to VFIO, or some of them bound to VFIO while others not being bound to anything at all.

If your device is behind a PCI-to-PCI bridge, the bridge will then be part of the IOMMU group in which your device is in. Therefore, the bridge driver should also be unbound from the bridge PCI device for VFIO to work with devices behind the bridge.

Warning

While any user can run the dpdk-devbind.py script to view the status of the network ports, binding or unbinding network ports requires root privileges.

Examples

To display current device status:

dpdk-devbind --status

To bind eth1 from the current driver and move to use igb_uio:

dpdk-devbind --bind=igb_uio eth1

To unbind 0000:01:00.0 from using any driver:

dpdk-devbind -u 0000:01:00.0

To bind 0000:02:00.0 and 0000:02:00.1 to the ixgbe kernel driver:

dpdk-devbind -b ixgbe 02:00.0 02:00.1

To check status of all network ports, assign one to the igb_uio driver and check status again:

# Check the status of the available devices.
dpdk-devbind --status
Network devices using DPDK-compatible driver
============================================
<none>

Network devices using kernel driver
===================================
0000:0a:00.0 '82599ES 10-Gigabit' if=eth2 drv=ixgbe unused=


# Bind the device to igb_uio.
sudo dpdk-devbind -b igb_uio 0000:0a:00.0


# Recheck the status of the devices.
dpdk-devbind --status
Network devices using DPDK-compatible driver
============================================
0000:0a:00.0 '82599ES 10-Gigabit' drv=igb_uio unused=

dpdk-test-crypto-perf Application

The dpdk-test-crypto-perf tool is a Data Plane Development Kit (DPDK) utility that allows measuring performance parameters of PMDs available in the crypto tree. There are available two measurement types: throughput and latency. User can use multiply cores to run tests on but only one type of crypto PMD can be measured during single application execution. Cipher parameters, type of device, type of operation and chain mode have to be specified in the command line as application parameters. These parameters are checked using device capabilities structure.

Limitations

On hardware devices the cycle-count doesn’t always represent the actual offload cost. The cycle-count only represents the offload cost when the hardware accelerator is not fully loaded, when loaded the cpu cycles freed up by the offload are still consumed by the test tool and included in the cycle-count. These cycles are consumed by retries and inefficient API calls enqueuing and dequeuing smaller bursts than specified by the cmdline parameter. This results in a larger cycle-count measurement and should not be interpreted as an offload cost measurement. Using “pmd-cyclecount” mode will give a better idea of actual costs of hardware acceleration.

On hardware devices the throughput measurement is not necessarily the maximum possible for the device, e.g. it may be necessary to use multiple cores to keep the hardware accelerator fully loaded and so measure maximum throughput.

Compiling the Application

Step 1: PMD setting

The dpdk-test-crypto-perf tool depends on crypto device drivers PMD which are disabled by default in the build configuration file common_base. The crypto device drivers PMD which should be tested can be enabled by setting:

CONFIG_RTE_LIBRTE_PMD_<name>=y

Setting example for open ssl PMD:

CONFIG_RTE_LIBRTE_PMD_OPENSSL=y

Step 2: Linearization setting

It is possible linearized input segmented packets just before crypto operation for devices which doesn’t support scatter-gather, and allows to measure performance also for this use case.

To set on the linearization options add below definition to the cperf_ops.h file:

#define CPERF_LINEARIZATION_ENABLE

Step 3: Build the application

Execute the dpdk-setup.sh script to build the DPDK library together with the dpdk-test-crypto-perf applcation.

Initially, the user must select a DPDK target to choose the correct target type and compiler options to use when building the libraries. The user must have all libraries, modules, updates and compilers installed in the system prior to this, as described in the earlier chapters in this Getting Started Guide.

Running the Application

The tool application has a number of command line options:

dpdk-test-crypto-perf [EAL Options] -- [Application Options]
EAL Options

The following are the EAL command-line options that can be used in conjunction with the dpdk-test-crypto-perf applcation. See the DPDK Getting Started Guides for more information on these options.

  • -c <COREMASK> or -l <CORELIST>

    Set the hexadecimal bitmask of the cores to run on. The corelist is a list cores to use.

  • -w <PCI>

    Add a PCI device in white list.

  • --vdev <driver><id>

    Add a virtual device.

Appication Options

The following are the appication command-line options:

  • --ptest type

    Set test type, where type is one of the following:

    throughput
    latency
    verify
    pmd-cyclecount
    
  • --silent

    Disable options dump.

  • --pool-sz <n>

    Set the number of mbufs to be allocated in the mbuf pool.

  • --total-ops <n>

    Set the number of total operations performed.

  • --burst-sz <n>

    Set the number of packets per burst.

    This can be set as:
    • Single value (i.e. --burst-sz 16)
    • Range of values, using the following structure min:inc:max, where min is minimum size, inc is the increment size and max is the maximum size (i.e. --burst-sz 16:2:32)
    • List of values, up to 32 values, separated in commas (i.e. --burst-sz 16,24,32)
  • --buffer-sz <n>

    Set the size of single packet (plaintext or ciphertext in it).

    This can be set as:
    • Single value (i.e. --buffer-sz 16)
    • Range of values, using the following structure min:inc:max, where min is minimum size, inc is the increment size and max is the maximum size (i.e. --buffer-sz 16:2:32)
    • List of values, up to 32 values, separated in commas (i.e. --buffer-sz 32,64,128)
  • --segment-sz <n>

    Set the size of the segment to use, for Scatter Gather List testing. By default, it is set to the size of the maximum buffer size, including the digest size, so a single segment is created.

  • --devtype <name>

    Set device type, where name is one of the following:

    crypto_null
    crypto_aesni_mb
    crypto_aesni_gcm
    crypto_openssl
    crypto_qat
    crypto_snow3g
    crypto_kasumi
    crypto_zuc
    crypto_dpaa_sec
    crypto_dpaa2_sec
    crypto_armv8
    crypto_scheduler
    crypto_mrvl
    
  • --optype <name>

    Set operation type, where name is one of the following:

    cipher-only
    auth-only
    cipher-then-auth
    auth-then-cipher
    aead
    

    For GCM/CCM algorithms you should use aead flag.

  • --sessionless

    Enable session-less crypto operations mode.

  • --out-of-place

    Enable out-of-place crypto operations mode.

  • --test-file <name>

    Set test vector file path. See the Test Vector File chapter.

  • --test-name <name>

    Set specific test name section in the test vector file.

  • --cipher-algo <name>

    Set cipher algorithm name, where name is one of the following:

    3des-cbc
    3des-ecb
    3des-ctr
    aes-cbc
    aes-ctr
    aes-ecb
    aes-f8
    aes-xts
    arc4
    null
    kasumi-f8
    snow3g-uea2
    zuc-eea3
    
  • --cipher-op <mode>

    Set cipher operation mode, where mode is one of the following:

    encrypt
    decrypt
    
  • --cipher-key-sz <n>

    Set the size of cipher key.

  • --cipher-iv-sz <n>

    Set the size of cipher iv.

  • --auth-algo <name>

    Set authentication algorithm name, where name is one of the following:

    3des-cbc
    aes-cbc-mac
    aes-cmac
    aes-gmac
    aes-xcbc-mac
    md5
    md5-hmac
    sha1
    sha1-hmac
    sha2-224
    sha2-224-hmac
    sha2-256
    sha2-256-hmac
    sha2-384
    sha2-384-hmac
    sha2-512
    sha2-512-hmac
    kasumi-f9
    snow3g-uia2
    zuc-eia3
    
  • --auth-op <mode>

    Set authentication operation mode, where mode is one of the following:

    verify
    generate
    
  • --auth-key-sz <n>

    Set the size of authentication key.

  • --auth-iv-sz <n>

    Set the size of auth iv.

  • --aead-algo <name>

    Set AEAD algorithm name, where name is one of the following:

    aes-ccm
    aes-gcm
    
  • --aead-op <mode>

    Set AEAD operation mode, where mode is one of the following:

    encrypt
    decrypt
    
  • --aead-key-sz <n>

    Set the size of AEAD key.

  • --aead-iv-sz <n>

    Set the size of AEAD iv.

  • --aead-aad-sz <n>

    Set the size of AEAD aad.

  • --digest-sz <n>

    Set the size of digest.

  • --desc-nb <n>

    Set number of descriptors for each crypto device.

  • --pmd-cyclecount-delay-ms <n>

    Add a delay (in milliseconds) between enqueue and dequeue in pmd-cyclecount benchmarking mode (useful when benchmarking hardware acceleration).

  • --csv-friendly

    Enable test result output CSV friendly rather than human friendly.

Test Vector File

The test vector file is a text file contain information about test vectors. The file is made of the sections. The first section doesn’t have header. It contain global information used in each test variant vectors - typically information about plaintext, ciphertext, cipher key, aut key, initial vector. All other sections begin header. The sections contain particular information typically digest.

Format of the file:

Each line beginig with sign ‘#’ contain comment and it is ignored by parser:

# <comment>

Header line is just name in square bracket:

[<section name>]

Data line contain information tocken then sign ‘=’ and a string of bytes in C byte array format:

<tocken> = <C byte array>

Tockens list:

  • plaintext

    Original plaintext to be crypted.

  • ciphertext

    Encrypted plaintext string.

  • cipher_key

    Key used in cipher operation.

  • auth_key

    Key used in auth operation.

  • cipher_iv

    Cipher Initial Vector.

  • auth_iv

    Auth Initial Vector.

  • aad

    Additional data.

  • digest

    Digest string.

Examples

Call application for performance throughput test of single Aesni MB PMD for cipher encryption aes-cbc and auth generation sha1-hmac, one million operations, burst size 32, packet size 64:

dpdk-test-crypto-perf -l 6-7 --vdev crypto_aesni_mb -w 0000:00:00.0 --
--ptest throughput --devtype crypto_aesni_mb --optype cipher-then-auth
--cipher-algo aes-cbc --cipher-op encrypt --cipher-key-sz 16 --auth-algo
sha1-hmac --auth-op generate --auth-key-sz 64 --digest-sz 12
--total-ops 10000000 --burst-sz 32 --buffer-sz 64

Call application for performance latency test of two Aesni MB PMD executed on two cores for cipher encryption aes-cbc, ten operations in silent mode:

dpdk-test-crypto-perf -l 4-7 --vdev crypto_aesni_mb1
--vdev crypto_aesni_mb2 -w 0000:00:00.0 -- --devtype crypto_aesni_mb
--cipher-algo aes-cbc --cipher-key-sz 16 --cipher-iv-sz 16
--cipher-op encrypt --optype cipher-only --silent
--ptest latency --total-ops 10

Call application for verification test of single open ssl PMD for cipher encryption aes-gcm and auth generation aes-gcm,ten operations in silent mode, test vector provide in file “test_aes_gcm.data” with packet verification:

dpdk-test-crypto-perf -l 4-7 --vdev crypto_openssl -w 0000:00:00.0 --
--devtype crypto_openssl --aead-algo aes-gcm --aead-key-sz 16
--aead-iv-sz 16 --aead-op encrypt --aead-aad-sz 16 --digest-sz 16
--optype aead --silent --ptest verify --total-ops 10
--test-file test_aes_gcm.data

Test vector file for cipher algorithm aes cbc 256 with authorization sha:

# Global Section
plaintext =
0xff, 0xca, 0xfb, 0xf1, 0x38, 0x20, 0x2f, 0x7b, 0x24, 0x98, 0x26, 0x7d, 0x1d, 0x9f, 0xb3, 0x93,
0xd9, 0xef, 0xbd, 0xad, 0x4e, 0x40, 0xbd, 0x60, 0xe9, 0x48, 0x59, 0x90, 0x67, 0xd7, 0x2b, 0x7b,
0x8a, 0xe0, 0x4d, 0xb0, 0x70, 0x38, 0xcc, 0x48, 0x61, 0x7d, 0xee, 0xd6, 0x35, 0x49, 0xae, 0xb4,
0xaf, 0x6b, 0xdd, 0xe6, 0x21, 0xc0, 0x60, 0xce, 0x0a, 0xf4, 0x1c, 0x2e, 0x1c, 0x8d, 0xe8, 0x7b
ciphertext =
0x77, 0xF9, 0xF7, 0x7A, 0xA3, 0xCB, 0x68, 0x1A, 0x11, 0x70, 0xD8, 0x7A, 0xB6, 0xE2, 0x37, 0x7E,
0xD1, 0x57, 0x1C, 0x8E, 0x85, 0xD8, 0x08, 0xBF, 0x57, 0x1F, 0x21, 0x6C, 0xAD, 0xAD, 0x47, 0x1E,
0x0D, 0x6B, 0x79, 0x39, 0x15, 0x4E, 0x5B, 0x59, 0x2D, 0x76, 0x87, 0xA6, 0xD6, 0x47, 0x8F, 0x82,
0xB8, 0x51, 0x91, 0x32, 0x60, 0xCB, 0x97, 0xDE, 0xBE, 0xF0, 0xAD, 0xFC, 0x23, 0x2E, 0x22, 0x02
cipher_key =
0xE4, 0x23, 0x33, 0x8A, 0x35, 0x64, 0x61, 0xE2, 0x49, 0x03, 0xDD, 0xC6, 0xB8, 0xCA, 0x55, 0x7A,
0xd0, 0xe7, 0x4b, 0xfb, 0x5d, 0xe5, 0x0c, 0xe7, 0x6f, 0x21, 0xb5, 0x52, 0x2a, 0xbb, 0xc7, 0xf7
auth_key =
0xaf, 0x96, 0x42, 0xf1, 0x8c, 0x50, 0xdc, 0x67, 0x1a, 0x43, 0x47, 0x62, 0xc7, 0x04, 0xab, 0x05,
0xf5, 0x0c, 0xe7, 0xa2, 0xa6, 0x23, 0xd5, 0x3d, 0x95, 0xd8, 0xcd, 0x86, 0x79, 0xf5, 0x01, 0x47,
0x4f, 0xf9, 0x1d, 0x9d, 0x36, 0xf7, 0x68, 0x1a, 0x64, 0x44, 0x58, 0x5d, 0xe5, 0x81, 0x15, 0x2a,
0x41, 0xe4, 0x0e, 0xaa, 0x1f, 0x04, 0x21, 0xff, 0x2c, 0xf3, 0x73, 0x2b, 0x48, 0x1e, 0xd2, 0xf7
cipher_iv =
0x00, 0x01, 0x02, 0x03, 0x04, 0x05, 0x06, 0x07, 0x08, 0x09, 0x0A, 0x0B, 0x0C, 0x0D, 0x0E, 0x0F
# Section sha 1 hmac buff 32
[sha1_hmac_buff_32]
digest =
0x36, 0xCA, 0x49, 0x6A, 0xE3, 0x54, 0xD8, 0x4F, 0x0B, 0x76, 0xD8, 0xAA, 0x78, 0xEB, 0x9D, 0x65,
0x2C, 0xCA, 0x1F, 0x97
# Section sha 256 hmac buff 32
[sha256_hmac_buff_32]
digest =
0x1C, 0xB2, 0x3D, 0xD1, 0xF9, 0xC7, 0x6C, 0x49, 0x2E, 0xDA, 0x94, 0x8B, 0xF1, 0xCF, 0x96, 0x43,
0x67, 0x50, 0x39, 0x76, 0xB5, 0xA1, 0xCE, 0xA1, 0xD7, 0x77, 0x10, 0x07, 0x43, 0x37, 0x05, 0xB4

dpdk-test-eventdev Application

The dpdk-test-eventdev tool is a Data Plane Development Kit (DPDK) application that allows exercising various eventdev use cases. This application has a generic framework to add new eventdev based test cases to verify functionality and measure the performance parameters of DPDK eventdev devices.

Compiling the Application

Build the application

Execute the dpdk-setup.sh script to build the DPDK library together with the dpdk-test-eventdev application.

Initially, the user must select a DPDK target to choose the correct target type and compiler options to use when building the libraries. The user must have all libraries, modules, updates and compilers installed in the system prior to this, as described in the earlier chapters in this Getting Started Guide.

Running the Application

The application has a number of command line options:

dpdk-test-eventdev [EAL Options] -- [application options]
EAL Options

The following are the EAL command-line options that can be used in conjunction with the dpdk-test-eventdev application. See the DPDK Getting Started Guides for more information on these options.

  • -c <COREMASK> or -l <CORELIST>

    Set the hexadecimal bitmask of the cores to run on. The corelist is a list of cores to use.

  • --vdev <driver><id>

    Add a virtual eventdev device.

Application Options

The following are the application command-line options:

  • --verbose

    Set verbose level. Default is 1. Value > 1 displays more details.

  • --dev <n>

    Set the device id of the event device.

  • --test <name>

    Set test name, where name is one of the following:

    order_queue
    order_atq
    perf_queue
    perf_atq
    
  • --socket_id <n>

    Set the socket id of the application resources.

  • --pool-sz <n>

    Set the number of mbufs to be allocated from the mempool.

  • --plcores <CORELIST>

    Set the list of cores to be used as producers.

  • --wlcores <CORELIST>

    Set the list of cores to be used as workers.

  • --stlist <type_list>

    Set the scheduled type of each stage where type_list size determines the number of stages used in the test application. Each type_list member can be one of the following:

    P or p : Parallel schedule type
    O or o : Ordered schedule type
    A or a : Atomic schedule type
    

    Application expects the type_list in comma separated form (i.e. --stlist o,a,a,a)

  • --nb_flows <n>

    Set the number of flows to produce.

  • --nb_pkts <n>

    Set the number of packets to produce. 0 implies no limit.

  • --worker_deq_depth <n>

    Set the dequeue depth of the worker.

  • --fwd_latency

    Perform forward latency measurement.

  • --queue_priority

    Enable queue priority.

Eventdev Tests

ORDER_QUEUE Test

This is a functional test case that aims at testing the following:

  1. Verify the ingress order maintenance.
  2. Verify the exclusive(atomic) access to given atomic flow per eventdev port.
Order queue test eventdev configuration.
# Items Value Comments
1 nb_queues 2 q0(ordered), q1(atomic)
2 nb_producers 1  
3 nb_workers >= 1  
4 nb_ports nb_workers + 1 Workers use port 0 to port n-1. Producer uses port n
_images/eventdev_order_queue_test.svg

order queue test operation.

The order queue test configures the eventdev with two queues and an event producer to inject the events to q0(ordered) queue. Both q0(ordered) and q1(atomic) are linked to all the workers.

The event producer maintains a sequence number per flow and injects the events to the ordered queue. The worker receives the events from ordered queue and forwards to atomic queue. Since the events from an ordered queue can be processed in parallel on the different workers, the ingress order of events might have changed on the downstream atomic queue enqueue. On enqueue to the atomic queue, the eventdev PMD driver reorders the event to the original ingress order(i.e producer ingress order).

When the event is dequeued from the atomic queue by the worker, this test verifies the expected sequence number of associated event per flow by comparing the free running expected sequence number per flow.

Application options

Supported application command line options are following:

--verbose
--dev
--test
--socket_id
--pool_sz
--plcores
--wlcores
--nb_flows
--nb_pkts
--worker_deq_depth
Example

Example command to run order queue test:

sudo build/app/dpdk-test-eventdev --vdev=event_sw0 -- \
             --test=order_queue --plcores 1 --wlcores 2,3
ORDER_ATQ Test

This test verifies the same aspects of order_queue test, the difference is the number of queues used, this test operates on a single all types queue(atq) instead of two different queues for ordered and atomic.

Order all types queue test eventdev configuration.
# Items Value Comments
1 nb_queues 1 q0(all types queue)
2 nb_producers 1  
3 nb_workers >= 1  
4 nb_ports nb_workers + 1 Workers use port 0 to port n-1.Producer uses port n.
_images/eventdev_order_atq_test.svg

order all types queue test operation.

Application options

Supported application command line options are following:

--verbose
--dev
--test
--socket_id
--pool_sz
--plcores
--wlcores
--nb_flows
--nb_pkts
--worker_deq_depth
Example

Example command to run order all types queue test:

sudo build/app/dpdk-test-eventdev --vdev=event_octeontx -- \
                     --test=order_atq --plcores 1 --wlcores 2,3
PERF_QUEUE Test

This is a performance test case that aims at testing the following:

  1. Measure the number of events can be processed in a second.
  2. Measure the latency to forward an event.
Perf queue test eventdev configuration.
# Items Value Comments
1 nb_queues nb_producers * nb_stages Queues will be configured based on the user requested sched type list(–stlist)
2 nb_producers >= 1 Selected through –plcores command line argument.
3 nb_workers >= 1 Selected through –wlcores command line argument
4 nb_ports nb_workers + nb_producers Workers use port 0 to port n-1. Producers use port n to port p
_images/eventdev_perf_queue_test.svg

perf queue test operation.

The perf queue test configures the eventdev with Q queues and P ports, where Q and P is a function of the number of workers, the number of producers and number of stages as mentioned in Table 6.3.

The user can choose the number of workers, the number of producers and number of stages through the --wlcores, --plcores and the --stlist application command line arguments respectively.

The producer(s) injects the events to eventdev based the first stage sched type list requested by the user through --stlist the command line argument.

Based on the number of stages to process(selected through --stlist), The application forwards the event to next upstream queue and terminates when it reaches the last stage in the pipeline. On event termination, application increments the number events processed and print periodically in one second to get the number of events processed in one second.

When --fwd_latency command line option selected, the application inserts the timestamp in the event on the first stage and then on termination, it updates the number of cycles to forward a packet. The application uses this value to compute the average latency to a forward packet.

Application options

Supported application command line options are following:

--verbose
--dev
--test
--socket_id
--pool_sz
--plcores
--wlcores
--stlist
--nb_flows
--nb_pkts
--worker_deq_depth
--fwd_latency
--queue_priority
Example

Example command to run perf queue test:

sudo build/app/dpdk-test-eventdev -c 0xf -s 0x1 --vdev=event_sw0 -- \
     --test=perf_queue --plcores=2 --wlcore=3 --stlist=p --nb_pkts=0
PERF_ATQ Test

This is a performance test case that aims at testing the following with all types queue eventdev scheme.

  1. Measure the number of events can be processed in a second.
  2. Measure the latency to forward an event.
Perf all types queue test eventdev configuration.
# Items Value Comments
1 nb_queues nb_producers Queues will be configured based on the user requested sched type list(–stlist)
2 nb_producers >= 1 Selected through –plcores command line argument.
3 nb_workers >= 1 Selected through –wlcores command line argument
4 nb_ports nb_workers + nb_producers Workers use port 0 to port n-1. Producers use port n to port p
_images/eventdev_perf_atq_test.svg

perf all types queue test operation.

The all types queues(atq) perf test configures the eventdev with Q queues and P ports, where Q and P is a function of the number of workers and number of producers as mentioned in Table 6.4.

The atq queue test functions as same as perf_queue test. The difference is, It uses, all type queue scheme instead of separate queues for each stage and thus reduces the number of queues required to realize the use case and enables flow pinning as the event does not move to the next queue.

Application options

Supported application command line options are following:

--verbose
--dev
--test
--socket_id
--pool_sz
--plcores
--wlcores
--stlist
--nb_flows
--nb_pkts
--worker_deq_depth
--fwd_latency
Example

Example command to run perf all types queue test:

sudo build/app/dpdk-test-eventdev --vdev=event_octeontx -- \
             --test=perf_atq --plcores=2 --wlcore=3 --stlist=p --nb_pkts=0

Testpmd Application User Guide

Introduction

This document is a user guide for the testpmd example application that is shipped as part of the Data Plane Development Kit.

The testpmd application can be used to test the DPDK in a packet forwarding mode and also to access NIC hardware features such as Flow Director. It also serves as a example of how to build a more fully-featured application using the DPDK SDK.

The guide shows how to build and run the testpmd application and how to configure the application from the command line and the run-time environment.

Compiling the Application

The testpmd application is compiled as part of the main compilation of the DPDK libraries and tools. Refer to the DPDK Getting Started Guides for details. The basic compilation steps are:

  1. Set the required environmental variables and go to the source directory:

    export RTE_SDK=/path/to/rte_sdk
    cd $RTE_SDK
    
  2. Set the compilation target. For example:

    export RTE_TARGET=x86_64-native-linuxapp-gcc
    
  3. Build the application:

    make install T=$RTE_TARGET
    

    The compiled application will be located at:

    $RTE_SDK/$RTE_TARGET/app/testpmd
    

Running the Application

EAL Command-line Options

The following are the EAL command-line options that can be used in conjunction with the testpmd, or any other DPDK application. See the DPDK Getting Started Guides for more information on these options.

  • -c COREMASK

    Set the hexadecimal bitmask of the cores to run on.

  • -l CORELIST

    List of cores to run on

    The argument format is <c1>[-c2][,c3[-c4],...] where c1, c2, etc are core indexes between 0 and 128.

  • --lcores COREMAP

    Map lcore set to physical cpu set

    The argument format is:

    <lcores[@cpus]>[<,lcores[@cpus]>...]
    

    Lcore and CPU lists are grouped by ( and ) Within the group. The - character is used as a range separator and , is used as a single number separator. The grouping () can be omitted for single element group. The @ can be omitted if cpus and lcores have the same value.

  • --master-lcore ID

    Core ID that is used as master.

  • -n NUM

    Set the number of memory channels to use.

  • -b, --pci-blacklist domain:bus:devid.func

    Blacklist a PCI device to prevent EAL from using it. Multiple -b options are allowed.

  • -d LIB.so

    Load an external driver. Multiple -d options are allowed.

  • -w, --pci-whitelist domain:bus:devid:func

    Add a PCI device in white list.

  • -m MB

    Memory to allocate. See also --socket-mem.

  • -r NUM

    Set the number of memory ranks (auto-detected by default).

  • -v

    Display the version information on startup.

  • --syslog

    Set the syslog facility.

  • --socket-mem

    Set the memory to allocate on specific sockets (use comma separated values).

  • --huge-dir

    Specify the directory where the hugetlbfs is mounted.

  • mbuf-pool-ops-name:

    Pool ops name for mbuf to use.

  • --proc-type

    Set the type of the current process.

  • --file-prefix

    Prefix for hugepage filenames.

  • -vmware-tsc-map

    Use VMware TSC map instead of native RDTSC.

  • --vdev

    Add a virtual device using the format:

    <driver><id>[,key=val, ...]
    

    For example:

    --vdev 'net_pcap0,rx_pcap=input.pcap,tx_pcap=output.pcap'
    
  • --base-virtaddr

    Specify base virtual address.

  • --create-uio-dev

    Create /dev/uioX (usually done by hotplug).

  • --no-shconf

    No shared config (mmap-ed files).

  • --no-pci

    Disable pci.

  • --no-hpet

    Disable hpet.

  • --no-huge

    Use malloc instead of hugetlbfs.

Testpmd Command-line Options

The following are the command-line options for the testpmd applications. They must be separated from the EAL options, shown in the previous section, with a -- separator:

sudo ./testpmd -l 0-3 -n 4 -- -i --portmask=0x1 --nb-cores=2

The commandline options are:

  • -i, --interactive

    Run testpmd in interactive mode. In this mode, the testpmd starts with a prompt that can be used to start and stop forwarding, configure the application and display stats on the current packet processing session. See Testpmd Runtime Functions for more details.

    In non-interactive mode, the application starts with the configuration specified on the command-line and immediately enters forwarding mode.

  • -h, --help

    Display a help message and quit.

  • -a, --auto-start

    Start forwarding on initialization.

  • --tx-first

    Start forwarding, after sending a burst of packets first.

Note

This flag should be only used in non-interactive mode.

  • --stats-period PERIOD

    Display statistics every PERIOD seconds, if interactive mode is disabled. The default value is 0, which means that the statistics will not be displayed.

  • --nb-cores=N

    Set the number of forwarding cores, where 1 <= N <= “number of cores” or CONFIG_RTE_MAX_LCORE from the configuration file. The default value is 1.

  • --nb-ports=N

    Set the number of forwarding ports, where 1 <= N <= “number of ports” on the board or CONFIG_RTE_MAX_ETHPORTS from the configuration file. The default value is the number of ports on the board.

  • --coremask=0xXX

    Set the hexadecimal bitmask of the cores running the packet forwarding test. The master lcore is reserved for command line parsing only and cannot be masked on for packet forwarding.

  • --portmask=0xXX

    Set the hexadecimal bitmask of the ports used by the packet forwarding test.

  • --numa

    Enable NUMA-aware allocation of RX/TX rings and of RX memory buffers (mbufs). [Default setting]

  • --no-numa

    Disable NUMA-aware allocation of RX/TX rings and of RX memory buffers (mbufs).

  • --port-numa-config=(port,socket)[,(port,socket)]

    Specify the socket on which the memory pool to be used by the port will be allocated.

  • --ring-numa-config=(port,flag,socket)[,(port,flag,socket)]

    Specify the socket on which the TX/RX rings for the port will be allocated. Where flag is 1 for RX, 2 for TX, and 3 for RX and TX.

  • --socket-num=N

    Set the socket from which all memory is allocated in NUMA mode, where 0 <= N < number of sockets on the board.

  • --mbuf-size=N

    Set the data size of the mbufs used to N bytes, where N < 65536. The default value is 2048.

  • --total-num-mbufs=N

    Set the number of mbufs to be allocated in the mbuf pools, where N > 1024.

  • --max-pkt-len=N

    Set the maximum packet size to N bytes, where N >= 64. The default value is 1518.

  • --eth-peers-configfile=name

    Use a configuration file containing the Ethernet addresses of the peer ports. The configuration file should contain the Ethernet addresses on separate lines:

    XX:XX:XX:XX:XX:01
    XX:XX:XX:XX:XX:02
    ...
    
  • --eth-peer=N,XX:XX:XX:XX:XX:XX

    Set the MAC address XX:XX:XX:XX:XX:XX of the peer port N, where 0 <= N < CONFIG_RTE_MAX_ETHPORTS from the configuration file.

  • --pkt-filter-mode=mode

    Set Flow Director mode where mode is either none (the default), signature or perfect. See flow_director_filter for more details.

  • --pkt-filter-report-hash=mode

    Set Flow Director hash match reporting mode where mode is none, match (the default) or always.

  • --pkt-filter-size=N

    Set Flow Director allocated memory size, where N is 64K, 128K or 256K. Sizes are in kilobytes. The default is 64.

  • --pkt-filter-flexbytes-offset=N

    Set the flexbytes offset. The offset is defined in words (not bytes) counted from the first byte of the destination Ethernet MAC address, where N is 0 <= N <= 32. The default value is 0x6.

  • --pkt-filter-drop-queue=N

    Set the drop-queue. In perfect filter mode, when a rule is added with queue = -1, the packet will be enqueued into the RX drop-queue. If the drop-queue does not exist, the packet is dropped. The default value is N=127.

  • --disable-crc-strip

    Disable hardware CRC stripping.

  • --enable-lro

    Enable large receive offload.

  • --enable-rx-cksum

    Enable hardware RX checksum offload.

  • --enable-scatter

    Enable scatter (multi-segment) RX.

  • --disable-hw-vlan

    Disable hardware VLAN.

  • --disable-hw-vlan-filter

    Disable hardware VLAN filter.

  • --disable-hw-vlan-strip

    Disable hardware VLAN strip.

  • --disable-hw-vlan-extend

    Disable hardware VLAN extend.

  • --enable-drop-en

    Enable per-queue packet drop for packets with no descriptors.

  • --disable-rss

    Disable RSS (Receive Side Scaling).

  • --port-topology=mode

    Set port topology, where mode is paired (the default) or chained.

    In paired mode, the forwarding is between pairs of ports, for example: (0,1), (2,3), (4,5).

    In chained mode, the forwarding is to the next available port in the port mask, for example: (0,1), (1,2), (2,0).

    The ordering of the ports can be changed using the portlist testpmd runtime function.

  • --forward-mode=mode

    Set the forwarding mode where mode is one of the following:

    io (the default)
    mac
    mac_swap
    flowgen
    rxonly
    txonly
    csum
    icmpecho
    ieee1588
    tm
    
  • --rss-ip

    Set RSS functions for IPv4/IPv6 only.

  • --rss-udp

    Set RSS functions for IPv4/IPv6 and UDP.

  • --rxq=N

    Set the number of RX queues per port to N, where 1 <= N <= 65535. The default value is 1.

  • --rxd=N

    Set the number of descriptors in the RX rings to N, where N > 0. The default value is 128.

  • --txq=N

    Set the number of TX queues per port to N, where 1 <= N <= 65535. The default value is 1.

  • --txd=N

    Set the number of descriptors in the TX rings to N, where N > 0. The default value is 512.

  • --burst=N

    Set the number of packets per burst to N, where 1 <= N <= 512. The default value is 16.

  • --mbcache=N

    Set the cache of mbuf memory pools to N, where 0 <= N <= 512. The default value is 16.

  • --rxpt=N

    Set the prefetch threshold register of RX rings to N, where N >= 0. The default value is 8.

  • --rxht=N

    Set the host threshold register of RX rings to N, where N >= 0. The default value is 8.

  • --rxfreet=N

    Set the free threshold of RX descriptors to N, where 0 <= N < value of –rxd. The default value is 0.

  • --rxwt=N

    Set the write-back threshold register of RX rings to N, where N >= 0. The default value is 4.

  • --txpt=N

    Set the prefetch threshold register of TX rings to N, where N >= 0. The default value is 36.

  • --txht=N

    Set the host threshold register of TX rings to N, where N >= 0. The default value is 0.

  • --txwt=N

    Set the write-back threshold register of TX rings to N, where N >= 0. The default value is 0.

  • --txfreet=N

    Set the transmit free threshold of TX rings to N, where 0 <= N <= value of --txd. The default value is 0.

  • --txrst=N

    Set the transmit RS bit threshold of TX rings to N, where 0 <= N <= value of --txd. The default value is 0.

  • --txqflags=0xXXXXXXXX

    Set the hexadecimal bitmask of TX queue flags, where 0 <= N <= 0x7FFFFFFF. The default value is 0.

    Note

    When using hardware offload functions such as vlan or checksum add txqflags=0 to force the full-featured TX code path. In some PMDs this may already be the default.

  • --rx-queue-stats-mapping=(port,queue,mapping)[,(port,queue,mapping)]

    Set the RX queues statistics counters mapping 0 <= mapping <= 15.

  • --tx-queue-stats-mapping=(port,queue,mapping)[,(port,queue,mapping)]

    Set the TX queues statistics counters mapping 0 <= mapping <= 15.

  • --no-flush-rx

    Don’t flush the RX streams before starting forwarding. Used mainly with the PCAP PMD.

  • --txpkts=X[,Y]

    Set TX segment sizes or total packet length. Valid for tx-only and flowgen forwarding modes.

  • --disable-link-check

    Disable check on link status when starting/stopping ports.

  • --no-lsc-interrupt

    Disable LSC interrupts for all ports, even those supporting it.

  • --no-rmv-interrupt

    Disable RMV interrupts for all ports, even those supporting it.

  • --bitrate-stats=N

    Set the logical core N to perform bitrate calculation.

  • --print-event <unknown|intr_lsc|queue_state|intr_reset|vf_mbox|macsec|intr_rmv|all>

    Enable printing the occurrence of the designated event. Using all will enable all of them.

  • --mask-event <unknown|intr_lsc|queue_state|intr_reset|vf_mbox|macsec|intr_rmv|all>

    Disable printing the occurrence of the designated event. Using all will disable all of them.

  • --flow-isolate-all

    Providing this parameter requests flow API isolated mode on all ports at initialization time. It ensures all traffic is received through the configured flow rules only (see flow command).

    Ports that do not support this mode are automatically discarded.

Testpmd Runtime Functions

Where the testpmd application is started in interactive mode, (-i|--interactive), it displays a prompt that can be used to start and stop forwarding, configure the application, display statistics (including the extended NIC statistics aka xstats) , set the Flow Director and other tasks:

testpmd>

The testpmd prompt has some, limited, readline support. Common bash command-line functions such as Ctrl+a and Ctrl+e to go to the start and end of the prompt line are supported as well as access to the command history via the up-arrow.

There is also support for tab completion. If you type a partial command and hit <TAB> you get a list of the available completions:

testpmd> show port <TAB>

    info [Mul-choice STRING]: show|clear port info|stats|xstats|fdir|stat_qmap|dcb_tc|cap X
    info [Mul-choice STRING]: show|clear port info|stats|xstats|fdir|stat_qmap|dcb_tc|cap all
    stats [Mul-choice STRING]: show|clear port info|stats|xstats|fdir|stat_qmap|dcb_tc|cap X
    stats [Mul-choice STRING]: show|clear port info|stats|xstats|fdir|stat_qmap|dcb_tc|cap all
    ...

Note

Some examples in this document are too long to fit on one line are are shown wrapped at “\” for display purposes:

testpmd> set flow_ctrl rx (on|off) tx (on|off) (high_water) (low_water) \
         (pause_time) (send_xon) (port_id)

In the real testpmd> prompt these commands should be on a single line.

Help Functions

The testpmd has on-line help for the functions that are available at runtime. These are divided into sections and can be accessed using help, help section or help all:

testpmd> help

    help control    : Start and stop forwarding.
    help display    : Displaying port, stats and config information.
    help config     : Configuration information.
    help ports      : Configuring ports.
    help registers  : Reading and setting port registers.
    help filters    : Filters configuration help.
    help all        : All of the above sections.

Command File Functions

To facilitate loading large number of commands or to avoid cutting and pasting where not practical or possible testpmd supports alternative methods for executing commands.

  • If started with the --cmdline-file=FILENAME command line argument testpmd will execute all CLI commands contained within the file immediately before starting packet forwarding or entering interactive mode.
./testpmd -n4 -r2 ... -- -i --cmdline-file=/home/ubuntu/flow-create-commands.txt
Interactive-mode selected
CLI commands to be read from /home/ubuntu/flow-create-commands.txt
Configuring Port 0 (socket 0)
Port 0: 7C:FE:90:CB:74:CE
Configuring Port 1 (socket 0)
Port 1: 7C:FE:90:CB:74:CA
Checking link statuses...
Port 0 Link Up - speed 10000 Mbps - full-duplex
Port 1 Link Up - speed 10000 Mbps - full-duplex
Done
Flow rule #0 created
Flow rule #1 created
...
...
Flow rule #498 created
Flow rule #499 created
Read all CLI commands from /home/ubuntu/flow-create-commands.txt
testpmd>
  • At run-time additional commands can be loaded in bulk by invoking the load FILENAME command.
testpmd> load /home/ubuntu/flow-create-commands.txt
Flow rule #0 created
Flow rule #1 created
...
...
Flow rule #498 created
Flow rule #499 created
Read all CLI commands from /home/ubuntu/flow-create-commands.txt
testpmd>

In all cases output from any included command will be displayed as standard output. Execution will continue until the end of the file is reached regardless of whether any errors occur. The end user must examine the output to determine if any failures occurred.

Control Functions

start

Start packet forwarding with current configuration:

testpmd> start
start tx_first

Start packet forwarding with current configuration after sending specified number of bursts of packets:

testpmd> start tx_first (""|burst_num)

The default burst number is 1 when burst_num not presented.

stop

Stop packet forwarding, and display accumulated statistics:

testpmd> stop
quit

Quit to prompt:

testpmd> quit

Display Functions

The functions in the following sections are used to display information about the testpmd configuration or the NIC status.

show port

Display information for a given port or all ports:

testpmd> show port (info|stats|xstats|fdir|stat_qmap|dcb_tc|cap) (port_id|all)

The available information categories are:

  • info: General port information such as MAC address.
  • stats: RX/TX statistics.
  • xstats: RX/TX extended NIC statistics.
  • fdir: Flow Director information and statistics.
  • stat_qmap: Queue statistics mapping.
  • dcb_tc: DCB information such as TC mapping.
  • cap: Supported offload capabilities.

For example:

testpmd> show port info 0

********************* Infos for port 0 *********************

MAC address: XX:XX:XX:XX:XX:XX
Connect to socket: 0
memory allocation on the socket: 0
Link status: up
Link speed: 40000 Mbps
Link duplex: full-duplex
Promiscuous mode: enabled
Allmulticast mode: disabled
Maximum number of MAC addresses: 64
Maximum number of MAC addresses of hash filtering: 0
VLAN offload:
    strip on
    filter on
    qinq(extend) off
Redirection table size: 512
Supported flow types:
  ipv4-frag
  ipv4-tcp
  ipv4-udp
  ipv4-sctp
  ipv4-other
  ipv6-frag
  ipv6-tcp
  ipv6-udp
  ipv6-sctp
  ipv6-other
  l2_payload
  port
  vxlan
  geneve
  nvgre
show port rss reta

Display the rss redirection table entry indicated by masks on port X:

testpmd> show port (port_id) rss reta (size) (mask0, mask1...)

size is used to indicate the hardware supported reta size

show port rss-hash

Display the RSS hash functions and RSS hash key of a port:

testpmd> show port (port_id) rss-hash ipv4|ipv4-frag|ipv4-tcp|ipv4-udp|ipv4-sctp|ipv4-other|ipv6|ipv6-frag|ipv6-tcp|ipv6-udp|ipv6-sctp|ipv6-other|l2-payload|ipv6-ex|ipv6-tcp-ex|ipv6-udp-ex [key]
clear port

Clear the port statistics for a given port or for all ports:

testpmd> clear port (info|stats|xstats|fdir|stat_qmap) (port_id|all)

For example:

testpmd> clear port stats all
show (rxq|txq)

Display information for a given port’s RX/TX queue:

testpmd> show (rxq|txq) info (port_id) (queue_id)
show config

Displays the configuration of the application. The configuration comes from the command-line, the runtime or the application defaults:

testpmd> show config (rxtx|cores|fwd|txpkts)

The available information categories are:

  • rxtx: RX/TX configuration items.
  • cores: List of forwarding cores.
  • fwd: Packet forwarding configuration.
  • txpkts: Packets to TX configuration.

For example:

testpmd> show config rxtx

io packet forwarding - CRC stripping disabled - packets/burst=16
nb forwarding cores=2 - nb forwarding ports=1
RX queues=1 - RX desc=128 - RX free threshold=0
RX threshold registers: pthresh=8 hthresh=8 wthresh=4
TX queues=1 - TX desc=512 - TX free threshold=0
TX threshold registers: pthresh=36 hthresh=0 wthresh=0
TX RS bit threshold=0 - TXQ flags=0x0
set fwd

Set the packet forwarding mode:

testpmd> set fwd (io|mac|macswap|flowgen| \
                  rxonly|txonly|csum|icmpecho) (""|retry)

retry can be specified for forwarding engines except rx_only.

The available information categories are:

  • io: Forwards packets “as-is” in I/O mode. This is the fastest possible forwarding operation as it does not access packets data. This is the default mode.
  • mac: Changes the source and the destination Ethernet addresses of packets before forwarding them. Default application behaviour is to set source Ethernet address to that of the transmitting interface, and destination address to a dummy value (set during init). The user may specify a target destination Ethernet address via the ‘eth-peer’ or ‘eth-peer-configfile’ command-line options. It is not currently possible to specify a specific source Ethernet address.
  • macswap: MAC swap forwarding mode. Swaps the source and the destination Ethernet addresses of packets before forwarding them.
  • flowgen: Multi-flow generation mode. Originates a number of flows (with varying destination IP addresses), and terminate receive traffic.
  • rxonly: Receives packets but doesn’t transmit them.
  • txonly: Generates and transmits packets without receiving any.
  • csum: Changes the checksum field with hardware or software methods depending on the offload flags on the packet.
  • icmpecho: Receives a burst of packets, lookup for IMCP echo requests and, if any, send back ICMP echo replies.
  • ieee1588: Demonstrate L2 IEEE1588 V2 PTP timestamping for RX and TX. Requires CONFIG_RTE_LIBRTE_IEEE1588=y.
  • tm: Traffic Management forwarding mode Demonstrates the use of ethdev traffic management APIs and softnic PMD for QoS traffic management. In this mode, 5-level hierarchical QoS scheduler is available as an default option that can be enabled through CLI. The user can also modify the default hierarchy or specify the new hierarchy through CLI for implementing QoS scheduler. Requires CONFIG_RTE_LIBRTE_PMD_SOFTNIC=y CONFIG_RTE_LIBRTE_SCHED=y.

Note: TX timestamping is only available in the “Full Featured” TX path. To force testpmd into this mode set --txqflags=0.

Example:

testpmd> set fwd rxonly

Set rxonly packet forwarding mode
read rxd

Display an RX descriptor for a port RX queue:

testpmd> read rxd (port_id) (queue_id) (rxd_id)

For example:

testpmd> read rxd 0 0 4
     0x0000000B - 0x001D0180 / 0x0000000B - 0x001D0180
read txd

Display a TX descriptor for a port TX queue:

testpmd> read txd (port_id) (queue_id) (txd_id)

For example:

testpmd> read txd 0 0 4
     0x00000001 - 0x24C3C440 / 0x000F0000 - 0x2330003C
ddp get list

Get loaded dynamic device personalization (DDP) package info list:

testpmd> ddp get list (port_id)
ddp get info

Display information about dynamic device personalization (DDP) profile:

testpmd> ddp get info (profile_path)
show vf stats

Display VF statistics:

testpmd> show vf stats (port_id) (vf_id)
clear vf stats

Reset VF statistics:

testpmd> clear vf stats (port_id) (vf_id)
show port pctype mapping

List all items from the pctype mapping table:

testpmd> show port (port_id) pctype mapping

Configuration Functions

The testpmd application can be configured from the runtime as well as from the command-line.

This section details the available configuration functions that are available.

Note

Configuration changes only become active when forwarding is started/restarted.

set default

Reset forwarding to the default configuration:

testpmd> set default
set verbose

Set the debug verbosity level:

testpmd> set verbose (level)

Currently the only available levels are 0 (silent except for error) and 1 (fully verbose).

set nbport

Set the number of ports used by the application:

set nbport (num)

This is equivalent to the --nb-ports command-line option.

set nbcore

Set the number of cores used by the application:

testpmd> set nbcore (num)

This is equivalent to the --nb-cores command-line option.

Note

The number of cores used must not be greater than number of ports used multiplied by the number of queues per port.

set coremask

Set the forwarding cores hexadecimal mask:

testpmd> set coremask (mask)

This is equivalent to the --coremask command-line option.

Note

The master lcore is reserved for command line parsing only and cannot be masked on for packet forwarding.

set portmask

Set the forwarding ports hexadecimal mask:

testpmd> set portmask (mask)

This is equivalent to the --portmask command-line option.

set burst

Set number of packets per burst:

testpmd> set burst (num)

This is equivalent to the --burst command-line option.

When retry is enabled, the transmit delay time and number of retries can also be set:

testpmd> set burst tx delay (microseconds) retry (num)
set txpkts

Set the length of each segment of the TX-ONLY packets or length of packet for FLOWGEN mode:

testpmd> set txpkts (x[,y]*)

Where x[,y]* represents a CSV list of values, without white space.

set txsplit

Set the split policy for the TX packets, applicable for TX-ONLY and CSUM forwarding modes:

testpmd> set txsplit (off|on|rand)

Where:

  • off disable packet copy & split for CSUM mode.
  • on split outgoing packet into multiple segments. Size of each segment and number of segments per packet is determined by set txpkts command (see above).
  • rand same as ‘on’, but number of segments per each packet is a random value between 1 and total number of segments.
set corelist

Set the list of forwarding cores:

testpmd> set corelist (x[,y]*)

For example, to change the forwarding cores:

testpmd> set corelist 3,1
testpmd> show config fwd

io packet forwarding - ports=2 - cores=2 - streams=2 - NUMA support disabled
Logical Core 3 (socket 0) forwards packets on 1 streams:
RX P=0/Q=0 (socket 0) -> TX P=1/Q=0 (socket 0) peer=02:00:00:00:00:01
Logical Core 1 (socket 0) forwards packets on 1 streams:
RX P=1/Q=0 (socket 0) -> TX P=0/Q=0 (socket 0) peer=02:00:00:00:00:00

Note

The cores are used in the same order as specified on the command line.

set portlist

Set the list of forwarding ports:

testpmd> set portlist (x[,y]*)

For example, to change the port forwarding:

testpmd> set portlist 0,2,1,3
testpmd> show config fwd

io packet forwarding - ports=4 - cores=1 - streams=4
Logical Core 3 (socket 0) forwards packets on 4 streams:
RX P=0/Q=0 (socket 0) -> TX P=2/Q=0 (socket 0) peer=02:00:00:00:00:01
RX P=2/Q=0 (socket 0) -> TX P=0/Q=0 (socket 0) peer=02:00:00:00:00:00
RX P=1/Q=0 (socket 0) -> TX P=3/Q=0 (socket 0) peer=02:00:00:00:00:03
RX P=3/Q=0 (socket 0) -> TX P=1/Q=0 (socket 0) peer=02:00:00:00:00:02
set tx loopback

Enable/disable tx loopback:

testpmd> set tx loopback (port_id) (on|off)
set drop enable

set drop enable bit for all queues:

testpmd> set all queues drop (port_id) (on|off)
set split drop enable (for VF)

set split drop enable bit for VF from PF:

testpmd> set vf split drop (port_id) (vf_id) (on|off)
set mac antispoof (for VF)

Set mac antispoof for a VF from the PF:

testpmd> set vf mac antispoof  (port_id) (vf_id) (on|off)
set macsec offload

Enable/disable MACsec offload:

testpmd> set macsec offload (port_id) on encrypt (on|off) replay-protect (on|off)
testpmd> set macsec offload (port_id) off
set macsec sc

Configure MACsec secure connection (SC):

testpmd> set macsec sc (tx|rx) (port_id) (mac) (pi)

Note

The pi argument is ignored for tx. Check the NIC Datasheet for hardware limits.

set macsec sa

Configure MACsec secure association (SA):

testpmd> set macsec sa (tx|rx) (port_id) (idx) (an) (pn) (key)

Note

The IDX value must be 0 or 1. Check the NIC Datasheet for hardware limits.

set broadcast mode (for VF)

Set broadcast mode for a VF from the PF:

testpmd> set vf broadcast (port_id) (vf_id) (on|off)
vlan set strip

Set the VLAN strip on a port:

testpmd> vlan set strip (on|off) (port_id)
vlan set stripq

Set the VLAN strip for a queue on a port:

testpmd> vlan set stripq (on|off) (port_id,queue_id)
vlan set stripq (for VF)

Set VLAN strip for all queues in a pool for a VF from the PF:

testpmd> set vf vlan stripq (port_id) (vf_id) (on|off)
vlan set insert (for VF)

Set VLAN insert for a VF from the PF:

testpmd> set vf vlan insert (port_id) (vf_id) (vlan_id)
vlan set tag (for VF)

Set VLAN tag for a VF from the PF:

testpmd> set vf vlan tag (port_id) (vf_id) (on|off)
vlan set antispoof (for VF)

Set VLAN antispoof for a VF from the PF:

testpmd> set vf vlan antispoof (port_id) (vf_id) (on|off)
vlan set filter

Set the VLAN filter on a port:

testpmd> vlan set filter (on|off) (port_id)
vlan set qinq

Set the VLAN QinQ (extended queue in queue) on for a port:

testpmd> vlan set qinq (on|off) (port_id)
vlan set tpid

Set the inner or outer VLAN TPID for packet filtering on a port:

testpmd> vlan set (inner|outer) tpid (value) (port_id)

Note

TPID value must be a 16-bit number (value <= 65536).

rx_vlan add

Add a VLAN ID, or all identifiers, to the set of VLAN identifiers filtered by port ID:

testpmd> rx_vlan add (vlan_id|all) (port_id)

Note

VLAN filter must be set on that port. VLAN ID < 4096. Depending on the NIC used, number of vlan_ids may be limited to the maximum entries in VFTA table. This is important if enabling all vlan_ids.

rx_vlan rm

Remove a VLAN ID, or all identifiers, from the set of VLAN identifiers filtered by port ID:

testpmd> rx_vlan rm (vlan_id|all) (port_id)
rx_vlan add (for VF)

Add a VLAN ID, to the set of VLAN identifiers filtered for VF(s) for port ID:

testpmd> rx_vlan add (vlan_id) port (port_id) vf (vf_mask)
rx_vlan rm (for VF)

Remove a VLAN ID, from the set of VLAN identifiers filtered for VF(s) for port ID:

testpmd> rx_vlan rm (vlan_id) port (port_id) vf (vf_mask)
tunnel_filter add

Add a tunnel filter on a port:

testpmd> tunnel_filter add (port_id) (outer_mac) (inner_mac) (ip_addr) \
         (inner_vlan) (vxlan|nvgre|ipingre) (imac-ivlan|imac-ivlan-tenid|\
         imac-tenid|imac|omac-imac-tenid|oip|iip) (tenant_id) (queue_id)

The available information categories are:

  • vxlan: Set tunnel type as VXLAN.
  • nvgre: Set tunnel type as NVGRE.
  • ipingre: Set tunnel type as IP-in-GRE.
  • imac-ivlan: Set filter type as Inner MAC and VLAN.
  • imac-ivlan-tenid: Set filter type as Inner MAC, VLAN and tenant ID.
  • imac-tenid: Set filter type as Inner MAC and tenant ID.
  • imac: Set filter type as Inner MAC.
  • omac-imac-tenid: Set filter type as Outer MAC, Inner MAC and tenant ID.
  • oip: Set filter type as Outer IP.
  • iip: Set filter type as Inner IP.

Example:

testpmd> tunnel_filter add 0 68:05:CA:28:09:82 00:00:00:00:00:00 \
         192.168.2.2 0 ipingre oip 1 1

Set an IP-in-GRE tunnel on port 0, and the filter type is Outer IP.
tunnel_filter remove

Remove a tunnel filter on a port:

testpmd> tunnel_filter rm (port_id) (outer_mac) (inner_mac) (ip_addr) \
         (inner_vlan) (vxlan|nvgre|ipingre) (imac-ivlan|imac-ivlan-tenid|\
         imac-tenid|imac|omac-imac-tenid|oip|iip) (tenant_id) (queue_id)
rx_vxlan_port add

Add an UDP port for VXLAN packet filter on a port:

testpmd> rx_vxlan_port add (udp_port) (port_id)
rx_vxlan_port remove

Remove an UDP port for VXLAN packet filter on a port:

testpmd> rx_vxlan_port rm (udp_port) (port_id)
tx_vlan set

Set hardware insertion of VLAN IDs in packets sent on a port:

testpmd> tx_vlan set (port_id) vlan_id[, vlan_id_outer]

For example, set a single VLAN ID (5) insertion on port 0:

tx_vlan set 0 5

Or, set double VLAN ID (inner: 2, outer: 3) insertion on port 1:

tx_vlan set 1 2 3
tx_vlan set pvid

Set port based hardware insertion of VLAN ID in packets sent on a port:

testpmd> tx_vlan set pvid (port_id) (vlan_id) (on|off)
tx_vlan reset

Disable hardware insertion of a VLAN header in packets sent on a port:

testpmd> tx_vlan reset (port_id)
csum set

Select hardware or software calculation of the checksum when transmitting a packet using the csum forwarding engine:

testpmd> csum set (ip|udp|tcp|sctp|outer-ip) (hw|sw) (port_id)

Where:

  • ip|udp|tcp|sctp always relate to the inner layer.
  • outer-ip relates to the outer IP layer (only for IPv4) in the case where the packet is recognized as a tunnel packet by the forwarding engine (vxlan, gre and ipip are supported). See also the csum parse-tunnel command.

Note

Check the NIC Datasheet for hardware limits.

RSS queue region

Set RSS queue region span on a port:

testpmd> set port (port_id) queue-region region_id (value) \
             queue_start_index (value) queue_num (value)

Set flowtype mapping on a RSS queue region on a port:

testpmd> set port (port_id) queue-region region_id (value) flowtype (value)

where:

  • For the flowtype(pctype) of packet,the specific index for each type has been defined in file i40e_type.h as enum i40e_filter_pctype.

Set user priority mapping on a RSS queue region on a port:

testpmd> set port (port_id) queue-region UP (value) region_id (value)

Flush all queue region related configuration on a port:

testpmd> set port (port_id) queue-region flush (on|off)

where:

  • “on”is just an enable function which server for other configuration, it is for all configuration about queue region from up layer, at first will only keep in DPDK softwarestored in driver, only after “flush on”, it commit all configuration to HW. “off” is just clean all configuration about queue region just now, and restore all to DPDK i40e driver default config when start up.

Show all queue region related configuration info on a port:

testpmd> show port (port_id) queue-region

Note

Queue region only support on PF by now, so these command is only for configuration of queue region on PF port.

csum parse-tunnel

Define how tunneled packets should be handled by the csum forward engine:

testpmd> csum parse-tunnel (on|off) (tx_port_id)

If enabled, the csum forward engine will try to recognize supported tunnel headers (vxlan, gre, ipip).

If disabled, treat tunnel packets as non-tunneled packets (a inner header is handled as a packet payload).

Note

The port argument is the TX port like in the csum set command.

Example:

Consider a packet in packet like the following:

eth_out/ipv4_out/udp_out/vxlan/eth_in/ipv4_in/tcp_in
  • If parse-tunnel is enabled, the ip|udp|tcp|sctp parameters of csum set command relate to the inner headers (here ipv4_in and tcp_in), and the outer-ip parameter relates to the outer headers (here ipv4_out).
  • If parse-tunnel is disabled, the ip|udp|tcp|sctp parameters of csum  set
    command relate to the outer headers, here ipv4_out and udp_out.
csum show

Display tx checksum offload configuration:

testpmd> csum show (port_id)
tso set

Enable TCP Segmentation Offload (TSO) in the csum forwarding engine:

testpmd> tso set (segsize) (port_id)

Note

Check the NIC datasheet for hardware limits.

tso show

Display the status of TCP Segmentation Offload:

testpmd> tso show (port_id)
set port - gro

Enable or disable GRO in csum forwarding engine:

testpmd> set port <port_id> gro on|off

If enabled, the csum forwarding engine will perform GRO on the TCP/IPv4 packets received from the given port.

If disabled, packets received from the given port won’t be performed GRO. By default, GRO is disabled for all ports.

Note

When enable GRO for a port, TCP/IPv4 packets received from the port will be performed GRO. After GRO, all merged packets have bad checksums, since the GRO library doesn’t re-calculate checksums for the merged packets. Therefore, if users want the merged packets to have correct checksums, please select HW IP checksum calculation and HW TCP checksum calculation for the port which the merged packets are transmitted to.

show port - gro

Display GRO configuration for a given port:

testpmd> show port <port_id> gro
set gro flush

Set the cycle to flush the GROed packets from reassembly tables:

testpmd> set gro flush <cycles>

When enable GRO, the csum forwarding engine performs GRO on received packets, and the GROed packets are stored in reassembly tables. Users can use this command to determine when the GROed packets are flushed from the reassembly tables.

The cycles is measured in GRO operation times. The csum forwarding engine flushes the GROed packets from the tables every cycles GRO operations.

By default, the value of cycles is 1, which means flush GROed packets from the reassembly tables as soon as one GRO operation finishes. The value of cycles should be in the range of 1 to GRO_MAX_FLUSH_CYCLES.

Please note that the large value of cycles may cause the poor TCP/IP stack performance. Because the GROed packets are delayed to arrive the stack, thus causing more duplicated ACKs and TCP retransmissions.

set port - gso

Toggle per-port GSO support in csum forwarding engine:

testpmd> set port <port_id> gso on|off

If enabled, the csum forwarding engine will perform GSO on supported IPv4 packets, transmitted on the given port.

If disabled, packets transmitted on the given port will not undergo GSO. By default, GSO is disabled for all ports.

Note

When GSO is enabled on a port, supported IPv4 packets transmitted on that port undergo GSO. Afterwards, the segmented packets are represented by multi-segment mbufs; however, the csum forwarding engine doesn’t calculation of checksums for GSO’d segments in SW. As a result, if users want correct checksums in GSO segments, they should enable HW checksum calculation for GSO-enabled ports.

For example, HW checksum calculation for VxLAN GSO’d packets may be enabled by setting the following options in the csum forwarding engine:

testpmd> csum set outer_ip hw <port_id>

testpmd> csum set ip hw <port_id>

testpmd> csum set tcp hw <port_id>

set gso segsz

Set the maximum GSO segment size (measured in bytes), which includes the packet header and the packet payload for GSO-enabled ports (global):

testpmd> set gso segsz <length>
show port - gso

Display the status of Generic Segmentation Offload for a given port:

testpmd> show port <port_id> gso
mac_addr add

Add an alternative MAC address to a port:

testpmd> mac_addr add (port_id) (XX:XX:XX:XX:XX:XX)
mac_addr remove

Remove a MAC address from a port:

testpmd> mac_addr remove (port_id) (XX:XX:XX:XX:XX:XX)
mac_addr add (for VF)

Add an alternative MAC address for a VF to a port:

testpmd> mac_add add port (port_id) vf (vf_id) (XX:XX:XX:XX:XX:XX)
mac_addr set

Set the default MAC address for a port:

testpmd> mac_addr set (port_id) (XX:XX:XX:XX:XX:XX)
mac_addr set (for VF)

Set the MAC address for a VF from the PF:

testpmd> set vf mac addr (port_id) (vf_id) (XX:XX:XX:XX:XX:XX)
set port-uta

Set the unicast hash filter(s) on/off for a port:

testpmd> set port (port_id) uta (XX:XX:XX:XX:XX:XX|all) (on|off)
set promisc

Set the promiscuous mode on for a port or for all ports. In promiscuous mode packets are not dropped if they aren’t for the specified MAC address:

testpmd> set promisc (port_id|all) (on|off)
set allmulti

Set the allmulti mode for a port or for all ports:

testpmd> set allmulti (port_id|all) (on|off)

Same as the ifconfig (8) option. Controls how multicast packets are handled.

set promisc (for VF)

Set the unicast promiscuous mode for a VF from PF. It’s supported by Intel i40e NICs now. In promiscuous mode packets are not dropped if they aren’t for the specified MAC address:

testpmd> set vf promisc (port_id) (vf_id) (on|off)
set allmulticast (for VF)

Set the multicast promiscuous mode for a VF from PF. It’s supported by Intel i40e NICs now. In promiscuous mode packets are not dropped if they aren’t for the specified MAC address:

testpmd> set vf allmulti (port_id) (vf_id) (on|off)
set tx max bandwidth (for VF)

Set TX max absolute bandwidth (Mbps) for a VF from PF:

testpmd> set vf tx max-bandwidth (port_id) (vf_id) (max_bandwidth)
set tc tx min bandwidth (for VF)

Set all TCs’ TX min relative bandwidth (%) for a VF from PF:

testpmd> set vf tc tx min-bandwidth (port_id) (vf_id) (bw1, bw2, ...)
set tc tx max bandwidth (for VF)

Set a TC’s TX max absolute bandwidth (Mbps) for a VF from PF:

testpmd> set vf tc tx max-bandwidth (port_id) (vf_id) (tc_no) (max_bandwidth)
set tc tx min bandwidth

Set all TCs’ TX min relative bandwidth (%) globally for all PF and VFs:

testpmd> set tc tx min-bandwidth (port_id) (bw1, bw2, ...)
set flow_ctrl rx

Set the link flow control parameter on a port:

testpmd> set flow_ctrl rx (on|off) tx (on|off) (high_water) (low_water) \
         (pause_time) (send_xon) mac_ctrl_frame_fwd (on|off) \
         autoneg (on|off) (port_id)

Where:

  • high_water (integer): High threshold value to trigger XOFF.
  • low_water (integer): Low threshold value to trigger XON.
  • pause_time (integer): Pause quota in the Pause frame.
  • send_xon (0/1): Send XON frame.
  • mac_ctrl_frame_fwd: Enable receiving MAC control frames.
  • autoneg: Change the auto-negotiation parameter.
set pfc_ctrl rx

Set the priority flow control parameter on a port:

testpmd> set pfc_ctrl rx (on|off) tx (on|off) (high_water) (low_water) \
         (pause_time) (priority) (port_id)

Where:

  • high_water (integer): High threshold value.
  • low_water (integer): Low threshold value.
  • pause_time (integer): Pause quota in the Pause frame.
  • priority (0-7): VLAN User Priority.
set stat_qmap

Set statistics mapping (qmapping 0..15) for RX/TX queue on port:

testpmd> set stat_qmap (tx|rx) (port_id) (queue_id) (qmapping)

For example, to set rx queue 2 on port 0 to mapping 5:

testpmd>set stat_qmap rx 0 2 5
set xstats-hide-zero

Set the option to hide zero values for xstats display:

testpmd> set xstats-hide-zero on|off

Note

By default, the zero values are displayed for xstats.

set port - rx/tx (for VF)

Set VF receive/transmit from a port:

testpmd> set port (port_id) vf (vf_id) (rx|tx) (on|off)
set port - mac address filter (for VF)

Add/Remove unicast or multicast MAC addr filter for a VF:

testpmd> set port (port_id) vf (vf_id) (mac_addr) \
         (exact-mac|exact-mac-vlan|hashmac|hashmac-vlan) (on|off)
set port - rx mode(for VF)

Set the VF receive mode of a port:

testpmd> set port (port_id) vf (vf_id) \
         rxmode (AUPE|ROPE|BAM|MPE) (on|off)

The available receive modes are:

  • AUPE: Accepts untagged VLAN.
  • ROPE: Accepts unicast hash.
  • BAM: Accepts broadcast packets.
  • MPE: Accepts all multicast packets.
set port - tx_rate (for Queue)

Set TX rate limitation for a queue on a port:

testpmd> set port (port_id) queue (queue_id) rate (rate_value)
set port - tx_rate (for VF)

Set TX rate limitation for queues in VF on a port:

testpmd> set port (port_id) vf (vf_id) rate (rate_value) queue_mask (queue_mask)
set port - mirror rule

Set pool or vlan type mirror rule for a port:

testpmd> set port (port_id) mirror-rule (rule_id) \
         (pool-mirror-up|pool-mirror-down|vlan-mirror) \
         (poolmask|vlanid[,vlanid]*) dst-pool (pool_id) (on|off)

Set link mirror rule for a port:

testpmd> set port (port_id) mirror-rule (rule_id) \
        (uplink-mirror|downlink-mirror) dst-pool (pool_id) (on|off)

For example to enable mirror traffic with vlan 0,1 to pool 0:

set port 0 mirror-rule 0 vlan-mirror 0,1 dst-pool 0 on
reset port - mirror rule

Reset a mirror rule for a port:

testpmd> reset port (port_id) mirror-rule (rule_id)
set flush_rx

Set the flush on RX streams before forwarding. The default is flush on. Mainly used with PCAP drivers to turn off the default behavior of flushing the first 512 packets on RX streams:

testpmd> set flush_rx off
set bypass mode

Set the bypass mode for the lowest port on bypass enabled NIC:

testpmd> set bypass mode (normal|bypass|isolate) (port_id)
set bypass event

Set the event required to initiate specified bypass mode for the lowest port on a bypass enabled:

testpmd> set bypass event (timeout|os_on|os_off|power_on|power_off) \
         mode (normal|bypass|isolate) (port_id)

Where:

  • timeout: Enable bypass after watchdog timeout.
  • os_on: Enable bypass when OS/board is powered on.
  • os_off: Enable bypass when OS/board is powered off.
  • power_on: Enable bypass when power supply is turned on.
  • power_off: Enable bypass when power supply is turned off.
set bypass timeout

Set the bypass watchdog timeout to n seconds where 0 = instant:

testpmd> set bypass timeout (0|1.5|2|3|4|8|16|32)
show bypass config

Show the bypass configuration for a bypass enabled NIC using the lowest port on the NIC:

testpmd> show bypass config (port_id)
E-tag set

Enable E-tag insertion for a VF on a port:

testpmd> E-tag set insertion on port-tag-id (value) port (port_id) vf (vf_id)

Disable E-tag insertion for a VF on a port:

testpmd> E-tag set insertion off port (port_id) vf (vf_id)

Enable/disable E-tag stripping on a port:

testpmd> E-tag set stripping (on|off) port (port_id)

Enable/disable E-tag based forwarding on a port:

testpmd> E-tag set forwarding (on|off) port (port_id)

Add an E-tag forwarding filter on a port:

testpmd> E-tag set filter add e-tag-id (value) dst-pool (pool_id) port (port_id)
Delete an E-tag forwarding filter on a port::
testpmd> E-tag set filter del e-tag-id (value) port (port_id)
ddp add

Load a dynamic device personalization (DDP) package:

testpmd> ddp add (port_id) (package_path[,output_path])
ddp del

Delete a dynamic device personalization package:

testpmd> ddp del (port_id) (package_path)
ptype mapping

List all items from the ptype mapping table:

testpmd> ptype mapping get (port_id) (valid_only)

Where:

  • valid_only: A flag indicates if only list valid items(=1) or all itemss(=0).

Replace a specific or a group of software defined ptype with a new one:

testpmd> ptype mapping replace  (port_id) (target) (mask) (pkt_type)

where:

  • target: A specific software ptype or a mask to represent a group of software ptypes.
  • mask: A flag indicate if “target” is a specific software ptype(=0) or a ptype mask(=1).
  • pkt_type: The new software ptype to replace the old ones.

Update hardware defined ptype to software defined packet type mapping table:

testpmd> ptype mapping update (port_id) (hw_ptype) (sw_ptype)

where:

  • hw_ptype: hardware ptype as the index of the ptype mapping table.
  • sw_ptype: software ptype as the value of the ptype mapping table.

Reset ptype mapping table:

testpmd> ptype mapping reset (port_id)

Port Functions

The following sections show functions for configuring ports.

Note

Port configuration changes only become active when forwarding is started/restarted.

port attach

Attach a port specified by pci address or virtual device args:

testpmd> port attach (identifier)

To attach a new pci device, the device should be recognized by kernel first. Then it should be moved under DPDK management. Finally the port can be attached to testpmd.

For example, to move a pci device using ixgbe under DPDK management:

# Check the status of the available devices.
./usertools/dpdk-devbind.py --status

Network devices using DPDK-compatible driver
============================================
<none>

Network devices using kernel driver
===================================
0000:0a:00.0 '82599ES 10-Gigabit' if=eth2 drv=ixgbe unused=


# Bind the device to igb_uio.
sudo ./usertools/dpdk-devbind.py -b igb_uio 0000:0a:00.0


# Recheck the status of the devices.
./usertools/dpdk-devbind.py --status
Network devices using DPDK-compatible driver
============================================
0000:0a:00.0 '82599ES 10-Gigabit' drv=igb_uio unused=

To attach a port created by virtual device, above steps are not needed.

For example, to attach a port whose pci address is 0000:0a:00.0.

testpmd> port attach 0000:0a:00.0
Attaching a new port...
EAL: PCI device 0000:0a:00.0 on NUMA socket -1
EAL:   probe driver: 8086:10fb rte_ixgbe_pmd
EAL:   PCI memory mapped at 0x7f83bfa00000
EAL:   PCI memory mapped at 0x7f83bfa80000
PMD: eth_ixgbe_dev_init(): MAC: 2, PHY: 18, SFP+: 5
PMD: eth_ixgbe_dev_init(): port 0 vendorID=0x8086 deviceID=0x10fb
Port 0 is attached. Now total ports is 1
Done

For example, to attach a port created by pcap PMD.

testpmd> port attach net_pcap0
Attaching a new port...
PMD: Initializing pmd_pcap for net_pcap0
PMD: Creating pcap-backed ethdev on numa socket 0
Port 0 is attached. Now total ports is 1
Done

In this case, identifier is net_pcap0. This identifier format is the same as --vdev format of DPDK applications.

For example, to re-attach a bonded port which has been previously detached, the mode and slave parameters must be given.

testpmd> port attach net_bond_0,mode=0,slave=1
Attaching a new port...
EAL: Initializing pmd_bond for net_bond_0
EAL: Create bonded device net_bond_0 on port 0 in mode 0 on socket 0.
Port 0 is attached. Now total ports is 1
Done
port detach

Detach a specific port:

testpmd> port detach (port_id)

Before detaching a port, the port should be stopped and closed.

For example, to detach a pci device port 0.

testpmd> port stop 0
Stopping ports...
Done
testpmd> port close 0
Closing ports...
Done

testpmd> port detach 0
Detaching a port...
EAL: PCI device 0000:0a:00.0 on NUMA socket -1
EAL:   remove driver: 8086:10fb rte_ixgbe_pmd
EAL:   PCI memory unmapped at 0x7f83bfa00000
EAL:   PCI memory unmapped at 0x7f83bfa80000
Done

For example, to detach a virtual device port 0.

testpmd> port stop 0
Stopping ports...
Done
testpmd> port close 0
Closing ports...
Done

testpmd> port detach 0
Detaching a port...
PMD: Closing pcap ethdev on numa socket 0
Port 'net_pcap0' is detached. Now total ports is 0
Done

To remove a pci device completely from the system, first detach the port from testpmd. Then the device should be moved under kernel management. Finally the device can be removed using kernel pci hotplug functionality.

For example, to move a pci device under kernel management:

sudo ./usertools/dpdk-devbind.py -b ixgbe 0000:0a:00.0

./usertools/dpdk-devbind.py --status

Network devices using DPDK-compatible driver
============================================
<none>

Network devices using kernel driver
===================================
0000:0a:00.0 '82599ES 10-Gigabit' if=eth2 drv=ixgbe unused=igb_uio

To remove a port created by a virtual device, above steps are not needed.

port start

Start all ports or a specific port:

testpmd> port start (port_id|all)
port stop

Stop all ports or a specific port:

testpmd> port stop (port_id|all)
port close

Close all ports or a specific port:

testpmd> port close (port_id|all)
port start/stop queue

Start/stop a rx/tx queue on a specific port:

testpmd> port (port_id) (rxq|txq) (queue_id) (start|stop)

Only take effect when port is started.

port config - speed

Set the speed and duplex mode for all ports or a specific port:

testpmd> port config (port_id|all) speed (10|100|1000|10000|25000|40000|50000|100000|auto) \
         duplex (half|full|auto)
port config - queues/descriptors

Set number of queues/descriptors for rxq, txq, rxd and txd:

testpmd> port config all (rxq|txq|rxd|txd) (value)

This is equivalent to the --rxq, --txq, --rxd and --txd command-line options.

port config - max-pkt-len

Set the maximum packet length:

testpmd> port config all max-pkt-len (value)

This is equivalent to the --max-pkt-len command-line option.

port config - CRC Strip

Set hardware CRC stripping on or off for all ports:

testpmd> port config all crc-strip (on|off)

CRC stripping is on by default.

The off option is equivalent to the --disable-crc-strip command-line option.

port config - scatter

Set RX scatter mode on or off for all ports:

testpmd> port config all scatter (on|off)

RX scatter mode is off by default.

The on option is equivalent to the --enable-scatter command-line option.

port config - TX queue flags

Set a hexadecimal bitmap of TX queue flags for all ports:

testpmd> port config all txqflags value

This command is equivalent to the --txqflags command-line option.

port config - RX Checksum

Set hardware RX checksum offload to on or off for all ports:

testpmd> port config all rx-cksum (on|off)

Checksum offload is off by default.

The on option is equivalent to the --enable-rx-cksum command-line option.

port config - VLAN

Set hardware VLAN on or off for all ports:

testpmd> port config all hw-vlan (on|off)

Hardware VLAN is on by default.

The off option is equivalent to the --disable-hw-vlan command-line option.

port config - VLAN filter

Set hardware VLAN filter on or off for all ports:

testpmd> port config all hw-vlan-filter (on|off)

Hardware VLAN filter is on by default.

The off option is equivalent to the --disable-hw-vlan-filter command-line option.

port config - VLAN strip

Set hardware VLAN strip on or off for all ports:

testpmd> port config all hw-vlan-strip (on|off)

Hardware VLAN strip is on by default.

The off option is equivalent to the --disable-hw-vlan-strip command-line option.

port config - VLAN extend

Set hardware VLAN extend on or off for all ports:

testpmd> port config all hw-vlan-extend (on|off)

Hardware VLAN extend is off by default.

The off option is equivalent to the --disable-hw-vlan-extend command-line option.

port config - Drop Packets

Set packet drop for packets with no descriptors on or off for all ports:

testpmd> port config all drop-en (on|off)

Packet dropping for packets with no descriptors is off by default.

The on option is equivalent to the --enable-drop-en command-line option.

port config - RSS

Set the RSS (Receive Side Scaling) mode on or off:

testpmd> port config all rss (all|ip|tcp|udp|sctp|ether|port|vxlan|geneve|nvgre|none)

RSS is on by default.

The none option is equivalent to the --disable-rss command-line option.

port config - RSS Reta

Set the RSS (Receive Side Scaling) redirection table:

testpmd> port config all rss reta (hash,queue)[,(hash,queue)]
port config - DCB

Set the DCB mode for an individual port:

testpmd> port config (port_id) dcb vt (on|off) (traffic_class) pfc (on|off)

The traffic class should be 4 or 8.

port config - Burst

Set the number of packets per burst:

testpmd> port config all burst (value)

This is equivalent to the --burst command-line option.

port config - Threshold

Set thresholds for TX/RX queues:

testpmd> port config all (threshold) (value)

Where the threshold type can be:

  • txpt: Set the prefetch threshold register of the TX rings, 0 <= value <= 255.
  • txht: Set the host threshold register of the TX rings, 0 <= value <= 255.
  • txwt: Set the write-back threshold register of the TX rings, 0 <= value <= 255.
  • rxpt: Set the prefetch threshold register of the RX rings, 0 <= value <= 255.
  • rxht: Set the host threshold register of the RX rings, 0 <= value <= 255.
  • rxwt: Set the write-back threshold register of the RX rings, 0 <= value <= 255.
  • txfreet: Set the transmit free threshold of the TX rings, 0 <= value <= txd.
  • rxfreet: Set the transmit free threshold of the RX rings, 0 <= value <= rxd.
  • txrst: Set the transmit RS bit threshold of TX rings, 0 <= value <= txd.

These threshold options are also available from the command-line.

port config - E-tag

Set the value of ether-type for E-tag:

testpmd> port config (port_id|all) l2-tunnel E-tag ether-type (value)

Enable/disable the E-tag support:

testpmd> port config (port_id|all) l2-tunnel E-tag (enable|disable)
port config pctype mapping

Reset pctype mapping table:

testpmd> port config (port_id) pctype mapping reset

Update hardware defined pctype to software defined flow type mapping table:

testpmd> port config (port_id) pctype mapping update (pctype_id_0[,pctype_id_1]*) (flow_type_id)

where:

  • pctype_id_x: hardware pctype id as index of bit in bitmask value of the pctype mapping table.
  • flow_type_id: software flow type id as the index of the pctype mapping table.

Register Functions

The Register Functions can be used to read from and write to registers on the network card referenced by a port number. This is mainly useful for debugging purposes. Reference should be made to the appropriate datasheet for the network card for details on the register addresses and fields that can be accessed.

read reg

Display the value of a port register:

testpmd> read reg (port_id) (address)

For example, to examine the Flow Director control register (FDIRCTL, 0x0000EE000) on an Intel 82599 10 GbE Controller:

testpmd> read reg 0 0xEE00
port 0 PCI register at offset 0xEE00: 0x4A060029 (1241907241)
read regfield

Display a port register bit field:

testpmd> read regfield (port_id) (address) (bit_x) (bit_y)

For example, reading the lowest two bits from the register in the example above:

testpmd> read regfield 0 0xEE00 0 1
port 0 PCI register at offset 0xEE00: bits[0, 1]=0x1 (1)
read regbit

Display a single port register bit:

testpmd> read regbit (port_id) (address) (bit_x)

For example, reading the lowest bit from the register in the example above:

testpmd> read regbit 0 0xEE00 0
port 0 PCI register at offset 0xEE00: bit 0=1
write reg

Set the value of a port register:

testpmd> write reg (port_id) (address) (value)

For example, to clear a register:

testpmd> write reg 0 0xEE00 0x0
port 0 PCI register at offset 0xEE00: 0x00000000 (0)
write regfield

Set bit field of a port register:

testpmd> write regfield (port_id) (address) (bit_x) (bit_y) (value)

For example, writing to the register cleared in the example above:

testpmd> write regfield 0 0xEE00 0 1 2
port 0 PCI register at offset 0xEE00: 0x00000002 (2)
write regbit

Set single bit value of a port register:

testpmd> write regbit (port_id) (address) (bit_x) (value)

For example, to set the high bit in the register from the example above:

testpmd> write regbit 0 0xEE00 31 1
port 0 PCI register at offset 0xEE00: 0x8000000A (2147483658)

Traffic Management

The following section shows functions for configuring traffic management on on the ethernet device through the use of generic TM API.

show port traffic management capability

Show traffic management capability of the port:

testpmd> show port tm cap (port_id)
show port traffic management capability (hierarchy level)

Show traffic management hierarchy level capability of the port:

testpmd> show port tm cap (port_id) (level_id)
show port traffic management capability (hierarchy node level)

Show the traffic management hierarchy node capability of the port:

testpmd> show port tm cap (port_id) (node_id)
show port traffic management hierarchy node type

Show the port traffic management hierarchy node type:

testpmd> show port tm node type (port_id) (node_id)
show port traffic management hierarchy node stats

Show the port traffic management hierarchy node statistics:

testpmd> show port tm node stats (port_id) (node_id) (clear)

where:

  • clear: When this parameter has a non-zero value, the statistics counters are cleared (i.e. set to zero) immediately after they have been read, otherwise the statistics counters are left untouched.
Add port traffic management private shaper profile

Add the port traffic management private shaper profile:

testpmd> add port tm node shaper profile (port_id) (shaper_profile_id) \
(tb_rate) (tb_size) (packet_length_adjust)

where:

  • shaper_profile id: Shaper profile ID for the new profile.
  • tb_rate: Token bucket rate (bytes per second).
  • tb_size: Token bucket size (bytes).
  • packet_length_adjust: The value (bytes) to be added to the length of each packet for the purpose of shaping. This parameter value can be used to correct the packet length with the framing overhead bytes that are consumed on the wire.
Delete port traffic management private shaper profile

Delete the port traffic management private shaper:

testpmd> del port tm node shaper profile (port_id) (shaper_profile_id)

where:

  • shaper_profile id: Shaper profile ID that needs to be deleted.
Add port traffic management shared shaper

Create the port traffic management shared shaper:

testpmd> add port tm node shared shaper (port_id) (shared_shaper_id) \
(shaper_profile_id)

where:

  • shared_shaper_id: Shared shaper ID to be created.
  • shaper_profile id: Shaper profile ID for shared shaper.
Set port traffic management shared shaper

Update the port traffic management shared shaper:

testpmd> set port tm node shared shaper (port_id) (shared_shaper_id) \
(shaper_profile_id)

where:

  • shared_shaper_id: Shared shaper ID to be update.
  • shaper_profile id: Shaper profile ID for shared shaper.
Delete port traffic management shared shaper

Delete the port traffic management shared shaper:

testpmd> del port tm node shared shaper (port_id) (shared_shaper_id)

where:

  • shared_shaper_id: Shared shaper ID to be deleted.
Set port traffic management hiearchy node private shaper

set the port traffic management hierarchy node private shaper:

testpmd> set port tm node shaper profile (port_id) (node_id) \
(shaper_profile_id)

where:

  • shaper_profile id: Private shaper profile ID to be enabled on the hierarchy node.
Add port traffic management WRED profile

Create a new WRED profile:

testpmd> add port tm node wred profile (port_id) (wred_profile_id) \
(color_g) (min_th_g) (max_th_g) (maxp_inv_g) (wq_log2_g) \
(color_y) (min_th_y) (max_th_y) (maxp_inv_y) (wq_log2_y) \
(color_r) (min_th_r) (max_th_r) (maxp_inv_r) (wq_log2_r)

where:

  • wred_profile id: Identifier for the newly create WRED profile
  • color_g: Packet color (green)
  • min_th_g: Minimum queue threshold for packet with green color
  • max_th_g: Minimum queue threshold for packet with green color
  • maxp_inv_g: Inverse of packet marking probability maximum value (maxp)
  • wq_log2_g: Negated log2 of queue weight (wq)
  • color_y: Packet color (yellow)
  • min_th_y: Minimum queue threshold for packet with yellow color
  • max_th_y: Minimum queue threshold for packet with yellow color
  • maxp_inv_y: Inverse of packet marking probability maximum value (maxp)
  • wq_log2_y: Negated log2 of queue weight (wq)
  • color_r: Packet color (red)
  • min_th_r: Minimum queue threshold for packet with yellow color
  • max_th_r: Minimum queue threshold for packet with yellow color
  • maxp_inv_r: Inverse of packet marking probability maximum value (maxp)
  • wq_log2_r: Negated log2 of queue weight (wq)
Delete port traffic management WRED profile

Delete the WRED profile:

testpmd> del port tm node wred profile (port_id) (wred_profile_id)
Add port traffic management hierarchy nonleaf node

Add nonleaf node to port traffic management hiearchy:

testpmd> add port tm nonleaf node (port_id) (node_id) (parent_node_id) \
(priority) (weight) (level_id) (shaper_profile_id) \
(n_sp_priorities) (stats_mask) (n_shared_shapers) \
[(shared_shaper_0) (shared_shaper_1) ...] \

where:

  • parent_node_id: Node ID of the parent.
  • priority: Node priority (highest node priority is zero). This is used by the SP algorithm running on the parent node for scheduling this node.
  • weight: Node weight (lowest weight is one). The node weight is relative to the weight sum of all siblings that have the same priority. It is used by the WFQ algorithm running on the parent node for scheduling this node.
  • level_id: Hiearchy level of the node.
  • shaper_profile_id: Shaper profile ID of the private shaper to be used by the node.
  • n_sp_priorities: Number of strict priorities.
  • stats_mask: Mask of statistics counter types to be enabled for this node.
  • n_shared_shapers: Number of shared shapers.
  • shared_shaper_id: Shared shaper id.
Add port traffic management hierarchy leaf node

Add leaf node to port traffic management hiearchy:

testpmd> add port tm leaf node (port_id) (node_id) (parent_node_id) \
(priority) (weight) (level_id) (shaper_profile_id) \
(cman_mode) (wred_profile_id) (stats_mask) (n_shared_shapers) \
[(shared_shaper_id) (shared_shaper_id) ...] \

where:

  • parent_node_id: Node ID of the parent.
  • priority: Node priority (highest node priority is zero). This is used by the SP algorithm running on the parent node for scheduling this node.
  • weight: Node weight (lowest weight is one). The node weight is relative to the weight sum of all siblings that have the same priority. It is used by the WFQ algorithm running on the parent node for scheduling this node.
  • level_id: Hiearchy level of the node.
  • shaper_profile_id: Shaper profile ID of the private shaper to be used by the node.
  • cman_mode: Congestion management mode to be enabled for this node.
  • wred_profile_id: WRED profile id to be enabled for this node.
  • stats_mask: Mask of statistics counter types to be enabled for this node.
  • n_shared_shapers: Number of shared shapers.
  • shared_shaper_id: Shared shaper id.
Delete port traffic management hierarchy node

Delete node from port traffic management hiearchy:

testpmd> del port tm node (port_id) (node_id)
Update port traffic management hierarchy parent node

Update port traffic management hierarchy parent node:

testpmd> set port tm node parent (port_id) (node_id) (parent_node_id) \
(priority) (weight)

This function can only be called after the hierarchy commit invocation. Its success depends on the port support for this operation, as advertised through the port capability set. This function is valid for all nodes of the traffic management hierarchy except root node.

Commit port traffic management hierarchy

Commit the traffic management hierarchy on the port:

testpmd> port tm hierarchy commit (port_id) (clean_on_fail)

where:

  • clean_on_fail: When set to non-zero, hierarchy is cleared on function call failure. On the other hand, hierarchy is preserved when this parameter is equal to zero.
Set port traffic management default hierarchy (tm forwarding mode)

set the traffic management default hierarchy on the port:

testpmd> set port tm hierarchy default (port_id)

Filter Functions

This section details the available filter functions that are available.

Note these functions interface the deprecated legacy filtering framework, superseded by rte_flow. See Flow rules management.

ethertype_filter

Add or delete a L2 Ethertype filter, which identify packets by their L2 Ethertype mainly assign them to a receive queue:

ethertype_filter (port_id) (add|del) (mac_addr|mac_ignr) (mac_address) \
                 ethertype (ether_type) (drop|fwd) queue (queue_id)

The available information parameters are:

  • port_id: The port which the Ethertype filter assigned on.
  • mac_addr: Compare destination mac address.
  • mac_ignr: Ignore destination mac address match.
  • mac_address: Destination mac address to match.
  • ether_type: The EtherType value want to match, for example 0x0806 for ARP packet. 0x0800 (IPv4) and 0x86DD (IPv6) are invalid.
  • queue_id: The receive queue associated with this EtherType filter. It is meaningless when deleting or dropping.

Example, to add/remove an ethertype filter rule:

testpmd> ethertype_filter 0 add mac_ignr 00:11:22:33:44:55 \
                          ethertype 0x0806 fwd queue 3

testpmd> ethertype_filter 0 del mac_ignr 00:11:22:33:44:55 \
                          ethertype 0x0806 fwd queue 3
2tuple_filter

Add or delete a 2-tuple filter, which identifies packets by specific protocol and destination TCP/UDP port and forwards packets into one of the receive queues:

2tuple_filter (port_id) (add|del) dst_port (dst_port_value) \
              protocol (protocol_value) mask (mask_value) \
              tcp_flags (tcp_flags_value) priority (prio_value) \
              queue (queue_id)

The available information parameters are:

  • port_id: The port which the 2-tuple filter assigned on.
  • dst_port_value: Destination port in L4.
  • protocol_value: IP L4 protocol.
  • mask_value: Participates in the match or not by bit for field above, 1b means participate.
  • tcp_flags_value: TCP control bits. The non-zero value is invalid, when the pro_value is not set to 0x06 (TCP).
  • prio_value: Priority of this filter.
  • queue_id: The receive queue associated with this 2-tuple filter.

Example, to add/remove an 2tuple filter rule:

testpmd> 2tuple_filter 0 add dst_port 32 protocol 0x06 mask 0x03 \
                       tcp_flags 0x02 priority 3 queue 3

testpmd> 2tuple_filter 0 del dst_port 32 protocol 0x06 mask 0x03 \
                       tcp_flags 0x02 priority 3 queue 3
5tuple_filter

Add or delete a 5-tuple filter, which consists of a 5-tuple (protocol, source and destination IP addresses, source and destination TCP/UDP/SCTP port) and routes packets into one of the receive queues:

5tuple_filter (port_id) (add|del) dst_ip (dst_address) src_ip \
              (src_address) dst_port (dst_port_value) \
              src_port (src_port_value) protocol (protocol_value) \
              mask (mask_value) tcp_flags (tcp_flags_value) \
              priority (prio_value) queue (queue_id)

The available information parameters are:

  • port_id: The port which the 5-tuple filter assigned on.
  • dst_address: Destination IP address.
  • src_address: Source IP address.
  • dst_port_value: TCP/UDP destination port.
  • src_port_value: TCP/UDP source port.
  • protocol_value: L4 protocol.
  • mask_value: Participates in the match or not by bit for field above, 1b means participate
  • tcp_flags_value: TCP control bits. The non-zero value is invalid, when the protocol_value is not set to 0x06 (TCP).
  • prio_value: The priority of this filter.
  • queue_id: The receive queue associated with this 5-tuple filter.

Example, to add/remove an 5tuple filter rule:

testpmd> 5tuple_filter 0 add dst_ip 2.2.2.5 src_ip 2.2.2.4 \
         dst_port 64 src_port 32 protocol 0x06 mask 0x1F \
         flags 0x0 priority 3 queue 3

testpmd> 5tuple_filter 0 del dst_ip 2.2.2.5 src_ip 2.2.2.4 \
         dst_port 64 src_port 32 protocol 0x06 mask 0x1F \
         flags 0x0 priority 3 queue 3
syn_filter

Using the SYN filter, TCP packets whose SYN flag is set can be forwarded to a separate queue:

syn_filter (port_id) (add|del) priority (high|low) queue (queue_id)

The available information parameters are:

  • port_id: The port which the SYN filter assigned on.
  • high: This SYN filter has higher priority than other filters.
  • low: This SYN filter has lower priority than other filters.
  • queue_id: The receive queue associated with this SYN filter

Example:

testpmd> syn_filter 0 add priority high queue 3
flex_filter

With flex filter, packets can be recognized by any arbitrary pattern within the first 128 bytes of the packet and routed into one of the receive queues:

flex_filter (port_id) (add|del) len (len_value) bytes (bytes_value) \
            mask (mask_value) priority (prio_value) queue (queue_id)

The available information parameters are:

  • port_id: The port which the Flex filter is assigned on.
  • len_value: Filter length in bytes, no greater than 128.
  • bytes_value: A string in hexadecimal, means the value the flex filter needs to match.
  • mask_value: A string in hexadecimal, bit 1 means corresponding byte participates in the match.
  • prio_value: The priority of this filter.
  • queue_id: The receive queue associated with this Flex filter.

Example:

testpmd> flex_filter 0 add len 16 bytes 0x00000000000000000000000008060000 \
                       mask 000C priority 3 queue 3

testpmd> flex_filter 0 del len 16 bytes 0x00000000000000000000000008060000 \
                       mask 000C priority 3 queue 3
flow_director_filter

The Flow Director works in receive mode to identify specific flows or sets of flows and route them to specific queues.

Four types of filtering are supported which are referred to as Perfect Match, Signature, Perfect-mac-vlan and Perfect-tunnel filters, the match mode is set by the --pkt-filter-mode command-line parameter:

  • Perfect match filters. The hardware checks a match between the masked fields of the received packets and the programmed filters. The masked fields are for IP flow.
  • Signature filters. The hardware checks a match between a hash-based signature of the masked fields of the received packet.
  • Perfect-mac-vlan match filters. The hardware checks a match between the masked fields of the received packets and the programmed filters. The masked fields are for MAC VLAN flow.
  • Perfect-tunnel match filters. The hardware checks a match between the masked fields of the received packets and the programmed filters. The masked fields are for tunnel flow.

The Flow Director filters can match the different fields for different type of packet: flow type, specific input set per flow type and the flexible payload.

The Flow Director can also mask out parts of all of these fields so that filters are only applied to certain fields or parts of the fields.

Different NICs may have different capabilities, command show port fdir (port_id) can be used to acquire the information.

# Commands to add flow director filters of different flow types:

flow_director_filter (port_id) mode IP (add|del|update) \
                     flow (ipv4-other|ipv4-frag|ipv6-other|ipv6-frag) \
                     src (src_ip_address) dst (dst_ip_address) \
                     tos (tos_value) proto (proto_value) ttl (ttl_value) \
                     vlan (vlan_value) flexbytes (flexbytes_value) \
                     (drop|fwd) pf|vf(vf_id) queue (queue_id) \
                     fd_id (fd_id_value)

flow_director_filter (port_id) mode IP (add|del|update) \
                     flow (ipv4-tcp|ipv4-udp|ipv6-tcp|ipv6-udp) \
                     src (src_ip_address) (src_port) \
                     dst (dst_ip_address) (dst_port) \
                     tos (tos_value) ttl (ttl_value) \
                     vlan (vlan_value) flexbytes (flexbytes_value) \
                     (drop|fwd) queue pf|vf(vf_id) (queue_id) \
                     fd_id (fd_id_value)

flow_director_filter (port_id) mode IP (add|del|update) \
                     flow (ipv4-sctp|ipv6-sctp) \
                     src (src_ip_address) (src_port) \
                     dst (dst_ip_address) (dst_port) \
                     tos (tos_value) ttl (ttl_value) \
                     tag (verification_tag) vlan (vlan_value) \
                     flexbytes (flexbytes_value) (drop|fwd) \
                     pf|vf(vf_id) queue (queue_id) fd_id (fd_id_value)

flow_director_filter (port_id) mode IP (add|del|update) flow l2_payload \
                     ether (ethertype) flexbytes (flexbytes_value) \
                     (drop|fwd) pf|vf(vf_id) queue (queue_id)
                     fd_id (fd_id_value)

flow_director_filter (port_id) mode MAC-VLAN (add|del|update) \
                     mac (mac_address) vlan (vlan_value) \
                     flexbytes (flexbytes_value) (drop|fwd) \
                     queue (queue_id) fd_id (fd_id_value)

flow_director_filter (port_id) mode Tunnel (add|del|update) \
                     mac (mac_address) vlan (vlan_value) \
                     tunnel (NVGRE|VxLAN) tunnel-id (tunnel_id_value) \
                     flexbytes (flexbytes_value) (drop|fwd) \
                     queue (queue_id) fd_id (fd_id_value)

For example, to add an ipv4-udp flow type filter:

testpmd> flow_director_filter 0 mode IP add flow ipv4-udp src 2.2.2.3 32 \
         dst 2.2.2.5 33 tos 2 ttl 40 vlan 0x1 flexbytes (0x88,0x48) \
         fwd pf queue 1 fd_id 1

For example, add an ipv4-other flow type filter:

testpmd> flow_director_filter 0 mode IP add flow ipv4-other src 2.2.2.3 \
          dst 2.2.2.5 tos 2 proto 20 ttl 40 vlan 0x1 \
          flexbytes (0x88,0x48) fwd pf queue 1 fd_id 1
flush_flow_director

Flush all flow director filters on a device:

testpmd> flush_flow_director (port_id)

Example, to flush all flow director filter on port 0:

testpmd> flush_flow_director 0
flow_director_mask

Set flow director’s input masks:

flow_director_mask (port_id) mode IP vlan (vlan_value) \
                   src_mask (ipv4_src) (ipv6_src) (src_port) \
                   dst_mask (ipv4_dst) (ipv6_dst) (dst_port)

flow_director_mask (port_id) mode MAC-VLAN vlan (vlan_value)

flow_director_mask (port_id) mode Tunnel vlan (vlan_value) \
                   mac (mac_value) tunnel-type (tunnel_type_value) \
                   tunnel-id (tunnel_id_value)

Example, to set flow director mask on port 0:

testpmd> flow_director_mask 0 mode IP vlan 0xefff \
         src_mask 255.255.255.255 \
             FFFF:FFFF:FFFF:FFFF:FFFF:FFFF:FFFF:FFFF 0xFFFF \
         dst_mask 255.255.255.255 \
             FFFF:FFFF:FFFF:FFFF:FFFF:FFFF:FFFF:FFFF 0xFFFF
flow_director_flex_mask

set masks of flow director’s flexible payload based on certain flow type:

testpmd> flow_director_flex_mask (port_id) \
         flow (none|ipv4-other|ipv4-frag|ipv4-tcp|ipv4-udp|ipv4-sctp| \
               ipv6-other|ipv6-frag|ipv6-tcp|ipv6-udp|ipv6-sctp| \
               l2_payload|all) (mask)

Example, to set flow director’s flex mask for all flow type on port 0:

testpmd> flow_director_flex_mask 0 flow all \
         (0xff,0xff,0,0,0,0,0,0,0,0,0,0,0,0,0,0)
flow_director_flex_payload

Configure flexible payload selection:

flow_director_flex_payload (port_id) (raw|l2|l3|l4) (config)

For example, to select the first 16 bytes from the offset 4 (bytes) of packet’s payload as flexible payload:

testpmd> flow_director_flex_payload 0 l4 \
         (4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19)
get_sym_hash_ena_per_port

Get symmetric hash enable configuration per port:

get_sym_hash_ena_per_port (port_id)

For example, to get symmetric hash enable configuration of port 1:

testpmd> get_sym_hash_ena_per_port 1
set_sym_hash_ena_per_port

Set symmetric hash enable configuration per port to enable or disable:

set_sym_hash_ena_per_port (port_id) (enable|disable)

For example, to set symmetric hash enable configuration of port 1 to enable:

testpmd> set_sym_hash_ena_per_port 1 enable
get_hash_global_config

Get the global configurations of hash filters:

get_hash_global_config (port_id)

For example, to get the global configurations of hash filters of port 1:

testpmd> get_hash_global_config 1
set_hash_global_config

Set the global configurations of hash filters:

set_hash_global_config (port_id) (toeplitz|simple_xor|default) \
(ipv4|ipv4-frag|ipv4-tcp|ipv4-udp|ipv4-sctp|ipv4-other|ipv6|ipv6-frag| \
ipv6-tcp|ipv6-udp|ipv6-sctp|ipv6-other|l2_payload) \
(enable|disable)

For example, to enable simple_xor for flow type of ipv6 on port 2:

testpmd> set_hash_global_config 2 simple_xor ipv6 enable
set_hash_input_set

Set the input set for hash:

set_hash_input_set (port_id) (ipv4-frag|ipv4-tcp|ipv4-udp|ipv4-sctp| \
ipv4-other|ipv6-frag|ipv6-tcp|ipv6-udp|ipv6-sctp|ipv6-other| \
l2_payload) (ovlan|ivlan|src-ipv4|dst-ipv4|src-ipv6|dst-ipv6|ipv4-tos| \
ipv4-proto|ipv6-tc|ipv6-next-header|udp-src-port|udp-dst-port| \
tcp-src-port|tcp-dst-port|sctp-src-port|sctp-dst-port|sctp-veri-tag| \
udp-key|gre-key|fld-1st|fld-2nd|fld-3rd|fld-4th|fld-5th|fld-6th|fld-7th| \
fld-8th|none) (select|add)

For example, to add source IP to hash input set for flow type of ipv4-udp on port 0:

testpmd> set_hash_input_set 0 ipv4-udp src-ipv4 add
set_fdir_input_set

The Flow Director filters can match the different fields for different type of packet, i.e. specific input set on per flow type and the flexible payload. This command can be used to change input set for each flow type.

Set the input set for flow director:

set_fdir_input_set (port_id) (ipv4-frag|ipv4-tcp|ipv4-udp|ipv4-sctp| \
ipv4-other|ipv6|ipv6-frag|ipv6-tcp|ipv6-udp|ipv6-sctp|ipv6-other| \
l2_payload) (ivlan|ethertype|src-ipv4|dst-ipv4|src-ipv6|dst-ipv6|ipv4-tos| \
ipv4-proto|ipv4-ttl|ipv6-tc|ipv6-next-header|ipv6-hop-limits| \
tudp-src-port|udp-dst-port|cp-src-port|tcp-dst-port|sctp-src-port| \
sctp-dst-port|sctp-veri-tag|none) (select|add)

For example to add source IP to FD input set for flow type of ipv4-udp on port 0:

testpmd> set_fdir_input_set 0 ipv4-udp src-ipv4 add
global_config

Set different GRE key length for input set:

global_config (port_id) gre-key-len (number in bytes)

For example to set GRE key length for input set to 4 bytes on port 0:

testpmd> global_config 0 gre-key-len 4

Flow rules management

Control of the generic flow API (rte_flow) is fully exposed through the flow command (validation, creation, destruction, queries and operation modes).

Considering rte_flow overlaps with all Filter Functions, using both features simultaneously may cause undefined side-effects and is therefore not recommended.

flow syntax

Because the flow command uses dynamic tokens to handle the large number of possible flow rules combinations, its behavior differs slightly from other commands, in particular:

  • Pressing ? or the <tab> key displays contextual help for the current token, not that of the entire command.
  • Optional and repeated parameters are supported (provided they are listed in the contextual help).

The first parameter stands for the operation mode. Possible operations and their general syntax are described below. They are covered in detail in the following sections.

  • Check whether a flow rule can be created:

    flow validate {port_id}
        [group {group_id}] [priority {level}] [ingress] [egress]
        pattern {item} [/ {item} [...]] / end
        actions {action} [/ {action} [...]] / end
    
  • Create a flow rule:

    flow create {port_id}
        [group {group_id}] [priority {level}] [ingress] [egress]
        pattern {item} [/ {item} [...]] / end
        actions {action} [/ {action} [...]] / end
    
  • Destroy specific flow rules:

    flow destroy {port_id} rule {rule_id} [...]
    
  • Destroy all flow rules:

    flow flush {port_id}
    
  • Query an existing flow rule:

    flow query {port_id} {rule_id} {action}
    
  • List existing flow rules sorted by priority, filtered by group identifiers:

    flow list {port_id} [group {group_id}] [...]
    
  • Restrict ingress traffic to the defined flow rules:

    flow isolate {port_id} {boolean}
    
Validating flow rules

flow validate reports whether a flow rule would be accepted by the underlying device in its current state but stops short of creating it. It is bound to rte_flow_validate():

flow validate {port_id}
   [group {group_id}] [priority {level}] [ingress] [egress]
   pattern {item} [/ {item} [...]] / end
   actions {action} [/ {action} [...]] / end

If successful, it will show:

Flow rule validated

Otherwise it will show an error message of the form:

Caught error type [...] ([...]): [...]

This command uses the same parameters as flow create, their format is described in Creating flow rules.

Check whether redirecting any Ethernet packet received on port 0 to RX queue index 6 is supported:

testpmd> flow validate 0 ingress pattern eth / end
   actions queue index 6 / end
Flow rule validated
testpmd>

Port 0 does not support TCPv6 rules:

testpmd> flow validate 0 ingress pattern eth / ipv6 / tcp / end
   actions drop / end
Caught error type 9 (specific pattern item): Invalid argument
testpmd>
Creating flow rules

flow create validates and creates the specified flow rule. It is bound to rte_flow_create():

flow create {port_id}
   [group {group_id}] [priority {level}] [ingress] [egress]
   pattern {item} [/ {item} [...]] / end
   actions {action} [/ {action} [...]] / end

If successful, it will return a flow rule ID usable with other commands:

Flow rule #[...] created

Otherwise it will show an error message of the form:

Caught error type [...] ([...]): [...]

Parameters describe in the following order:

  • Attributes (group, priority, ingress, egress tokens).
  • A matching pattern, starting with the pattern token and terminated by an end pattern item.
  • Actions, starting with the actions token and terminated by an end action.

These translate directly to rte_flow objects provided as-is to the underlying functions.

The shortest valid definition only comprises mandatory tokens:

testpmd> flow create 0 pattern end actions end

Note that PMDs may refuse rules that essentially do nothing such as this one.

All unspecified object values are automatically initialized to 0.

Attributes

These tokens affect flow rule attributes (struct rte_flow_attr) and are specified before the pattern token.

  • group {group id}: priority group.
  • priority {level}: priority level within group.
  • ingress: rule applies to ingress traffic.
  • egress: rule applies to egress traffic.

Each instance of an attribute specified several times overrides the previous value as shown below (group 4 is used):

testpmd> flow create 0 group 42 group 24 group 4 [...]

Note that once enabled, ingress and egress cannot be disabled.

While not specifying a direction is an error, some rules may allow both simultaneously.

Most rules affect RX therefore contain the ingress token:

testpmd> flow create 0 ingress pattern [...]
Matching pattern

A matching pattern starts after the pattern token. It is made of pattern items and is terminated by a mandatory end item.

Items are named after their type (RTE_FLOW_ITEM_TYPE_ from enum rte_flow_item_type).

The / token is used as a separator between pattern items as shown below:

testpmd> flow create 0 ingress pattern eth / ipv4 / udp / end [...]

Note that protocol items like these must be stacked from lowest to highest layer to make sense. For instance, the following rule is either invalid or unlikely to match any packet:

testpmd> flow create 0 ingress pattern eth / udp / ipv4 / end [...]

More information on these restrictions can be found in the rte_flow documentation.

Several items support additional specification structures, for example ipv4 allows specifying source and destination addresses as follows:

testpmd> flow create 0 ingress pattern eth / ipv4 src is 10.1.1.1
   dst is 10.2.0.0 / end [...]

This rule matches all IPv4 traffic with the specified properties.

In this example, src and dst are field names of the underlying struct rte_flow_item_ipv4 object. All item properties can be specified in a similar fashion.

The is token means that the subsequent value must be matched exactly, and assigns spec and mask fields in struct rte_flow_item accordingly. Possible assignment tokens are:

  • is: match value perfectly (with full bit-mask).
  • spec: match value according to configured bit-mask.
  • last: specify upper bound to establish a range.
  • mask: specify bit-mask with relevant bits set to one.
  • prefix: generate bit-mask from a prefix length.

These yield identical results:

ipv4 src is 10.1.1.1
ipv4 src spec 10.1.1.1 src mask 255.255.255.255
ipv4 src spec 10.1.1.1 src prefix 32
ipv4 src is 10.1.1.1 src last 10.1.1.1 # range with a single value
ipv4 src is 10.1.1.1 src last 0 # 0 disables range

Inclusive ranges can be defined with last:

ipv4 src is 10.1.1.1 src last 10.2.3.4 # 10.1.1.1 to 10.2.3.4

Note that mask affects both spec and last:

ipv4 src is 10.1.1.1 src last 10.2.3.4 src mask 255.255.0.0
   # matches 10.1.0.0 to 10.2.255.255

Properties can be modified multiple times:

ipv4 src is 10.1.1.1 src is 10.1.2.3 src is 10.2.3.4 # matches 10.2.3.4
ipv4 src is 10.1.1.1 src prefix 24 src prefix 16 # matches 10.1.0.0/16
Pattern items

This section lists supported pattern items and their attributes, if any.

  • end: end list of pattern items.
  • void: no-op pattern item.
  • invert: perform actions when pattern does not match.
  • any: match any protocol for the current layer.
    • num {unsigned}: number of layers covered.
  • pf: match packets addressed to the physical function.
  • vf: match packets addressed to a virtual function ID.
    • id {unsigned}: destination VF ID.
  • port: device-specific physical port index to use.
    • index {unsigned}: physical port index.
  • raw: match an arbitrary byte string.
    • relative {boolean}: look for pattern after the previous item.
    • search {boolean}: search pattern from offset (see also limit).
    • offset {integer}: absolute or relative offset for pattern.
    • limit {unsigned}: search area limit for start of pattern.
    • pattern {string}: byte string to look for.
  • eth: match Ethernet header.
    • dst {MAC-48}: destination MAC.
    • src {MAC-48}: source MAC.
    • type {unsigned}: EtherType.
  • vlan: match 802.1Q/ad VLAN tag.
    • tpid {unsigned}: tag protocol identifier.
    • tci {unsigned}: tag control information.
    • pcp {unsigned}: priority code point.
    • dei {unsigned}: drop eligible indicator.
    • vid {unsigned}: VLAN identifier.
  • ipv4: match IPv4 header.
    • tos {unsigned}: type of service.
    • ttl {unsigned}: time to live.
    • proto {unsigned}: next protocol ID.
    • src {ipv4 address}: source address.
    • dst {ipv4 address}: destination address.
  • ipv6: match IPv6 header.
    • tc {unsigned}: traffic class.
    • flow {unsigned}: flow label.
    • proto {unsigned}: protocol (next header).
    • hop {unsigned}: hop limit.
    • src {ipv6 address}: source address.
    • dst {ipv6 address}: destination address.
  • icmp: match ICMP header.
    • type {unsigned}: ICMP packet type.
    • code {unsigned}: ICMP packet code.
  • udp: match UDP header.
    • src {unsigned}: UDP source port.
    • dst {unsigned}: UDP destination port.
  • tcp: match TCP header.
    • src {unsigned}: TCP source port.
    • dst {unsigned}: TCP destination port.
  • sctp: match SCTP header.
    • src {unsigned}: SCTP source port.
    • dst {unsigned}: SCTP destination port.
    • tag {unsigned}: validation tag.
    • cksum {unsigned}: checksum.
  • vxlan: match VXLAN header.
    • vni {unsigned}: VXLAN identifier.
  • e_tag: match IEEE 802.1BR E-Tag header.
    • grp_ecid_b {unsigned}: GRP and E-CID base.
  • nvgre: match NVGRE header.
    • tni {unsigned}: virtual subnet ID.
  • mpls: match MPLS header.
    • label {unsigned}: MPLS label.
  • gre: match GRE header.
    • protocol {unsigned}: protocol type.
  • fuzzy: fuzzy pattern match, expect faster than default.
    • thresh {unsigned}: accuracy threshold.
  • gtp, gtpc, gtpu: match GTPv1 header.
    • teid {unsigned}: tunnel endpoint identifier.
Actions list

A list of actions starts after the actions token in the same fashion as Matching pattern; actions are separated by / tokens and the list is terminated by a mandatory end action.

Actions are named after their type (RTE_FLOW_ACTION_TYPE_ from enum rte_flow_action_type).

Dropping all incoming UDPv4 packets can be expressed as follows:

testpmd> flow create 0 ingress pattern eth / ipv4 / udp / end
   actions drop / end

Several actions have configurable properties which must be specified when there is no valid default value. For example, queue requires a target queue index.

This rule redirects incoming UDPv4 traffic to queue index 6:

testpmd> flow create 0 ingress pattern eth / ipv4 / udp / end
   actions queue index 6 / end

While this one could be rejected by PMDs (unspecified queue index):

testpmd> flow create 0 ingress pattern eth / ipv4 / udp / end
   actions queue / end

As defined by rte_flow, the list is not ordered, all actions of a given rule are performed simultaneously. These are equivalent:

queue index 6 / void / mark id 42 / end
void / mark id 42 / queue index 6 / end

All actions in a list should have different types, otherwise only the last action of a given type is taken into account:

queue index 4 / queue index 5 / queue index 6 / end # will use queue 6
drop / drop / drop / end # drop is performed only once
mark id 42 / queue index 3 / mark id 24 / end # mark will be 24

Considering they are performed simultaneously, opposite and overlapping actions can sometimes be combined when the end result is unambiguous:

drop / queue index 6 / end # drop has no effect
drop / dup index 6 / end # same as above
queue index 6 / rss queues 6 7 8 / end # queue has no effect
drop / passthru / end # drop has no effect

Note that PMDs may still refuse such combinations.

Actions

This section lists supported actions and their attributes, if any.

  • end: end list of actions.
  • void: no-op action.
  • passthru: let subsequent rule process matched packets.
  • mark: attach 32 bit value to packets.
    • id {unsigned}: 32 bit value to return with packets.
  • flag: flag packets.
  • queue: assign packets to a given queue index.
    • index {unsigned}: queue index to use.
  • drop: drop packets (note: passthru has priority).
  • count: enable counters for this rule.
  • dup: duplicate packets to a given queue index.
    • index {unsigned}: queue index to duplicate packets to.
  • rss: spread packets among several queues.
    • queues [{unsigned} [...]] end: queue indices to use.
  • pf: redirect packets to physical device function.
  • vf: redirect packets to virtual device function.
    • original {boolean}: use original VF ID if possible.
    • id {unsigned}: VF ID to redirect packets to.
Destroying flow rules

flow destroy destroys one or more rules from their rule ID (as returned by flow create), this command calls rte_flow_destroy() as many times as necessary:

flow destroy {port_id} rule {rule_id} [...]

If successful, it will show:

Flow rule #[...] destroyed

It does not report anything for rule IDs that do not exist. The usual error message is shown when a rule cannot be destroyed:

Caught error type [...] ([...]): [...]

flow flush destroys all rules on a device and does not take extra arguments. It is bound to rte_flow_flush():

flow flush {port_id}

Any errors are reported as above.

Creating several rules and destroying them:

testpmd> flow create 0 ingress pattern eth / ipv6 / end
   actions queue index 2 / end
Flow rule #0 created
testpmd> flow create 0 ingress pattern eth / ipv4 / end
   actions queue index 3 / end
Flow rule #1 created
testpmd> flow destroy 0 rule 0 rule 1
Flow rule #1 destroyed
Flow rule #0 destroyed
testpmd>

The same result can be achieved using flow flush:

testpmd> flow create 0 ingress pattern eth / ipv6 / end
   actions queue index 2 / end
Flow rule #0 created
testpmd> flow create 0 ingress pattern eth / ipv4 / end
   actions queue index 3 / end
Flow rule #1 created
testpmd> flow flush 0
testpmd>

Non-existent rule IDs are ignored:

testpmd> flow create 0 ingress pattern eth / ipv6 / end
   actions queue index 2 / end
Flow rule #0 created
testpmd> flow create 0 ingress pattern eth / ipv4 / end
   actions queue index 3 / end
Flow rule #1 created
testpmd> flow destroy 0 rule 42 rule 10 rule 2
testpmd>
testpmd> flow destroy 0 rule 0
Flow rule #0 destroyed
testpmd>
Querying flow rules

flow query queries a specific action of a flow rule having that ability. Such actions collect information that can be reported using this command. It is bound to rte_flow_query():

flow query {port_id} {rule_id} {action}

If successful, it will display either the retrieved data for known actions or the following message:

Cannot display result for action type [...] ([...])

Otherwise, it will complain either that the rule does not exist or that some error occurred:

Flow rule #[...] not found
Caught error type [...] ([...]): [...]

Currently only the count action is supported. This action reports the number of packets that hit the flow rule and the total number of bytes. Its output has the following format:

count:
 hits_set: [...] # whether "hits" contains a valid value
 bytes_set: [...] # whether "bytes" contains a valid value
 hits: [...] # number of packets
 bytes: [...] # number of bytes

Querying counters for TCPv6 packets redirected to queue 6:

testpmd> flow create 0 ingress pattern eth / ipv6 / tcp / end
   actions queue index 6 / count / end
Flow rule #4 created
testpmd> flow query 0 4 count
count:
 hits_set: 1
 bytes_set: 0
 hits: 386446
 bytes: 0
testpmd>
Listing flow rules

flow list lists existing flow rules sorted by priority and optionally filtered by group identifiers:

flow list {port_id} [group {group_id}] [...]

This command only fails with the following message if the device does not exist:

Invalid port [...]

Output consists of a header line followed by a short description of each flow rule, one per line. There is no output at all when no flow rules are configured on the device:

ID      Group   Prio    Attr    Rule
[...]   [...]   [...]   [...]   [...]

Attr column flags:

  • i for ingress.
  • e for egress.

Creating several flow rules and listing them:

testpmd> flow create 0 ingress pattern eth / ipv4 / end
   actions queue index 6 / end
Flow rule #0 created
testpmd> flow create 0 ingress pattern eth / ipv6 / end
   actions queue index 2 / end
Flow rule #1 created
testpmd> flow create 0 priority 5 ingress pattern eth / ipv4 / udp / end
   actions rss queues 6 7 8 end / end
Flow rule #2 created
testpmd> flow list 0
ID      Group   Prio    Attr    Rule
0       0       0       i-      ETH IPV4 => QUEUE
1       0       0       i-      ETH IPV6 => QUEUE
2       0       5       i-      ETH IPV4 UDP => RSS
testpmd>

Rules are sorted by priority (i.e. group ID first, then priority level):

testpmd> flow list 1
ID      Group   Prio    Attr    Rule
0       0       0       i-      ETH => COUNT
6       0       500     i-      ETH IPV6 TCP => DROP COUNT
5       0       1000    i-      ETH IPV6 ICMP => QUEUE
1       24      0       i-      ETH IPV4 UDP => QUEUE
4       24      10      i-      ETH IPV4 TCP => DROP
3       24      20      i-      ETH IPV4 => DROP
2       24      42      i-      ETH IPV4 UDP => QUEUE
7       63      0       i-      ETH IPV6 UDP VXLAN => MARK QUEUE
testpmd>

Output can be limited to specific groups:

testpmd> flow list 1 group 0 group 63
ID      Group   Prio    Attr    Rule
0       0       0       i-      ETH => COUNT
6       0       500     i-      ETH IPV6 TCP => DROP COUNT
5       0       1000    i-      ETH IPV6 ICMP => QUEUE
7       63      0       i-      ETH IPV6 UDP VXLAN => MARK QUEUE
testpmd>
Toggling isolated mode

flow isolate can be used to tell the underlying PMD that ingress traffic must only be injected from the defined flow rules; that no default traffic is expected outside those rules and the driver is free to assign more resources to handle them. It is bound to rte_flow_isolate():

flow isolate {port_id} {boolean}

If successful, enabling or disabling isolated mode shows either:

Ingress traffic on port [...]
   is now restricted to the defined flow rules

Or:

Ingress traffic on port [...]
   is not restricted anymore to the defined flow rules

Otherwise, in case of error:

Caught error type [...] ([...]): [...]

Mainly due to its side effects, PMDs supporting this mode may not have the ability to toggle it more than once without reinitializing affected ports first (e.g. by exiting testpmd).

Enabling isolated mode:

testpmd> flow isolate 0 true
Ingress traffic on port 0 is now restricted to the defined flow rules
testpmd>

Disabling isolated mode:

testpmd> flow isolate 0 false
Ingress traffic on port 0 is not restricted anymore to the defined flow rules
testpmd>
Sample QinQ flow rules

Before creating QinQ rule(s) the following commands should be issued to enable QinQ:

testpmd> port stop 0
testpmd> vlan set qinq on 0

The above command sets the inner and outer TPID’s to 0x8100.

To change the TPID’s the following commands should be used:

testpmd> vlan set outer tpid 0xa100 0
testpmd> vlan set inner tpid 0x9100 0
testpmd> port start 0

Validate and create a QinQ rule on port 0 to steer traffic to a VF queue in a VM.

testpmd> flow validate 0 ingress pattern eth / vlan tci is 123 /
    vlan tci is 456 / end actions vf id 1 / queue index 0 / end
Flow rule #0 validated

testpmd> flow create 0 ingress pattern eth / vlan tci is 4 /
    vlan tci is 456 / end actions vf id 123 / queue index 0 / end
Flow rule #0 created

testpmd> flow list 0
ID      Group   Prio    Attr    Rule
0       0       0       i-      ETH VLAN VLAN=>VF QUEUE

Validate and create a QinQ rule on port 0 to steer traffic to a queue on the host.

testpmd> flow validate 0 ingress pattern eth / vlan tci is 321 /
     vlan tci is 654 / end actions pf / queue index 0 / end
Flow rule #1 validated

testpmd> flow create 0 ingress pattern eth / vlan tci is 321 /
     vlan tci is 654 / end actions pf / queue index 1 / end
Flow rule #1 created

testpmd> flow list 0
ID      Group   Prio    Attr    Rule
0       0       0       i-      ETH VLAN VLAN=>VF QUEUE
1       0       0       i-      ETH VLAN VLAN=>PF QUEUE

Network Interface Controller Drivers

Overview of Networking Drivers

The networking drivers may be classified in two categories:

  • physical for real devices
  • virtual for emulated devices

Some physical devices may be shaped through a virtual layer as for SR-IOV. The interface seen in the virtual environment is a VF (Virtual Function).

The ethdev layer exposes an API to use the networking functions of these devices. The bottom half part of ethdev is implemented by the drivers. Thus some features may not be implemented.

There are more differences between drivers regarding some internal properties, portability or even documentation availability. Most of these differences are summarized below.

More details about features can be found in Features Overview.

Features availability in networking drivers
Feature a f p a c k e t a r k a v p b n x 2 x b n x 2 x v f b n x t b o n d i n g c x g b e d p a a d p a a 2 e 1 0 0 0 e n a e n i c f a i l s a f e f m 1 0 k f m 1 0 k v f i 4 0 e i 4 0 e . . . v e c i 4 0 e v f i 4 0 e v f . v e c i g b i g b v f i x g b e i x g b e . . v e c i x g b e v f i x g b e v f v e c k n i l i q u i d i o m l x 4 m l x 5 m r v l n f p _ p f n f p v f n u l l o c t e o n t x p c a p q e d e q e d e v f r i n g s f c _ e f x s z e d a t a 2 t a p t h u n d e r x v h o s t v i r t i o v i r t i o . v e c v m x n e t 3
Speed capabilities   Y   P P     Y P P P P     P P Y Y     P   Y Y       Y P Y Y Y Y   Y   Y Y   Y P P Y   P P P
Link status     Y Y Y Y   Y Y Y Y   Y Y     Y Y     Y   Y Y Y Y   Y Y Y Y Y Y   Y   Y Y   Y Y Y Y Y Y Y Y
Link status event       Y Y         Y Y   Y Y     Y Y     Y   Y Y       Y Y Y   Y Y       Y Y   Y   Y Y Y     Y
Removal event                                                         Y Y                                  
Queue status event                                                                                       Y      
Rx interrupt           Y         Y       Y Y Y Y Y Y Y Y Y Y Y Y     Y Y   Y Y                       Y Y  
Lock-free Tx queue                                                                     Y                        
Fast mbuf free                                                                                              
Free Tx mbuf on demand                     Y                                                                 Y      
Queue start/stop   Y       Y   Y   Y   Y Y   Y Y Y Y Y Y     Y Y         Y Y   Y Y   P         Y Y   Y   Y Y Y
MTU update           Y   Y Y Y Y Y Y Y             Y   Y Y Y Y   Y Y Y Y Y Y       Y Y   Y   Y Y   Y   Y
Jumbo frame   Y Y     Y   Y Y Y Y Y Y Y Y Y Y Y Y Y Y   Y Y Y Y   Y Y Y Y Y Y   Y Y Y Y   Y   Y Y       Y
Scattered Rx   Y Y         Y Y   Y Y Y   Y Y Y Y Y Y Y Y Y Y Y Y   Y Y Y             Y Y   Y Y   Y   Y    
LRO           Y                                 Y Y Y Y                     Y Y                 Y
TSO           Y   Y       Y Y   Y Y Y Y Y Y Y Y Y Y Y Y       Y   Y Y       Y Y   Y             Y
Promiscuous mode     Y Y Y Y   Y Y Y Y   Y Y Y   Y Y Y Y Y Y Y Y       Y Y Y Y Y Y   Y   Y Y   Y Y Y Y   Y Y Y
Allmulticast mode           Y   Y Y Y Y     Y Y   Y Y Y Y Y Y Y Y Y Y   Y Y Y Y           Y Y   Y Y Y Y   Y Y Y
Unicast MAC filter     Y Y Y Y     Y Y Y   Y Y Y   Y Y Y Y Y Y Y Y Y Y     Y Y Y       P   Y Y       Y     Y Y Y
Multicast MAC filter       Y Y Y             Y Y Y   Y Y Y Y                 Y Y Y           Y Y   Y   Y     Y Y  
RSS hash           Y   Y Y Y   Y Y   Y Y Y Y Y Y Y   Y Y Y Y   Y Y Y Y Y Y       Y Y   Y     Y       Y
RSS key update           Y           Y     Y Y Y Y Y Y Y   Y Y Y Y   Y   Y   Y Y       Y Y   Y     Y        
RSS reta update           Y           Y     Y Y Y Y Y Y Y   Y Y Y Y   Y   Y   Y Y       Y Y   Y     Y        
VMDq           Y                 Y   Y Y     Y   Y Y                                              
SR-IOV         Y Y           Y Y       Y Y     Y   Y Y         Y Y               Y   Y     Y        
DCB                                 Y Y     Y   Y Y                                              
VLAN filter           Y       Y Y   Y Y Y   Y Y Y Y Y Y Y Y Y Y   Y Y Y Y           Y Y             Y Y Y
Ethertype filter           Y                     Y Y     Y   Y Y                                              
N-tuple filter           Y                             Y   Y Y                         Y                    
SYN filter                                         Y   Y Y                                              
Tunnel filter                                 Y Y         Y Y                         Y                    
Flexible filter                                         Y                                                    
Hash filter                                 Y Y Y Y                                                      
Flow director                         Y       Y Y         Y Y           Y             Y                    
Flow control           Y   Y Y Y Y     Y     Y Y     Y   Y Y               Y Y       Y Y   Y   Y          
Flow API           Y             Y Y     Y       Y   Y             Y                   Y   Y          
Rate limitation                                             Y Y                                              
Traffic mirroring                                 Y Y         Y Y                                              
Inline crypto                                             Y Y Y Y                                          
CRC offload               Y     Y Y Y   Y Y Y   Y   Y Y Y   Y     Y   Y Y       Y   Y Y         Y        
VLAN offload     Y     Y   Y   Y Y Y Y   Y Y Y   Y   Y Y Y   Y     P   Y   Y Y       Y Y   P     P       Y
QinQ offload                     Y           Y   Y   Y Y Y   Y                                            
L3 checksum offload           Y   Y Y Y Y Y Y   Y Y Y   Y   Y Y Y   Y     Y Y Y Y Y Y       Y Y   Y   Y Y        
L4 checksum offload           Y   Y Y Y Y Y Y   Y Y Y   Y   Y Y Y   Y     Y Y Y Y Y Y       Y Y   Y   Y Y       Y
Timestamp offload                                                                                              
MACsec offload                                             Y                                                
Inner L3 checksum           Y           Y         Y   Y       Y   Y     Y Y               Y                    
Inner L4 checksum           Y           Y         Y   Y       Y   Y     Y Y               Y                    
Packet type parsing           Y   Y Y Y     Y Y Y Y Y   Y   Y Y Y   Y       Y Y Y       Y   Y Y   Y   Y Y       Y
Timesync                                 Y Y     Y   Y Y                                              
Rx descriptor status           Y         Y           Y Y Y Y Y Y Y Y Y Y       Y                   Y              
Tx descriptor status           Y         Y           Y Y Y Y Y Y Y Y Y Y       Y                   Y              
Basic stats   Y Y Y Y Y   Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y   Y Y Y Y Y Y   Y Y Y Y   Y Y Y Y Y Y Y Y
Extended stats       Y Y Y     Y Y   Y     Y Y Y Y Y Y Y Y Y Y Y Y   Y   Y             Y Y   Y Y     Y Y   Y
Stats per queue   Y Y         Y           Y Y Y             Y Y         Y Y   Y Y       Y Y     Y   Y   Y Y Y
FW version           Y     Y Y             Y       Y   Y                                 Y              
EEPROM dump           Y   Y                         Y   Y Y                                              
Registers dump               Y                         Y Y Y Y Y Y                                 Y        
LED           Y                                                                                  
Multiprocess aware                         Y   Y Y Y Y Y Y     Y Y Y Y   Y   Y           Y Y Y   Y     Y   Y    
BSD nic_uio               Y     Y   Y   Y Y Y Y Y Y Y Y Y Y Y Y                           Y         Y Y  
Linux UIO   Y Y Y Y Y   Y     Y Y Y   Y Y Y Y Y Y Y Y Y Y Y Y   Y       Y Y       Y Y   Y         Y Y Y
Linux VFIO           Y   Y   Y Y   Y   Y Y Y Y Y Y Y Y Y Y Y Y   Y       Y Y   Y         Y     Y   Y Y Y
Other kdrv                                                         Y Y                     Y Y          
ARMv7                           Y                                           Y           Y     Y Y  
ARMv8                 Y Y       Y     Y Y   Y     Y Y Y Y       Y Y       Y Y Y Y       Y Y   Y Y  
Power8                           Y     Y Y                     Y Y           Y           Y          
x86-32               Y     Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y     Y             Y Y Y       Y   Y Y Y Y
x86-64   Y Y Y Y Y   Y     Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y   Y Y Y   Y Y     Y Y Y   Y Y Y   Y Y Y Y
Usage doc   Y   Y Y     Y Y Y     Y Y                         Y Y Y Y Y Y Y   Y Y Y Y     Y Y Y   Y   Y
Design doc                                                                                              
Perf doc                                                                                              

Note

Features marked with “P” are partially supported. Refer to the appropriate NIC guide in the following sections for details.

Features Overview

This section explains the supported features that are listed in the Overview of Networking Drivers.

As a guide to implementers it also shows the structs where the features are defined and the APIs that can be use to get/set the values.

Following tags used for feature details, these are from driver point of view:

[uses] : Driver uses some kind of input from the application.

[implements] : Driver implements a functionality.

[provides] : Driver provides some kind of data to the application. It is possible to provide data by implementing some function, but “provides” is used for cases where provided data can’t be represented simply by a function.

[related] : Related API with that feature.

Speed capabilities

Supports getting the speed capabilities that the current device is capable of.

  • [provides] rte_eth_dev_info: speed_capa:ETH_LINK_SPEED_*.
  • [related] API: rte_eth_dev_info_get().

Removal event

Supports device removal interrupts.

  • [uses] user config: dev_conf.intr_conf.rmv.
  • [uses] rte_eth_dev_data: dev_flags:RTE_ETH_DEV_INTR_RMV.
  • [uses] rte_eth_event_type: RTE_ETH_EVENT_INTR_RMV.
  • [provides] rte_pci_driver.drv_flags: RTE_PCI_DRV_INTR_RMV.

Queue status event

Supports queue enable/disable events.

  • [uses] rte_eth_event_type: RTE_ETH_EVENT_QUEUE_STATE.

Rx interrupt

Supports Rx interrupts.

  • [uses] user config: dev_conf.intr_conf.rxq.
  • [implements] eth_dev_ops: rx_queue_intr_enable, rx_queue_intr_disable.
  • [related] API: rte_eth_dev_rx_intr_enable(), rte_eth_dev_rx_intr_disable().

Lock-free Tx queue

If a PMD advertises DEV_TX_OFFLOAD_MT_LOCKFREE capable, multiple threads can invoke rte_eth_tx_burst() concurrently on the same Tx queue without SW lock.

  • [uses] rte_eth_txconf,rte_eth_txmode: offloads:DEV_TX_OFFLOAD_MT_LOCKFREE.
  • [provides] rte_eth_dev_info: tx_offload_capa,tx_queue_offload_capa:DEV_TX_OFFLOAD_MT_LOCKFREE.
  • [related] API: rte_eth_tx_burst().

Fast mbuf free

Supports optimization for fast release of mbufs following successful Tx. Requires that per queue, all mbufs come from the same mempool and has refcnt = 1.

  • [uses] rte_eth_txconf,rte_eth_txmode: offloads:DEV_TX_OFFLOAD_MBUF_FAST_FREE.
  • [provides] rte_eth_dev_info: tx_offload_capa,tx_queue_offload_capa:DEV_TX_OFFLOAD_MBUF_FAST_FREE.

Free Tx mbuf on demand

Supports freeing consumed buffers on a Tx ring.

  • [implements] eth_dev_ops: tx_done_cleanup.
  • [related] API: rte_eth_tx_done_cleanup().

Queue start/stop

Supports starting/stopping a specific Rx/Tx queue of a port.

  • [implements] eth_dev_ops: rx_queue_start, rx_queue_stop, tx_queue_start, tx_queue_stop.
  • [related] API: rte_eth_dev_rx_queue_start(), rte_eth_dev_rx_queue_stop(), rte_eth_dev_tx_queue_start(), rte_eth_dev_tx_queue_stop().

MTU update

Supports updating port MTU.

  • [implements] eth_dev_ops: mtu_set.
  • [implements] rte_eth_dev_data: mtu.
  • [provides] rte_eth_dev_info: max_rx_pktlen.
  • [related] API: rte_eth_dev_set_mtu(), rte_eth_dev_get_mtu().

Jumbo frame

Supports Rx jumbo frames.

  • [uses] rte_eth_rxconf,rte_eth_rxmode: offloads:DEV_RX_OFFLOAD_JUMBO_FRAME. dev_conf.rxmode.max_rx_pkt_len.
  • [related] rte_eth_dev_info: max_rx_pktlen.
  • [related] API: rte_eth_dev_set_mtu().

Scattered Rx

Supports receiving segmented mbufs.

  • [uses] rte_eth_rxconf,rte_eth_rxmode: offloads:DEV_RX_OFFLOAD_SCATTER.
  • [implements] datapath: Scattered Rx function.
  • [implements] rte_eth_dev_data: scattered_rx.
  • [provides] eth_dev_ops: rxq_info_get:scattered_rx.
  • [related] eth_dev_ops: rx_pkt_burst.

LRO

Supports Large Receive Offload.

  • [uses] rte_eth_rxconf,rte_eth_rxmode: offloads:DEV_RX_OFFLOAD_TCP_LRO.
  • [implements] datapath: LRO functionality.
  • [implements] rte_eth_dev_data: lro.
  • [provides] mbuf: mbuf.ol_flags:PKT_RX_LRO, mbuf.tso_segsz.
  • [provides] rte_eth_dev_info: rx_offload_capa,rx_queue_offload_capa:DEV_RX_OFFLOAD_TCP_LRO.

TSO

Supports TCP Segmentation Offloading.

  • [uses] rte_eth_txconf,rte_eth_txmode: offloads:DEV_TX_OFFLOAD_TCP_TSO.
  • [uses] rte_eth_desc_lim: nb_seg_max, nb_mtu_seg_max.
  • [uses] mbuf: mbuf.ol_flags:PKT_TX_TCP_SEG.
  • [uses] mbuf: mbuf.tso_segsz, mbuf.l2_len, mbuf.l3_len, mbuf.l4_len.
  • [implements] datapath: TSO functionality.
  • [provides] rte_eth_dev_info: tx_offload_capa,tx_queue_offload_capa:DEV_TX_OFFLOAD_TCP_TSO,DEV_TX_OFFLOAD_UDP_TSO.

Promiscuous mode

Supports enabling/disabling promiscuous mode for a port.

  • [implements] eth_dev_ops: promiscuous_enable, promiscuous_disable.
  • [implements] rte_eth_dev_data: promiscuous.
  • [related] API: rte_eth_promiscuous_enable(), rte_eth_promiscuous_disable(), rte_eth_promiscuous_get().

Allmulticast mode

Supports enabling/disabling receiving multicast frames.

  • [implements] eth_dev_ops: allmulticast_enable, allmulticast_disable.
  • [implements] rte_eth_dev_data: all_multicast.
  • [related] API: rte_eth_allmulticast_enable(), rte_eth_allmulticast_disable(), rte_eth_allmulticast_get().

Unicast MAC filter

Supports adding MAC addresses to enable whitelist filtering to accept packets.

  • [implements] eth_dev_ops: mac_addr_set, mac_addr_add, mac_addr_remove.
  • [implements] rte_eth_dev_data: mac_addrs.
  • [related] API: rte_eth_dev_default_mac_addr_set(), rte_eth_dev_mac_addr_add(), rte_eth_dev_mac_addr_remove(), rte_eth_macaddr_get().

Multicast MAC filter

Supports setting multicast addresses to filter.

  • [implements] eth_dev_ops: set_mc_addr_list.
  • [related] API: rte_eth_dev_set_mc_addr_list().

RSS hash

Supports RSS hashing on RX.

  • [uses] user config: dev_conf.rxmode.mq_mode = ETH_MQ_RX_RSS_FLAG.
  • [uses] user config: dev_conf.rx_adv_conf.rss_conf.
  • [provides] rte_eth_dev_info: flow_type_rss_offloads.
  • [provides] mbuf: mbuf.ol_flags:PKT_RX_RSS_HASH, mbuf.rss.

RSS key update

Supports configuration of Receive Side Scaling (RSS) hash computation. Updating Receive Side Scaling (RSS) hash key.

  • [implements] eth_dev_ops: rss_hash_update, rss_hash_conf_get.
  • [provides] rte_eth_dev_info: hash_key_size.
  • [related] API: rte_eth_dev_rss_hash_update(), rte_eth_dev_rss_hash_conf_get().

RSS reta update

Supports updating Redirection Table of the Receive Side Scaling (RSS).

  • [implements] eth_dev_ops: reta_update, reta_query.
  • [provides] rte_eth_dev_info: reta_size.
  • [related] API: rte_eth_dev_rss_reta_update(), rte_eth_dev_rss_reta_query().

VMDq

Supports Virtual Machine Device Queues (VMDq).

  • [uses] user config: dev_conf.rxmode.mq_mode = ETH_MQ_RX_VMDQ_FLAG.
  • [uses] user config: dev_conf.rx_adv_conf.vmdq_dcb_conf.
  • [uses] user config: dev_conf.rx_adv_conf.vmdq_rx_conf.
  • [uses] user config: dev_conf.tx_adv_conf.vmdq_dcb_tx_conf.
  • [uses] user config: dev_conf.tx_adv_conf.vmdq_tx_conf.

SR-IOV

Driver supports creating Virtual Functions.

  • [implements] rte_eth_dev_data: sriov.

DCB

Supports Data Center Bridging (DCB).

  • [uses] user config: dev_conf.rxmode.mq_mode = ETH_MQ_RX_DCB_FLAG.
  • [uses] user config: dev_conf.rx_adv_conf.vmdq_dcb_conf.
  • [uses] user config: dev_conf.rx_adv_conf.dcb_rx_conf.
  • [uses] user config: dev_conf.tx_adv_conf.vmdq_dcb_tx_conf.
  • [uses] user config: dev_conf.tx_adv_conf.vmdq_tx_conf.
  • [implements] eth_dev_ops: get_dcb_info.
  • [related] API: rte_eth_dev_get_dcb_info().

VLAN filter

Supports filtering of a VLAN Tag identifier.

  • [uses] rte_eth_rxconf,rte_eth_rxmode: offloads:DEV_RX_OFFLOAD_VLAN_FILTER.
  • [implements] eth_dev_ops: vlan_filter_set.
  • [related] API: rte_eth_dev_vlan_filter().

Ethertype filter

Supports filtering on Ethernet type.

  • [implements] eth_dev_ops: filter_ctrl:RTE_ETH_FILTER_ETHERTYPE.
  • [related] API: rte_eth_dev_filter_ctrl(), rte_eth_dev_filter_supported().

N-tuple filter

Supports filtering on N-tuple values.

  • [implements] eth_dev_ops: filter_ctrl:RTE_ETH_FILTER_NTUPLE.
  • [related] API: rte_eth_dev_filter_ctrl(), rte_eth_dev_filter_supported().

SYN filter

Supports TCP syn filtering.

  • [implements] eth_dev_ops: filter_ctrl:RTE_ETH_FILTER_SYN.
  • [related] API: rte_eth_dev_filter_ctrl(), rte_eth_dev_filter_supported().

Tunnel filter

Supports tunnel filtering.

  • [implements] eth_dev_ops: filter_ctrl:RTE_ETH_FILTER_TUNNEL.
  • [related] API: rte_eth_dev_filter_ctrl(), rte_eth_dev_filter_supported().

Flexible filter

Supports a flexible (non-tuple or Ethertype) filter.

  • [implements] eth_dev_ops: filter_ctrl:RTE_ETH_FILTER_FLEXIBLE.
  • [related] API: rte_eth_dev_filter_ctrl(), rte_eth_dev_filter_supported().

Hash filter

Supports Hash filtering.

  • [implements] eth_dev_ops: filter_ctrl:RTE_ETH_FILTER_HASH.
  • [related] API: rte_eth_dev_filter_ctrl(), rte_eth_dev_filter_supported().

Flow director

Supports Flow Director style filtering to queues.

  • [implements] eth_dev_ops: filter_ctrl:RTE_ETH_FILTER_FDIR.
  • [provides] mbuf: mbuf.ol_flags: PKT_RX_FDIR, PKT_RX_FDIR_ID, PKT_RX_FDIR_FLX.
  • [related] API: rte_eth_dev_filter_ctrl(), rte_eth_dev_filter_supported().

Flow control

Supports configuring link flow control.

  • [implements] eth_dev_ops: flow_ctrl_get, flow_ctrl_set, priority_flow_ctrl_set.
  • [related] API: rte_eth_dev_flow_ctrl_get(), rte_eth_dev_flow_ctrl_set(), rte_eth_dev_priority_flow_ctrl_set().

Flow API

Supports the DPDK Flow API for generic filtering.

  • [implements] eth_dev_ops: filter_ctrl:RTE_ETH_FILTER_GENERIC.
  • [implements] rte_flow_ops: All.

Rate limitation

Supports Tx rate limitation for a queue.

  • [implements] eth_dev_ops: set_queue_rate_limit.
  • [related] API: rte_eth_set_queue_rate_limit().

Traffic mirroring

Supports adding traffic mirroring rules.

  • [implements] eth_dev_ops: mirror_rule_set, mirror_rule_reset.
  • [related] API: rte_eth_mirror_rule_set(), rte_eth_mirror_rule_reset().

Inline crypto

Supports inline crypto processing (eg. inline IPsec). See Security library and PMD documentation for more details.

  • [uses] rte_eth_rxconf,rte_eth_rxmode: offloads:DEV_RX_OFFLOAD_SECURITY,
  • [uses] rte_eth_txconf,rte_eth_txmode: offloads:DEV_TX_OFFLOAD_SECURITY.
  • [implements] rte_security_ops: session_create, session_update, session_stats_get, session_destroy, set_pkt_metadata, capabilities_get.
  • [provides] rte_eth_dev_info: rx_offload_capa,rx_queue_offload_capa:DEV_RX_OFFLOAD_SECURITY, tx_offload_capa,tx_queue_offload_capa:DEV_TX_OFFLOAD_SECURITY.
  • [provides] mbuf: mbuf.ol_flags:PKT_RX_SEC_OFFLOAD, mbuf.ol_flags:PKT_TX_SEC_OFFLOAD, mbuf.ol_flags:PKT_RX_SEC_OFFLOAD_FAILED.

CRC offload

Supports CRC stripping by hardware.

  • [uses] rte_eth_rxconf,rte_eth_rxmode: offloads:DEV_RX_OFFLOAD_CRC_STRIP.

VLAN offload

Supports VLAN offload to hardware.

  • [uses] rte_eth_rxconf,rte_eth_rxmode: offloads:DEV_RX_OFFLOAD_VLAN_STRIP,DEV_RX_OFFLOAD_VLAN_FILTER,DEV_RX_OFFLOAD_VLAN_EXTEND.
  • [uses] rte_eth_txconf,rte_eth_txmode: offloads:DEV_TX_OFFLOAD_VLAN_INSERT.
  • [implements] eth_dev_ops: vlan_offload_set.
  • [provides] mbuf: mbuf.ol_flags:PKT_RX_VLAN_STRIPPED, mbuf.vlan_tci.
  • [provides] rte_eth_dev_info: rx_offload_capa,rx_queue_offload_capa:DEV_RX_OFFLOAD_VLAN_STRIP, tx_offload_capa,tx_queue_offload_capa:DEV_TX_OFFLOAD_VLAN_INSERT.
  • [related] API: rte_eth_dev_set_vlan_offload(), rte_eth_dev_get_vlan_offload().

QinQ offload

Supports QinQ (queue in queue) offload.

  • [uses] rte_eth_rxconf,rte_eth_rxmode: offloads:DEV_RX_OFFLOAD_QINQ_STRIP.
  • [uses] rte_eth_txconf,rte_eth_txmode: offloads:DEV_TX_OFFLOAD_QINQ_INSERT.
  • [uses] mbuf: mbuf.ol_flags:PKT_TX_QINQ_PKT.
  • [provides] mbuf: mbuf.ol_flags:PKT_RX_QINQ_STRIPPED, mbuf.vlan_tci,
    mbuf.vlan_tci_outer.
  • [provides] rte_eth_dev_info: rx_offload_capa,rx_queue_offload_capa:DEV_RX_OFFLOAD_QINQ_STRIP, tx_offload_capa,tx_queue_offload_capa:DEV_TX_OFFLOAD_QINQ_INSERT.

L3 checksum offload

Supports L3 checksum offload.

  • [uses] rte_eth_rxconf,rte_eth_rxmode: offloads:DEV_RX_OFFLOAD_IPV4_CKSUM.
  • [uses] rte_eth_txconf,rte_eth_txmode: offloads:DEV_TX_OFFLOAD_IPV4_CKSUM.
  • [uses] mbuf: mbuf.ol_flags:PKT_TX_IP_CKSUM, mbuf.ol_flags:PKT_TX_IPV4 | PKT_TX_IPV6.
  • [provides] mbuf: mbuf.ol_flags:PKT_RX_IP_CKSUM_UNKNOWN | PKT_RX_IP_CKSUM_BAD | PKT_RX_IP_CKSUM_GOOD | PKT_RX_IP_CKSUM_NONE.
  • [provides] rte_eth_dev_info: rx_offload_capa,rx_queue_offload_capa:DEV_RX_OFFLOAD_IPV4_CKSUM, tx_offload_capa,tx_queue_offload_capa:DEV_TX_OFFLOAD_IPV4_CKSUM.

L4 checksum offload

Supports L4 checksum offload.

  • [uses] rte_eth_rxconf,rte_eth_rxmode: offloads:DEV_RX_OFFLOAD_UDP_CKSUM,DEV_RX_OFFLOAD_TCP_CKSUM.
  • [uses] rte_eth_txconf,rte_eth_txmode: offloads:DEV_TX_OFFLOAD_UDP_CKSUM,DEV_TX_OFFLOAD_TCP_CKSUM,DEV_TX_OFFLOAD_SCTP_CKSUM.
  • [uses] user config: dev_conf.rxmode.hw_ip_checksum.
  • [uses] mbuf: mbuf.ol_flags:PKT_TX_IPV4 | PKT_TX_IPV6, mbuf.ol_flags:PKT_TX_L4_NO_CKSUM | PKT_TX_TCP_CKSUM | PKT_TX_SCTP_CKSUM | PKT_TX_UDP_CKSUM.
  • [provides] mbuf: mbuf.ol_flags:PKT_RX_L4_CKSUM_UNKNOWN | PKT_RX_L4_CKSUM_BAD | PKT_RX_L4_CKSUM_GOOD | PKT_RX_L4_CKSUM_NONE.
  • [provides] rte_eth_dev_info: rx_offload_capa,rx_queue_offload_capa:DEV_RX_OFFLOAD_UDP_CKSUM,DEV_RX_OFFLOAD_TCP_CKSUM, tx_offload_capa,tx_queue_offload_capa:DEV_TX_OFFLOAD_UDP_CKSUM,DEV_TX_OFFLOAD_TCP_CKSUM,DEV_TX_OFFLOAD_SCTP_CKSUM.

Timestamp offload

Supports Timestamp.

  • [uses] rte_eth_rxconf,rte_eth_rxmode: offloads:DEV_RX_OFFLOAD_TIMESTAMP.
  • [provides] mbuf: mbuf.ol_flags:PKT_RX_TIMESTAMP.
  • [provides] mbuf: mbuf.timestamp.
  • [provides] rte_eth_dev_info: rx_offload_capa,rx_queue_offload_capa: DEV_RX_OFFLOAD_TIMESTAMP.

MACsec offload

Supports MACsec.

  • [uses] rte_eth_rxconf,rte_eth_rxmode: offloads:DEV_RX_OFFLOAD_MACSEC_STRIP.
  • [uses] rte_eth_txconf,rte_eth_txmode: offloads:DEV_TX_OFFLOAD_MACSEC_INSERT.
  • [uses] mbuf: mbuf.ol_flags:PKT_TX_MACSEC.
  • [provides] rte_eth_dev_info: rx_offload_capa,rx_queue_offload_capa:DEV_RX_OFFLOAD_MACSEC_STRIP, tx_offload_capa,tx_queue_offload_capa:DEV_TX_OFFLOAD_MACSEC_INSERT.

Inner L3 checksum

Supports inner packet L3 checksum.

  • [uses] rte_eth_rxconf,rte_eth_rxmode: offloads:DEV_RX_OFFLOAD_OUTER_IPV4_CKSUM.
  • [uses] rte_eth_txconf,rte_eth_txmode: offloads:DEV_TX_OFFLOAD_OUTER_IPV4_CKSUM.
  • [uses] mbuf: mbuf.ol_flags:PKT_TX_IP_CKSUM, mbuf.ol_flags:PKT_TX_IPV4 | PKT_TX_IPV6, mbuf.ol_flags:PKT_TX_OUTER_IP_CKSUM, mbuf.ol_flags:PKT_TX_OUTER_IPV4 | PKT_TX_OUTER_IPV6.
  • [uses] mbuf: mbuf.outer_l2_len, mbuf.outer_l3_len.
  • [provides] mbuf: mbuf.ol_flags:PKT_RX_EIP_CKSUM_BAD.
  • [provides] rte_eth_dev_info: rx_offload_capa,rx_queue_offload_capa:DEV_RX_OFFLOAD_OUTER_IPV4_CKSUM, tx_offload_capa,tx_queue_offload_capa:DEV_TX_OFFLOAD_OUTER_IPV4_CKSUM.

Inner L4 checksum

Supports inner packet L4 checksum.

Packet type parsing

Supports packet type parsing and returns a list of supported types.

  • [implements] eth_dev_ops: dev_supported_ptypes_get.
  • [related] API: rte_eth_dev_get_supported_ptypes().

Timesync

Supports IEEE1588/802.1AS timestamping.

  • [implements] eth_dev_ops: timesync_enable, timesync_disable timesync_read_rx_timestamp, timesync_read_tx_timestamp, timesync_adjust_time, timesync_read_time, timesync_write_time.
  • [related] API: rte_eth_timesync_enable(), rte_eth_timesync_disable(), rte_eth_timesync_read_rx_timestamp(), rte_eth_timesync_read_tx_timestamp, rte_eth_timesync_adjust_time(), rte_eth_timesync_read_time(), rte_eth_timesync_write_time().

Rx descriptor status

Supports check the status of a Rx descriptor. When rx_descriptor_status is used, status can be “Available”, “Done” or “Unavailable”. When rx_descriptor_done is used, status can be “DD bit is set” or “DD bit is not set”.

  • [implements] eth_dev_ops: rx_descriptor_status.
  • [related] API: rte_eth_rx_descriptor_status().
  • [implements] eth_dev_ops: rx_descriptor_done.
  • [related] API: rte_eth_rx_descriptor_done().

Tx descriptor status

Supports checking the status of a Tx descriptor. Status can be “Full”, “Done” or “Unavailable.”

  • [implements] eth_dev_ops: tx_descriptor_status.
  • [related] API: rte_eth_tx_descriptor_status().

Basic stats

Support basic statistics such as: ipackets, opackets, ibytes, obytes, imissed, ierrors, oerrors, rx_nombuf.

And per queue stats: q_ipackets, q_opackets, q_ibytes, q_obytes, q_errors.

These apply to all drivers.

  • [implements] eth_dev_ops: stats_get, stats_reset.
  • [related] API: rte_eth_stats_get, rte_eth_stats_reset().

Extended stats

Supports Extended Statistics, changes from driver to driver.

  • [implements] eth_dev_ops: xstats_get, xstats_reset, xstats_get_names.
  • [implements] eth_dev_ops: xstats_get_by_id, xstats_get_names_by_id.
  • [related] API: rte_eth_xstats_get(), rte_eth_xstats_reset(), rte_eth_xstats_get_names, rte_eth_xstats_get_by_id(), rte_eth_xstats_get_names_by_id(), rte_eth_xstats_get_id_by_name().

Stats per queue

Supports configuring per-queue stat counter mapping.

  • [implements] eth_dev_ops: queue_stats_mapping_set.
  • [related] API: rte_eth_dev_set_rx_queue_stats_mapping(), rte_eth_dev_set_tx_queue_stats_mapping().

FW version

Supports getting device hardware firmware information.

  • [implements] eth_dev_ops: fw_version_get.
  • [related] API: rte_eth_dev_fw_version_get().

EEPROM dump

Supports getting/setting device eeprom data.

  • [implements] eth_dev_ops: get_eeprom_length, get_eeprom, set_eeprom.
  • [related] API: rte_eth_dev_get_eeprom_length(), rte_eth_dev_get_eeprom(), rte_eth_dev_set_eeprom().

Registers dump

Supports retrieving device registers and registering attributes (number of registers and register size).

  • [implements] eth_dev_ops: get_reg.
  • [related] API: rte_eth_dev_get_reg_info().

LED

Supports turning on/off a software controllable LED on a device.

  • [implements] eth_dev_ops: dev_led_on, dev_led_off.
  • [related] API: rte_eth_led_on(), rte_eth_led_off().

Multiprocess aware

Driver can be used for primary-secondary process model.

BSD nic_uio

BSD nic_uio module supported.

Linux UIO

Works with igb_uio kernel module.

  • [provides] RTE_PMD_REGISTER_KMOD_DEP: igb_uio.

Linux VFIO

Works with vfio-pci kernel module.

  • [provides] RTE_PMD_REGISTER_KMOD_DEP: vfio-pci.

Other kdrv

Kernel module other than above ones supported.

ARMv7

Support armv7 architecture.

Use defconfig_arm-armv7a-*-*.

ARMv8

Support armv8a (64bit) architecture.

Use defconfig_arm64-armv8a-*-*

Power8

Support PowerPC architecture.

Use defconfig_ppc_64-power8-*-*

x86-32

Support 32bits x86 architecture.

Use defconfig_x86_x32-native-*-* and defconfig_i686-native-*-*.

x86-64

Support 64bits x86 architecture.

Use defconfig_x86_64-native-*-*.

Usage doc

Documentation describes usage.

See doc/guides/nics/*.rst

Design doc

Documentation describes design.

See doc/guides/nics/*.rst.

Perf doc

Documentation describes performance values.

See dpdk.org/doc/perf/*.

Other dev ops not represented by a Feature

  • rxq_info_get
  • txq_info_get
  • vlan_tpid_set
  • vlan_strip_queue_set
  • vlan_pvid_set
  • rx_queue_count
  • l2_tunnel_offload_set
  • uc_hash_table_set
  • uc_all_hash_table_set
  • udp_tunnel_port_add
  • udp_tunnel_port_del
  • l2_tunnel_eth_type_conf
  • l2_tunnel_offload_set
  • tx_pkt_prepare

Compiling and testing a PMD for a NIC

This section demonstrates how to compile and run a Poll Mode Driver (PMD) for the available Network Interface Cards in DPDK using TestPMD.

TestPMD is one of the reference applications distributed with the DPDK. Its main purpose is to forward packets between Ethernet ports on a network interface and as such is the best way to test a PMD.

Refer to the testpmd application user guide for detailed information on how to build and run testpmd.

Driver Compilation

To compile a PMD for a platform, run make with appropriate target as shown below. Use “make” command in Linux and “gmake” in FreeBSD. This will also build testpmd.

To check available targets:

cd <DPDK-source-directory>
make showconfigs

Example output:

arm-armv7a-linuxapp-gcc
arm64-armv8a-linuxapp-gcc
arm64-dpaa2-linuxapp-gcc
arm64-thunderx-linuxapp-gcc
arm64-xgene1-linuxapp-gcc
i686-native-linuxapp-gcc
i686-native-linuxapp-icc
ppc_64-power8-linuxapp-gcc
x86_64-native-bsdapp-clang
x86_64-native-bsdapp-gcc
x86_64-native-linuxapp-clang
x86_64-native-linuxapp-gcc
x86_64-native-linuxapp-icc
x86_x32-native-linuxapp-gcc

To compile a PMD for Linux x86_64 gcc target, run the following “make” command:

make install T=x86_64-native-linuxapp-gcc

Use ARM (ThunderX, DPAA, X-Gene) or PowerPC target for respective platform.

For more information, refer to the Getting Started Guide for Linux or Getting Started Guide for FreeBSD depending on your platform.

Running testpmd in Linux

This section demonstrates how to setup and run testpmd in Linux.

  1. Mount huge pages:

    mkdir /mnt/huge
    mount -t hugetlbfs nodev /mnt/huge
    
  2. Request huge pages:

    Hugepage memory should be reserved as per application requirement. Check hugepage size configured in the system and calculate the number of pages required.

    To reserve 1024 pages of 2MB:

    echo 1024 > /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
    

    Note

    Check /proc/meminfo to find system hugepage size:

    grep "Hugepagesize:" /proc/meminfo
    

    Example output:

    Hugepagesize:       2048 kB
    
  3. Load igb_uio or vfio-pci driver:

    modprobe uio
    insmod ./x86_64-native-linuxapp-gcc/kmod/igb_uio.ko
    

    or

    modprobe vfio-pci
    
  4. Setup VFIO permissions for regular users before binding to vfio-pci:

    sudo chmod a+x /dev/vfio
    
    sudo chmod 0666 /dev/vfio/*
    
  5. Bind the adapters to igb_uio or vfio-pci loaded in the previous step:

    ./usertools/dpdk-devbind.py --bind igb_uio DEVICE1 DEVICE2 ...
    

    Or setup VFIO permissions for regular users and then bind to vfio-pci:

    ./usertools/dpdk-devbind.py --bind vfio-pci DEVICE1 DEVICE2 ...
    

    Note

    DEVICE1, DEVICE2 are specified via PCI “domain:bus:slot.func” syntax or “bus:slot.func” syntax.

  6. Start testpmd with basic parameters:

    ./x86_64-native-linuxapp-gcc/app/testpmd -l 0-3 -n 4 -- -i
    

    Successful execution will show initialization messages from EAL, PMD and testpmd application. A prompt will be displayed at the end for user commands as interactive mode (-i) is on.

    testpmd>
    

    Refer to the testpmd runtime functions for a list of available commands.

ARK Poll Mode Driver

The ARK PMD is a DPDK poll-mode driver for the Atomic Rules Arkville (ARK) family of devices.

More information can be found at the Atomic Rules website.

Overview

The Atomic Rules Arkville product is DPDK and AXI compliant product that marshals packets across a PCIe conduit between host DPDK mbufs and FPGA AXI streams.

The ARK PMD, and the spirit of the overall Arkville product, has been to take the DPDK API/ABI as a fixed specification; then implement much of the business logic in FPGA RTL circuits. The approach of working backwards from the DPDK API/ABI and having the GPP host software dictate, while the FPGA hardware copes, results in significant performance gains over a naive implementation.

While this document describes the ARK PMD software, it is helpful to understand what the FPGA hardware is and is not. The Arkville RTL component provides a single PCIe Physical Function (PF) supporting some number of RX/Ingress and TX/Egress Queues. The ARK PMD controls the Arkville core through a dedicated opaque Core BAR (CBAR). To allow users full freedom for their own FPGA application IP, an independent FPGA Application BAR (ABAR) is provided.

One popular way to imagine Arkville’s FPGA hardware aspect is as the FPGA PCIe-facing side of a so-called Smart NIC. The Arkville core does not contain any MACs, and is link-speed independent, as well as agnostic to the number of physical ports the application chooses to use. The ARK driver exposes the familiar PMD interface to allow packet movement to and from mbufs across multiple queues.

However FPGA RTL applications could contain a universe of added functionality that an Arkville RTL core does not provide or can not anticipate. To allow for this expectation of user-defined innovation, the ARK PMD provides a dynamic mechanism of adding capabilities without having to modify the ARK PMD.

The ARK PMD is intended to support all instances of the Arkville RTL Core, regardless of configuration, FPGA vendor, or target board. While specific capabilities such as number of physical hardware queue-pairs are negotiated; the driver is designed to remain constant over a broad and extendable feature set.

Intentionally, Arkville by itself DOES NOT provide common NIC capabilities such as offload or receive-side scaling (RSS). These capabilities would be viewed as a gate-level “tax” on Green-box FPGA applications that do not require such function. Instead, they can be added as needed with essentially no overhead to the FPGA Application.

The ARK PMD also supports optional user extensions, through dynamic linking. The ARK PMD user extensions are a feature of Arkville’s DPDK net/ark poll mode driver, allowing users to add their own code to extend the net/ark functionality without having to make source code changes to the driver. One motivation for this capability is that while DPDK provides a rich set of functions to interact with NIC-like capabilities (e.g. MAC addresses and statistics), the Arkville RTL IP does not include a MAC. Users can supply their own MAC or custom FPGA applications, which may require control from the PMD. The user extension is the means providing the control between the user’s FPGA application and the existing DPDK features via the PMD.

Device Parameters

The ARK PMD supports device parameters that are used for packet routing and for internal packet generation and packet checking. This section describes the supported parameters. These features are primarily used for diagnostics, testing, and performance verification under the guidance of an Arkville specialist. The nominal use of Arkville does not require any configuration using these parameters.

“Pkt_dir”

The Packet Director controls connectivity between Arkville’s internal hardware components. The features of the Pkt_dir are only used for diagnostics and testing; it is not intended for nominal use. The full set of features are not published at this level.

Format: Pkt_dir=0x00110F10

“Pkt_gen”

The packet generator parameter takes a file as its argument. The file contains configuration parameters used internally for regression testing and are not intended to be published at this level. The packet generator is an internal Arkville hardware component.

Format: Pkt_gen=./config/pg.conf

“Pkt_chkr”

The packet checker parameter takes a file as its argument. The file contains configuration parameters used internally for regression testing and are not intended to be published at this level. The packet checker is an internal Arkville hardware component.

Format: Pkt_chkr=./config/pc.conf

Data Path Interface

Ingress RX and Egress TX operation is by the nominal DPDK API . The driver supports single-port, multi-queue for both RX and TX.

Configuration Information

DPDK Configuration Parameters

The following configuration options are available for the ARK PMD:

  • CONFIG_RTE_LIBRTE_ARK_PMD (default y): Enables or disables inclusion of the ARK PMD driver in the DPDK compilation.
  • CONFIG_RTE_LIBRTE_ARK_PAD_TX (default y): When enabled TX packets are padded to 60 bytes to support downstream MACS.
  • CONFIG_RTE_LIBRTE_ARK_DEBUG_RX (default n): Enables or disables debug logging and internal checking of RX ingress logic within the ARK PMD driver.
  • CONFIG_RTE_LIBRTE_ARK_DEBUG_TX (default n): Enables or disables debug logging and internal checking of TX egress logic within the ARK PMD driver.
  • CONFIG_RTE_LIBRTE_ARK_DEBUG_STATS (default n): Enables or disables debug logging of detailed packet and performance statistics gathered in the PMD and FPGA.
  • CONFIG_RTE_LIBRTE_ARK_DEBUG_TRACE (default n): Enables or disables debug logging of detailed PMD events and status.

Building DPDK

See the DPDK Getting Started Guide for Linux for instructions on how to build DPDK.

By default the ARK PMD library will be built into the DPDK library.

For configuring and using UIO and VFIO frameworks, please also refer the documentation that comes with DPDK suite.

Supported ARK RTL PCIe Instances

ARK PMD supports the following Arkville RTL PCIe instances including:

  • 1d6c:100d - AR-ARKA-FX0 [Arkville 32B DPDK Data Mover]
  • 1d6c:100e - AR-ARKA-FX1 [Arkville 64B DPDK Data Mover]

Supported Operating Systems

Any Linux distribution fulfilling the conditions described in System Requirements section of the DPDK documentation or refer to DPDK Release Notes. ARM and PowerPC architectures are not supported at this time.

Supported Features

  • Dynamic ARK PMD extensions
  • Multiple receive and transmit queues
  • Jumbo frames up to 9K
  • Hardware Statistics

Unsupported Features

Features that may be part of, or become part of, the Arkville RTL IP that are not currently supported or exposed by the ARK PMD include:

  • PCIe SR-IOV Virtual Functions (VFs)
  • Arkville’s Packet Generator Control and Status
  • Arkville’s Packet Director Control and Status
  • Arkville’s Packet Checker Control and Status
  • Arkville’s Timebase Management

Pre-Requisites

  1. Prepare the system as recommended by DPDK suite. This includes environment variables, hugepages configuration, tool-chains and configuration
  2. Insert igb_uio kernel module using the command ‘modprobe igb_uio’
  3. Bind the intended ARK device to igb_uio module

At this point the system should be ready to run DPDK applications. Once the application runs to completion, the ARK PMD can be detached from igb_uio if necessary.

Usage Example

Follow instructions available in the document compiling and testing a PMD for a NIC to launch testpmd with Atomic Rules ARK devices managed by librte_pmd_ark.

Example output:

[...]
EAL: PCI device 0000:01:00.0 on NUMA socket -1
EAL:   probe driver: 1d6c:100e rte_ark_pmd
EAL:   PCI memory mapped at 0x7f9b6c400000
PMD: eth_ark_dev_init(): Initializing 0:2:0.1
ARKP PMD CommitID: 378f3a67
Configuring Port 0 (socket 0)
Port 0: DC:3C:F6:00:00:01
Checking link statuses...
Port 0 Link Up - speed 100000 Mbps - full-duplex
Done
testpmd>

AVP Poll Mode Driver

The Accelerated Virtual Port (AVP) device is a shared memory based device only available on virtualization platforms from Wind River Systems. The Wind River Systems virtualization platform currently uses QEMU/KVM as its hypervisor and as such provides support for all of the QEMU supported virtual and/or emulated devices (e.g., virtio, e1000, etc.). The platform offers the virtio device type as the default device when launching a virtual machine or creating a virtual machine port. The AVP device is a specialized device available to customers that require increased throughput and decreased latency to meet the demands of their performance focused applications.

The AVP driver binds to any AVP PCI devices that have been exported by the Wind River Systems QEMU/KVM hypervisor. As a user of the DPDK driver API it supports a subset of the full Ethernet device API to enable the application to use the standard device configuration functions and packet receive/transmit functions.

These devices enable optimized packet throughput by bypassing QEMU and delivering packets directly to the virtual switch via a shared memory mechanism. This provides DPDK applications running in virtual machines with significantly improved throughput and latency over other device types.

The AVP device implementation is integrated with the QEMU/KVM live-migration mechanism to allow applications to seamlessly migrate from one hypervisor node to another with minimal packet loss.

Features and Limitations of the AVP PMD

The AVP PMD driver provides the following functionality.

  • Receive and transmit of both simple and chained mbuf packets,
  • Chained mbufs may include up to 5 chained segments,
  • Up to 8 receive and transmit queues per device,
  • Only a single MAC address is supported,
  • The MAC address cannot be modified,
  • The maximum receive packet length is 9238 bytes,
  • VLAN header stripping and inserting,
  • Promiscuous mode
  • VM live-migration
  • PCI hotplug insertion and removal

Prerequisites

The following prerequisites apply:

  • A virtual machine running in a Wind River Systems virtualization environment and configured with at least one neutron port defined with a vif-model set to “avp”.

Launching a VM with an AVP type network attachment

The following example will launch a VM with three network attachments. The first attachment will have a default vif-model of “virtio”. The next two network attachments will have a vif-model of “avp” and may be used with a DPDK application which is built to include the AVP PMD driver.

nova boot --flavor small --image my-image \
   --nic net-id=${NETWORK1_UUID} \
   --nic net-id=${NETWORK2_UUID},vif-model=avp \
   --nic net-id=${NETWORK3_UUID},vif-model=avp \
   --security-group default my-instance1

BNX2X Poll Mode Driver

The BNX2X poll mode driver library (librte_pmd_bnx2x) implements support for QLogic 578xx 10/20 Gbps family of adapters as well as their virtual functions (VF) in SR-IOV context. It is supported on several standard Linux distros like Red Hat 7.x and SLES12 OS. It is compile-tested under FreeBSD OS.

More information can be found at QLogic Corporation’s Official Website.

Supported Features

BNX2X PMD has support for:

  • Base L2 features
  • Unicast/multicast filtering
  • Promiscuous mode
  • Port hardware statistics
  • SR-IOV VF

Non-supported Features

The features not yet supported include:

  • TSS (Transmit Side Scaling)
  • RSS (Receive Side Scaling)
  • LRO/TSO offload
  • Checksum offload
  • SR-IOV PF
  • Rx TX scatter gather

Co-existence considerations

  • BCM578xx being a CNA can have both NIC and Storage personalities. However, coexistence with storage protocol drivers (cnic, bnx2fc and bnx2fi) is not supported on the same adapter. So storage personality has to be disabled on that adapter when used in DPDK applications.
  • For SR-IOV case, bnx2x PMD will be used to bind to SR-IOV VF device and Linux native kernel driver (bnx2x) will be attached to SR-IOV PF.

Supported QLogic NICs

  • 578xx

Prerequisites

  • Requires firmware version 7.2.51.0. It is included in most of the standard Linux distros. If it is not available visit QLogic Driver Download Center to get the required firmware.

Pre-Installation Configuration

Config File Options

The following options can be modified in the .config file. Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_LIBRTE_BNX2X_PMD (default n)

    Toggle compilation of bnx2x driver. To use bnx2x PMD set this config parameter to ‘y’. Also, in order for firmware binary to load user will need zlib devel package installed.

  • CONFIG_RTE_LIBRTE_BNX2X_DEBUG (default n)

    Toggle display of generic debugging messages.

  • CONFIG_RTE_LIBRTE_BNX2X_DEBUG_INIT (default n)

    Toggle display of initialization related messages.

  • CONFIG_RTE_LIBRTE_BNX2X_DEBUG_TX (default n)

    Toggle display of transmit fast path run-time messages.

  • CONFIG_RTE_LIBRTE_BNX2X_DEBUG_RX (default n)

    Toggle display of receive fast path run-time messages.

  • CONFIG_RTE_LIBRTE_BNX2X_DEBUG_PERIODIC (default n)

    Toggle display of register reads and writes.

Driver compilation and testing

Refer to the document compiling and testing a PMD for a NIC for details.

SR-IOV: Prerequisites and sample Application Notes

This section provides instructions to configure SR-IOV with Linux OS.

  1. Verify SR-IOV and ARI capabilities are enabled on the adapter using lspci:

    lspci -s <slot> -vvv
    

    Example output:

    [...]
    Capabilities: [1b8 v1] Alternative Routing-ID Interpretation (ARI)
    [...]
    Capabilities: [1c0 v1] Single Root I/O Virtualization (SR-IOV)
    [...]
    Kernel driver in use: igb_uio
    
  2. Load the kernel module:

    modprobe bnx2x
    

    Example output:

    systemd-udevd[4848]: renamed network interface eth0 to ens5f0
    systemd-udevd[4848]: renamed network interface eth1 to ens5f1
    
  3. Bring up the PF ports:

    ifconfig ens5f0 up
    ifconfig ens5f1 up
    
  4. Create VF device(s):

    Echo the number of VFs to be created into “sriov_numvfs” sysfs entry of the parent PF.

    Example output:

    echo 2 > /sys/devices/pci0000:00/0000:00:03.0/0000:81:00.0/sriov_numvfs
    
  5. Assign VF MAC address:

    Assign MAC address to the VF using iproute2 utility. The syntax is: ip link set <PF iface> vf <VF id> mac <macaddr>

    Example output:

    ip link set ens5f0 vf 0 mac 52:54:00:2f:9d:e8
    
  6. PCI Passthrough:

    The VF devices may be passed through to the guest VM using virt-manager or virsh etc. bnx2x PMD should be used to bind the VF devices in the guest VM using the instructions outlined in the Application notes below.

  7. Running testpmd:

    Follow instructions available in the document compiling and testing a PMD for a NIC to run testpmd.

    Example output:

    [...]
    EAL: PCI device 0000:84:00.0 on NUMA socket 1
    EAL:   probe driver: 14e4:168e rte_bnx2x_pmd
    EAL:   PCI memory mapped at 0x7f14f6fe5000
    EAL:   PCI memory mapped at 0x7f14f67e5000
    EAL:   PCI memory mapped at 0x7f15fbd9b000
    EAL: PCI device 0000:84:00.1 on NUMA socket 1
    EAL:   probe driver: 14e4:168e rte_bnx2x_pmd
    EAL:   PCI memory mapped at 0x7f14f5fe5000
    EAL:   PCI memory mapped at 0x7f14f57e5000
    EAL:   PCI memory mapped at 0x7f15fbd4f000
    Interactive-mode selected
    Configuring Port 0 (socket 0)
    PMD: bnx2x_dev_tx_queue_setup(): fp[00] req_bd=512, thresh=512,
                 usable_bd=1020, total_bd=1024,
                              tx_pages=4
    PMD: bnx2x_dev_rx_queue_setup(): fp[00] req_bd=128, thresh=0,
                 usable_bd=510, total_bd=512,
                              rx_pages=1, cq_pages=8
    PMD: bnx2x_print_adapter_info():
    [...]
    Checking link statuses...
    Port 0 Link Up - speed 10000 Mbps - full-duplex
    Port 1 Link Up - speed 10000 Mbps - full-duplex
    Done
    testpmd>
    

BNXT Poll Mode Driver

The bnxt poll mode library (librte_pmd_bnxt) implements support for:

  • Broadcom NetXtreme-C®/NetXtreme-E® BCM5730X and BCM574XX family of Ethernet Network Controllers

    These adapters support Standards compliant 10/25/50/100Gbps 30MPPS full-duplex throughput.

    Information about the NetXtreme family of adapters can be found in the NetXtreme® Brand section of the Broadcom website.

  • Broadcom StrataGX® BCM5871X Series of Communucations Processors

    These ARM based processors target a broad range of networking applications including virtual CPE (vCPE) and NFV appliances, 10G service routers and gateways, control plane processing for Ethernet switches and network attached storage (NAS).

    Information about the StrataGX family of adapters can be found in the StrataGX® BCM5871X Series section of the Broadcom website.

Limitations

With the current driver, allocated mbufs must be large enough to hold the entire received frame. If the mbufs are not large enough, the packets will be dropped. This is most limiting when jumbo frames are used.

CXGBE Poll Mode Driver

The CXGBE PMD (librte_pmd_cxgbe) provides poll mode driver support for Chelsio Terminator 10/25/40/100 Gbps family of adapters. CXGBE PMD has support for the latest Linux and FreeBSD operating systems.

More information can be found at Chelsio Communications Official Website.

Features

CXGBE PMD has support for:

  • Multiple queues for TX and RX
  • Receiver Side Steering (RSS)
  • VLAN filtering
  • Checksum offload
  • Promiscuous mode
  • All multicast mode
  • Port hardware statistics
  • Jumbo frames

Limitations

The Chelsio Terminator series of devices provide two/four ports but expose a single PCI bus address, thus, librte_pmd_cxgbe registers itself as a PCI driver that allocates one Ethernet device per detected port.

For this reason, one cannot whitelist/blacklist a single port without whitelisting/blacklisting the other ports on the same device.

Supported Chelsio T5 NICs

  • 1G NICs: T502-BT
  • 10G NICs: T520-BT, T520-CR, T520-LL-CR, T520-SO-CR, T540-CR
  • 40G NICs: T580-CR, T580-LP-CR, T580-SO-CR
  • Other T5 NICs: T522-CR

Supported Chelsio T6 NICs

  • 25G NICs: T6425-CR, T6225-CR, T6225-LL-CR, T6225-SO-CR
  • 100G NICs: T62100-CR, T62100-LP-CR, T62100-SO-CR

Prerequisites

  • Requires firmware version 1.16.43.0 and higher. Visit Chelsio Download Center to get latest firmware bundled with the latest Chelsio Unified Wire package.

    For Linux, installing and loading the latest cxgb4 kernel driver from the Chelsio Unified Wire package should get you the latest firmware. More information can be obtained from the User Guide that is bundled with the Chelsio Unified Wire package.

    For FreeBSD, the latest firmware obtained from the Chelsio Unified Wire package must be manually flashed via cxgbetool available in FreeBSD source repository.

    Instructions on how to manually flash the firmware are given in section Linux Installation for Linux and section FreeBSD Installation for FreeBSD.

Pre-Installation Configuration

Config File Options

The following options can be modified in the .config file. Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_LIBRTE_CXGBE_PMD (default y)

    Toggle compilation of librte_pmd_cxgbe driver.

  • CONFIG_RTE_LIBRTE_CXGBE_DEBUG (default n)

    Toggle display of generic debugging messages.

  • CONFIG_RTE_LIBRTE_CXGBE_DEBUG_REG (default n)

    Toggle display of registers related run-time check messages.

  • CONFIG_RTE_LIBRTE_CXGBE_DEBUG_MBOX (default n)

    Toggle display of firmware mailbox related run-time check messages.

  • CONFIG_RTE_LIBRTE_CXGBE_DEBUG_TX (default n)

    Toggle display of transmission data path run-time check messages.

  • CONFIG_RTE_LIBRTE_CXGBE_DEBUG_RX (default n)

    Toggle display of receiving data path run-time check messages.

  • CONFIG_RTE_LIBRTE_CXGBE_TPUT (default y)

    Toggle behaviour to prefer Throughput or Latency.

Driver compilation and testing

Refer to the document compiling and testing a PMD for a NIC for details.

Linux

Linux Installation

Steps to manually install the latest firmware from the downloaded Chelsio Unified Wire package for Linux operating system are as follows:

  1. Load the kernel module:

    modprobe cxgb4
    
  2. Use ifconfig to get the interface name assigned to Chelsio card:

    ifconfig -a | grep "00:07:43"
    

    Example output:

    p1p1      Link encap:Ethernet  HWaddr 00:07:43:2D:EA:C0
    p1p2      Link encap:Ethernet  HWaddr 00:07:43:2D:EA:C8
    
  3. Install cxgbtool:

    cd <path_to_uwire>/tools/cxgbtool
    make install
    
  4. Use cxgbtool to load the firmware config file onto the card:

    cxgbtool p1p1 loadcfg <path_to_uwire>/src/network/firmware/t5-config.txt
    
  5. Use cxgbtool to load the firmware image onto the card:

    cxgbtool p1p1 loadfw <path_to_uwire>/src/network/firmware/t5fw-*.bin
    
  6. Unload and reload the kernel module:

    modprobe -r cxgb4
    modprobe cxgb4
    
  7. Verify with ethtool:

    ethtool -i p1p1 | grep "firmware"
    

    Example output:

    firmware-version: 1.16.43.0, TP 0.1.4.9
    
Running testpmd

This section demonstrates how to launch testpmd with Chelsio devices managed by librte_pmd_cxgbe in Linux operating system.

  1. Load the kernel module:

    modprobe cxgb4
    
  2. Get the PCI bus addresses of the interfaces bound to cxgb4 driver:

    dmesg | tail -2
    

    Example output:

    cxgb4 0000:02:00.4 p1p1: renamed from eth0
    cxgb4 0000:02:00.4 p1p2: renamed from eth1
    

    Note

    Both the interfaces of a Chelsio 2-port adapter are bound to the same PCI bus address.

  3. Unload the kernel module:

    modprobe -ar cxgb4 csiostor
    
  4. Running testpmd

    Follow instructions available in the document compiling and testing a PMD for a NIC to run testpmd.

    Note

    Currently, CXGBE PMD only supports the binding of PF4 for Chelsio NICs.

    Example output:

    [...]
    EAL: PCI device 0000:02:00.4 on NUMA socket -1
    EAL:   probe driver: 1425:5401 rte_cxgbe_pmd
    EAL:   PCI memory mapped at 0x7fd7c0200000
    EAL:   PCI memory mapped at 0x7fd77cdfd000
    EAL:   PCI memory mapped at 0x7fd7c10b7000
    PMD: rte_cxgbe_pmd: fw: 1.16.43.0, TP: 0.1.4.9
    PMD: rte_cxgbe_pmd: Coming up as MASTER: Initializing adapter
    Interactive-mode selected
    Configuring Port 0 (socket 0)
    Port 0: 00:07:43:2D:EA:C0
    Configuring Port 1 (socket 0)
    Port 1: 00:07:43:2D:EA:C8
    Checking link statuses...
    PMD: rte_cxgbe_pmd: Port0: passive DA port module inserted
    PMD: rte_cxgbe_pmd: Port1: passive DA port module inserted
    Port 0 Link Up - speed 10000 Mbps - full-duplex
    Port 1 Link Up - speed 10000 Mbps - full-duplex
    Done
    testpmd>
    

    Note

    Flow control pause TX/RX is disabled by default and can be enabled via testpmd. Refer section Enable/Disable Flow Control for more details.

FreeBSD

FreeBSD Installation

Steps to manually install the latest firmware from the downloaded Chelsio Unified Wire package for FreeBSD operating system are as follows:

  1. Load the kernel module:

    kldload if_cxgbe
    
  2. Use dmesg to get the t5nex instance assigned to the Chelsio card:

    dmesg | grep "t5nex"
    

    Example output:

    t5nex0: <Chelsio T520-CR> irq 16 at device 0.4 on pci2
    cxl0: <port 0> on t5nex0
    cxl1: <port 1> on t5nex0
    t5nex0: PCIe x8, 2 ports, 14 MSI-X interrupts, 31 eq, 13 iq
    

    In the example above, a Chelsio T520-CR card is bound to a t5nex0 instance.

  3. Install cxgbetool from FreeBSD source repository:

    cd <path_to_FreeBSD_source>/tools/tools/cxgbetool/
    make && make install
    
  4. Use cxgbetool to load the firmware image onto the card:

    cxgbetool t5nex0 loadfw <path_to_uwire>/src/network/firmware/t5fw-*.bin
    
  5. Unload and reload the kernel module:

    kldunload if_cxgbe
    kldload if_cxgbe
    
  6. Verify with sysctl:

    sysctl -a | grep "t5nex" | grep "firmware"
    

    Example output:

    dev.t5nex.0.firmware_version: 1.16.43.0
    
Running testpmd

This section demonstrates how to launch testpmd with Chelsio devices managed by librte_pmd_cxgbe in FreeBSD operating system.

  1. Change to DPDK source directory where the target has been compiled in section Driver compilation and testing:

    cd <DPDK-source-directory>
    
  2. Copy the contigmem kernel module to /boot/kernel directory:

    cp x86_64-native-bsdapp-clang/kmod/contigmem.ko /boot/kernel/
    
  3. Add the following lines to /boot/loader.conf:

    # reserve 2 x 1G blocks of contiguous memory using contigmem driver
    hw.contigmem.num_buffers=2
    hw.contigmem.buffer_size=1073741824
    # load contigmem module during boot process
    contigmem_load="YES"
    

    The above lines load the contigmem kernel module during boot process and allocate 2 x 1G blocks of contiguous memory to be used for DPDK later on. This is to avoid issues with potential memory fragmentation during later system up time, which may result in failure of allocating the contiguous memory required for the contigmem kernel module.

  4. Restart the system and ensure the contigmem module is loaded successfully:

    reboot
    kldstat | grep "contigmem"
    

    Example output:

    2    1 0xffffffff817f1000 3118     contigmem.ko
    
  5. Repeat step 1 to ensure that you are in the DPDK source directory.

  6. Load the cxgbe kernel module:

    kldload if_cxgbe
    
  7. Get the PCI bus addresses of the interfaces bound to t5nex driver:

    pciconf -l | grep "t5nex"
    

    Example output:

    t5nex0@pci0:2:0:4: class=0x020000 card=0x00001425 chip=0x54011425 rev=0x00
    

    In the above example, the t5nex0 is bound to 2:0:4 bus address.

    Note

    Both the interfaces of a Chelsio 2-port adapter are bound to the same PCI bus address.

  8. Unload the kernel module:

    kldunload if_cxgbe
    
  9. Set the PCI bus addresses to hw.nic_uio.bdfs kernel environment parameter:

    kenv hw.nic_uio.bdfs="2:0:4"
    

    This automatically binds 2:0:4 to nic_uio kernel driver when it is loaded in the next step.

    Note

    Currently, CXGBE PMD only supports the binding of PF4 for Chelsio NICs.

  10. Load nic_uio kernel driver:

    kldload ./x86_64-native-bsdapp-clang/kmod/nic_uio.ko
    
  11. Start testpmd with basic parameters:

    ./x86_64-native-bsdapp-clang/app/testpmd -l 0-3 -n 4 -w 0000:02:00.4 -- -i
    

    Example output:

    [...]
    EAL: PCI device 0000:02:00.4 on NUMA socket 0
    EAL:   probe driver: 1425:5401 rte_cxgbe_pmd
    EAL:   PCI memory mapped at 0x8007ec000
    EAL:   PCI memory mapped at 0x842800000
    EAL:   PCI memory mapped at 0x80086c000
    PMD: rte_cxgbe_pmd: fw: 1.16.43.0, TP: 0.1.4.9
    PMD: rte_cxgbe_pmd: Coming up as MASTER: Initializing adapter
    Interactive-mode selected
    Configuring Port 0 (socket 0)
    Port 0: 00:07:43:2D:EA:C0
    Configuring Port 1 (socket 0)
    Port 1: 00:07:43:2D:EA:C8
    Checking link statuses...
    PMD: rte_cxgbe_pmd: Port0: passive DA port module inserted
    PMD: rte_cxgbe_pmd: Port1: passive DA port module inserted
    Port 0 Link Up - speed 10000 Mbps - full-duplex
    Port 1 Link Up - speed 10000 Mbps - full-duplex
    Done
    testpmd>
    

Note

Flow control pause TX/RX is disabled by default and can be enabled via testpmd. Refer section Enable/Disable Flow Control for more details.

Sample Application Notes

Enable/Disable Flow Control

Flow control pause TX/RX is disabled by default and can be enabled via testpmd as follows:

testpmd> set flow_ctrl rx on tx on 0 0 0 0 mac_ctrl_frame_fwd off autoneg on 0
testpmd> set flow_ctrl rx on tx on 0 0 0 0 mac_ctrl_frame_fwd off autoneg on 1

To disable again, run:

testpmd> set flow_ctrl rx off tx off 0 0 0 0 mac_ctrl_frame_fwd off autoneg off 0
testpmd> set flow_ctrl rx off tx off 0 0 0 0 mac_ctrl_frame_fwd off autoneg off 1
Jumbo Mode

There are two ways to enable sending and receiving of jumbo frames via testpmd. One method involves using the mtu command, which changes the mtu of an individual port without having to stop the selected port. Another method involves stopping all the ports first and then running max-pkt-len command to configure the mtu of all the ports with a single command.

  • To configure each port individually, run the mtu command as follows:

    testpmd> port config mtu 0 9000
    testpmd> port config mtu 1 9000
    
  • To configure all the ports at once, stop all the ports first and run the max-pkt-len command as follows:

    testpmd> port stop all
    testpmd> port config all max-pkt-len 9000
    

DPAA Poll Mode Driver

The DPAA NIC PMD (librte_pmd_dpaa) provides poll mode driver support for the inbuilt NIC found in the NXP DPAA SoC family.

More information can be found at NXP Official Website.

NXP DPAA (Data Path Acceleration Architecture - Gen 1)

This section provides an overview of the NXP DPAA architecture and how it is integrated into the DPDK.

Contents summary

  • DPAA overview
  • DPAA driver architecture overview
DPAA Overview

Reference: FSL DPAA Architecture.

The QorIQ Data Path Acceleration Architecture (DPAA) is a set of hardware components on specific QorIQ series multicore processors. This architecture provides the infrastructure to support simplified sharing of networking interfaces and accelerators by multiple CPU cores, and the accelerators themselves.

DPAA includes:

  • Cores
  • Network and packet I/O
  • Hardware offload accelerators
  • Infrastructure required to facilitate flow of packets between the components above

Infrastructure components are:

  • The Queue Manager (QMan) is a hardware accelerator that manages frame queues. It allows CPUs and other accelerators connected to the SoC datapath to enqueue and dequeue ethernet frames, thus providing the infrastructure for data exchange among CPUs and datapath accelerators.
  • The Buffer Manager (BMan) is a hardware buffer pool management block that allows software and accelerators on the datapath to acquire and release buffers in order to build frames.

Hardware accelerators are:

  • SEC - Cryptographic accelerator
  • PME - Pattern matching engine

The Network and packet I/O component:

  • The Frame Manager (FMan) is a key component in the DPAA and makes use of the DPAA infrastructure (QMan and BMan). FMan is responsible for packet distribution and policing. Each frame can be parsed, classified and results may be attached to the frame. This meta data can be used to select particular QMan queue, which the packet is forwarded to.

DPAA DPDK - Poll Mode Driver Overview

This section provides an overview of the drivers for DPAA:

  • Bus driver and associated “DPAA infrastructure” drivers
  • Functional object drivers (such as Ethernet).

Brief description of each driver is provided in layout below as well as in the following sections.

                                   +------------+
                                   | DPDK DPAA  |
                                   |    PMD     |
                                   +-----+------+
                                         |
                                   +-----+------+       +---------------+
                                   :  Ethernet  :.......| DPDK DPAA     |
                . . . . . . . . .  :   (FMAN)   :       | Mempool driver|
               .                   +---+---+----+       |  (BMAN)       |
              .                        ^   |            +-----+---------+
             .                         |   |<enqueue,         .
            .                          |   | dequeue>         .
           .                           |   |                  .
          .                        +---+---V----+             .
         .      . . . . . . . . . .: Portal drv :             .
        .      .                   :            :             .
       .      .                    +-----+------+             .
      .      .                     :   QMAN     :             .
     .      .                      :  Driver    :             .
+----+------+-------+              +-----+------+             .
|   DPDK DPAA Bus   |                    |                    .
|   driver          |....................|.....................
|   /bus/dpaa       |                    |
+-------------------+                    |
                                         |
========================== HARDWARE =====|========================
                                        PHY
=========================================|========================

In the above representation, solid lines represent components which interface with DPDK RTE Framework and dotted lines represent DPAA internal components.

DPAA Bus driver

The DPAA bus driver is a rte_bus driver which scans the platform like bus. Key functions include:

  • Scanning and parsing the various objects and adding them to their respective device list.
  • Performing probe for available drivers against each scanned device
  • Creating necessary ethernet instance before passing control to the PMD
DPAA NIC Driver (PMD)

DPAA PMD is traditional DPDK PMD which provides necessary interface between RTE framework and DPAA internal components/drivers.

  • Once devices have been identified by DPAA Bus, each device is associated with the PMD
  • PMD is responsible for implementing necessary glue layer between RTE APIs and lower level QMan and FMan blocks. The Ethernet driver is bound to a FMAN port and implements the interfaces needed to connect the DPAA network interface to the network stack. Each FMAN Port corresponds to a DPDK network interface.
Features

Features of the DPAA PMD are:

  • Multiple queues for TX and RX
  • Receive Side Scaling (RSS)
  • Packet type information
  • Checksum offload
  • Promiscuous mode
DPAA Mempool Driver

DPAA has a hardware offloaded buffer pool manager, called BMan, or Buffer Manager.

  • Using standard Mempools operations RTE API, the mempool driver interfaces with RTE to service each mempool creation, deletion, buffer allocation and deallocation requests.
  • Each FMAN instance has a BMan pool attached to it during initialization. Each Tx frame can be automatically released by hardware, if allocated from this pool.

Supported DPAA SoCs

  • LS1043A/LS1023A
  • LS1046A/LS1026A

Prerequisites

There are three main pre-requisities for executing DPAA PMD on a DPAA compatible board:

  1. ARM 64 Tool Chain

    For example, the *aarch64* Linaro Toolchain.

  2. Linux Kernel

    It can be obtained from NXP’s Github hosting.

  3. Rootfile system

    Any aarch64 supporting filesystem can be used. For example, Ubuntu 15.10 (Wily) or 16.04 LTS (Xenial) userland which can be obtained from here.

  4. FMC Tool

    Before any DPDK application can be executed, the Frame Manager Configuration Tool (FMC) need to be executed to set the configurations of the queues. This includes the queue state, RSS and other policies. This tool can be obtained from NXP (Freescale) Public Git Repository.

    This tool needs configuration files which are available in the DPDK Extra Scripts, described below for DPDK usages.

As an alternative method, DPAA PMD can also be executed using images provided as part of SDK from NXP. The SDK includes all the above prerequisites necessary to bring up a DPAA board.

The following dependencies are not part of DPDK and must be installed separately:

  • NXP Linux SDK

    NXP Linux software development kit (SDK) includes support for family of QorIQ® ARM-Architecture-based system on chip (SoC) processors and corresponding boards.

    It includes the Linux board support packages (BSPs) for NXP SoCs, a fully operational tool chain, kernel and board specific modules.

    SDK and related information can be obtained from: NXP QorIQ SDK.

  • DPDK Extra Scripts

    DPAA based resources can be configured easily with the help of ready scripts as provided in the DPDK Extra repository.

    DPDK Extras Scripts.

Currently supported by DPDK:

Note

Some part of dpaa bus code (qbman and fman - library) routines are dual licensed (BSD & GPLv2), however they are used as BSD in DPDK in userspace.

Pre-Installation Configuration

Config File Options

The following options can be modified in the config file. Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_LIBRTE_DPAA_BUS (default n)

    By default it is enabled only for defconfig_arm64-dpaa-* config. Toggle compilation of the librte_bus_dpaa driver.

  • CONFIG_RTE_LIBRTE_DPAA_PMD (default n)

    By default it is enabled only for defconfig_arm64-dpaa-* config. Toggle compilation of the librte_pmd_dpaa driver.

  • CONFIG_RTE_LIBRTE_DPAA_DEBUG_DRIVER (default n)

    Toggles display of bus configurations and enables a debugging queue to fetch error (Rx/Tx) packets to driver. By default, packets with errors (like wrong checksum) are dropped by the hardware.

  • CONFIG_RTE_LIBRTE_DPAA_HWDEBUG (default n)

    Enables debugging of the Queue and Buffer Manager layer which interacts with the DPAA hardware.

  • CONFIG_RTE_MBUF_DEFAULT_MEMPOOL_OPS (default dpaa)

    This is not a DPAA specific configuration - it is a generic RTE config. For optimal performance and hardware utilization, it is expected that DPAA Mempool driver is used for mempools. For that, this configuration needs to enabled.

Environment Variables

DPAA drivers uses the following environment variables to configure its state during application initialization:

  • DPAA_NUM_RX_QUEUES (default 1)

    This defines the number of Rx queues configured for an application, per port. Hardware would distribute across these many number of queues on Rx of packets. In case the application is configured to use lesser number of queues than configured above, it might result in packet loss (because of distribution).

Driver compilation and testing

Refer to the document compiling and testing a PMD for a NIC for details.

  1. Running testpmd:

    Follow instructions available in the document compiling and testing a PMD for a NIC to run testpmd.

    Example output:

    ./arm64-dpaa-linuxapp-gcc/testpmd -c 0xff -n 1 \
      -- -i --portmask=0x3 --nb-cores=1 --no-flush-rx
    
    .....
    EAL: Registered [pci] bus.
    EAL: Registered [dpaa] bus.
    EAL: Detected 4 lcore(s)
    .....
    EAL: dpaa: Bus scan completed
    .....
    Configuring Port 0 (socket 0)
    Port 0: 00:00:00:00:00:01
    Configuring Port 1 (socket 0)
    Port 1: 00:00:00:00:00:02
    .....
    Checking link statuses...
    Port 0 Link Up - speed 10000 Mbps - full-duplex
    Port 1 Link Up - speed 10000 Mbps - full-duplex
    Done
    testpmd>
    

Limitations

Platform Requirement

DPAA drivers for DPDK can only work on NXP SoCs as listed in the Supported DPAA SoCs.

Maximum packet length

The DPAA SoC family support a maximum of a 10240 jumbo frame. The value is fixed and cannot be changed. So, even when the rxmode.max_rx_pkt_len member of struct rte_eth_conf is set to a value lower than 10240, frames up to 10240 bytes can still reach the host interface.

Multiprocess Support

Current version of DPAA driver doesn’t support multi-process applications where I/O is performed using secondary processes. This feature would be implemented in subsequent versions.

DPAA2 Poll Mode Driver

The DPAA2 NIC PMD (librte_pmd_dpaa2) provides poll mode driver support for the inbuilt NIC found in the NXP DPAA2 SoC family.

More information can be found at NXP Official Website.

NXP DPAA2 (Data Path Acceleration Architecture Gen2)

This section provides an overview of the NXP DPAA2 architecture and how it is integrated into the DPDK.

Contents summary

  • DPAA2 overview
  • Overview of DPAA2 objects
  • DPAA2 driver architecture overview
DPAA2 Overview

Reference: FSL MC BUS in Linux Kernel.

DPAA2 is a hardware architecture designed for high-speed network packet processing. DPAA2 consists of sophisticated mechanisms for processing Ethernet packets, queue management, buffer management, autonomous L2 switching, virtual Ethernet bridging, and accelerator (e.g. crypto) sharing.

A DPAA2 hardware component called the Management Complex (or MC) manages the DPAA2 hardware resources. The MC provides an object-based abstraction for software drivers to use the DPAA2 hardware.

The MC uses DPAA2 hardware resources such as queues, buffer pools, and network ports to create functional objects/devices such as network interfaces, an L2 switch, or accelerator instances.

The MC provides memory-mapped I/O command interfaces (MC portals) which DPAA2 software drivers use to operate on DPAA2 objects:

The diagram below shows an overview of the DPAA2 resource management architecture:

+--------------------------------------+
|                  OS                  |
|                        DPAA2 drivers |
|                             |        |
+-----------------------------|--------+
                              |
                              | (create,discover,connect
                              |  config,use,destroy)
                              |
                DPAA2         |
+------------------------| mc portal |-+
|                             |        |
|   +- - - - - - - - - - - - -V- - -+  |
|   |                               |  |
|   |   Management Complex (MC)     |  |
|   |                               |  |
|   +- - - - - - - - - - - - - - - -+  |
|                                      |
| Hardware                  Hardware   |
| Resources                 Objects    |
| ---------                 -------    |
| -queues                   -DPRC      |
| -buffer pools             -DPMCP     |
| -Eth MACs/ports           -DPIO      |
| -network interface        -DPNI      |
|  profiles                 -DPMAC     |
| -queue portals            -DPBP      |
| -MC portals                ...       |
|  ...                                 |
|                                      |
+--------------------------------------+

The MC mediates operations such as create, discover, connect, configuration, and destroy. Fast-path operations on data, such as packet transmit/receive, are not mediated by the MC and are done directly using memory mapped regions in DPIO objects.

Overview of DPAA2 Objects

The section provides a brief overview of some key DPAA2 objects. A simple scenario is described illustrating the objects involved in creating a network interfaces.

DPRC (Datapath Resource Container)

A DPRC is a container object that holds all the other types of DPAA2 objects. In the example diagram below there are 8 objects of 5 types (DPMCP, DPIO, DPBP, DPNI, and DPMAC) in the container.
+---------------------------------------------------------+
| DPRC                                                    |
|                                                         |
|  +-------+  +-------+  +-------+  +-------+  +-------+  |
|  | DPMCP |  | DPIO  |  | DPBP  |  | DPNI  |  | DPMAC |  |
|  +-------+  +-------+  +-------+  +---+---+  +---+---+  |
|  | DPMCP |  | DPIO  |                                   |
|  +-------+  +-------+                                   |
|  | DPMCP |                                              |
|  +-------+                                              |
|                                                         |
+---------------------------------------------------------+

From the point of view of an OS, a DPRC behaves similar to a plug and play bus, like PCI. DPRC commands can be used to enumerate the contents of the DPRC, discover the hardware objects present (including mappable regions and interrupts).

DPRC.1 (bus)
  |
  +--+--------+-------+-------+-------+
     |        |       |       |       |
   DPMCP.1  DPIO.1  DPBP.1  DPNI.1  DPMAC.1
   DPMCP.2  DPIO.2
   DPMCP.3

Hardware objects can be created and destroyed dynamically, providing the ability to hot plug/unplug objects in and out of the DPRC.

A DPRC has a mappable MMIO region (an MC portal) that can be used to send MC commands. It has an interrupt for status events (like hotplug).

All objects in a container share the same hardware “isolation context”. This means that with respect to an IOMMU the isolation granularity is at the DPRC (container) level, not at the individual object level.

DPRCs can be defined statically and populated with objects via a config file passed to the MC when firmware starts it. There is also a Linux user space tool called “restool” that can be used to create/destroy containers and objects dynamically.

DPAA2 Objects for an Ethernet Network Interface

A typical Ethernet NIC is monolithic– the NIC device contains TX/RX queuing mechanisms, configuration mechanisms, buffer management, physical ports, and interrupts. DPAA2 uses a more granular approach utilizing multiple hardware objects. Each object provides specialized functions. Groups of these objects are used by software to provide Ethernet network interface functionality. This approach provides efficient use of finite hardware resources, flexibility, and performance advantages.

The diagram below shows the objects needed for a simple network interface configuration on a system with 2 CPUs.

+---+---+ +---+---+
   CPU0     CPU1
+---+---+ +---+---+
    |         |
+---+---+ +---+---+
   DPIO     DPIO
+---+---+ +---+---+
      \     /
       \   /
        \ /
     +---+---+
        DPNI  --- DPBP,DPMCP
     +---+---+
         |
         |
     +---+---+
       DPMAC
     +---+---+
         |
      port/PHY

Below the objects are described. For each object a brief description is provided along with a summary of the kinds of operations the object supports and a summary of key resources of the object (MMIO regions and IRQs).

DPMAC (Datapath Ethernet MAC): represents an Ethernet MAC, a hardware device that connects to an Ethernet PHY and allows physical transmission and reception of Ethernet frames.

  • MMIO regions: none
  • IRQs: DPNI link change
  • commands: set link up/down, link config, get stats, IRQ config, enable, reset

DPNI (Datapath Network Interface): contains TX/RX queues, network interface configuration, and RX buffer pool configuration mechanisms. The TX/RX queues are in memory and are identified by queue number.

  • MMIO regions: none
  • IRQs: link state
  • commands: port config, offload config, queue config, parse/classify config, IRQ config, enable, reset

DPIO (Datapath I/O): provides interfaces to enqueue and dequeue packets and do hardware buffer pool management operations. The DPAA2 architecture separates the mechanism to access queues (the DPIO object) from the queues themselves. The DPIO provides an MMIO interface to enqueue/dequeue packets. To enqueue something a descriptor is written to the DPIO MMIO region, which includes the target queue number. There will typically be one DPIO assigned to each CPU. This allows all CPUs to simultaneously perform enqueue/dequeued operations. DPIOs are expected to be shared by different DPAA2 drivers.

  • MMIO regions: queue operations, buffer management
  • IRQs: data availability, congestion notification, buffer pool depletion
  • commands: IRQ config, enable, reset

DPBP (Datapath Buffer Pool): represents a hardware buffer pool.

  • MMIO regions: none
  • IRQs: none
  • commands: enable, reset

DPMCP (Datapath MC Portal): provides an MC command portal. Used by drivers to send commands to the MC to manage objects.

  • MMIO regions: MC command portal
  • IRQs: command completion
  • commands: IRQ config, enable, reset
Object Connections

Some objects have explicit relationships that must be configured:

  • DPNI <–> DPMAC
  • DPNI <–> DPNI
  • DPNI <–> L2-switch-port

A DPNI must be connected to something such as a DPMAC, another DPNI, or L2 switch port. The DPNI connection is made via a DPRC command.

+-------+  +-------+
| DPNI  |  | DPMAC |
+---+---+  +---+---+
    |          |
    +==========+
  • DPNI <–> DPBP

A network interface requires a ‘buffer pool’ (DPBP object) which provides a list of pointers to memory where received Ethernet data is to be copied. The Ethernet driver configures the DPBPs associated with the network interface.

Interrupts

All interrupts generated by DPAA2 objects are message interrupts. At the hardware level message interrupts generated by devices will normally have 3 components– 1) a non-spoofable ‘device-id’ expressed on the hardware bus, 2) an address, 3) a data value.

In the case of DPAA2 devices/objects, all objects in the same container/DPRC share the same ‘device-id’. For ARM-based SoC this is the same as the stream ID.

DPAA2 DPDK - Poll Mode Driver Overview

This section provides an overview of the drivers for DPAA2– 1) the bus driver and associated “DPAA2 infrastructure” drivers and 2) functional object drivers (such as Ethernet).

As described previously, a DPRC is a container that holds the other types of DPAA2 objects. It is functionally similar to a plug-and-play bus controller.

Each object in the DPRC is a Linux “device” and is bound to a driver. The diagram below shows the dpaa2 drivers involved in a networking scenario and the objects bound to each driver. A brief description of each driver follows.

A brief description of each driver is provided below.

DPAA2 bus driver

The DPAA2 bus driver is a rte_bus driver which scans the fsl-mc bus. Key functions include:

  • Reading the container and setting up vfio group
  • Scanning and parsing the various MC objects and adding them to their respective device list.

Additionally, it also provides the object driver for generic MC objects.

DPIO driver

The DPIO driver is bound to DPIO objects and provides services that allow other drivers such as the Ethernet driver to enqueue and dequeue data for their respective objects. Key services include:

  • Data availability notifications
  • Hardware queuing operations (enqueue and dequeue of data)
  • Hardware buffer pool management

To transmit a packet the Ethernet driver puts data on a queue and invokes a DPIO API. For receive, the Ethernet driver registers a data availability notification callback. To dequeue a packet a DPIO API is used.

There is typically one DPIO object per physical CPU for optimum performance, allowing different CPUs to simultaneously enqueue and dequeue data.

The DPIO driver operates on behalf of all DPAA2 drivers active – Ethernet, crypto, compression, etc.

DPBP based Mempool driver

The DPBP driver is bound to a DPBP objects and provides sevices to create a hardware offloaded packet buffer mempool.

DPAA2 NIC Driver

The Ethernet driver is bound to a DPNI and implements the kernel interfaces needed to connect the DPAA2 network interface to the network stack.

Each DPNI corresponds to a DPDK network interface.

Features

Features of the DPAA2 PMD are:

  • Multiple queues for TX and RX
  • Receive Side Scaling (RSS)
  • MAC/VLAN filtering
  • Packet type information
  • Checksum offload
  • Promiscuous mode
  • Multicast mode
  • Port hardware statistics
  • Jumbo frames
  • Link flow control
  • Scattered and gather for TX and RX

Supported DPAA2 SoCs

  • LS2080A/LS2040A
  • LS2084A/LS2044A
  • LS2088A/LS2048A
  • LS1088A/LS1048A

Prerequisites

There are three main pre-requisities for executing DPAA2 PMD on a DPAA2 compatible board:

  1. ARM 64 Tool Chain

    For example, the *aarch64* Linaro Toolchain.

  2. Linux Kernel

    It can be obtained from NXP’s Github hosting.

  3. Rootfile system

    Any aarch64 supporting filesystem can be used. For example, Ubuntu 15.10 (Wily) or 16.04 LTS (Xenial) userland which can be obtained from here.

As an alternative method, DPAA2 PMD can also be executed using images provided as part of SDK from NXP. The SDK includes all the above prerequisites necessary to bring up a DPAA2 board.

The following dependencies are not part of DPDK and must be installed separately:

  • NXP Linux SDK

    NXP Linux software development kit (SDK) includes support for family of QorIQ® ARM-Architecture-based system on chip (SoC) processors and corresponding boards.

    It includes the Linux board support packages (BSPs) for NXP SoCs, a fully operational tool chain, kernel and board specific modules.

    SDK and related information can be obtained from: NXP QorIQ SDK.

  • DPDK Extra Scripts

    DPAA2 based resources can be configured easily with the help of ready scripts as provided in the DPDK Extra repository.

    DPDK Extras Scripts.

Currently supported by DPDK:

  • NXP SDK 17.08+.
  • MC Firmware version 10.3.1 and higher.
  • Supported architectures: arm64 LE.
  • Follow the DPDK Getting Started Guide for Linux to setup the basic DPDK environment.

Note

Some part of fslmc bus code (mc flib - object library) routines are dual licensed (BSD & GPLv2), however they are used as BSD in DPDK in userspace.

Pre-Installation Configuration

Config File Options

The following options can be modified in the config file. Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_LIBRTE_FSLMC_BUS (default n)

    By default it is enabled only for defconfig_arm64-dpaa2-* config. Toggle compilation of the librte_bus_fslmc driver.

  • CONFIG_RTE_LIBRTE_DPAA2_PMD (default n)

    By default it is enabled only for defconfig_arm64-dpaa2-* config. Toggle compilation of the librte_pmd_dpaa2 driver.

  • CONFIG_RTE_LIBRTE_DPAA2_DEBUG_DRIVER (default n)

    Toggle display of generic debugging messages

  • CONFIG_RTE_LIBRTE_DPAA2_USE_PHYS_IOVA (default y)

    Toggle to use physical address vs virtual address for hardware accelerators.

  • CONFIG_RTE_LIBRTE_DPAA2_DEBUG_INIT (default n)

    Toggle display of initialization related messages.

  • CONFIG_RTE_LIBRTE_DPAA2_DEBUG_RX (default n)

    Toggle display of receive fast path run-time message

  • CONFIG_RTE_LIBRTE_DPAA2_DEBUG_TX (default n)

    Toggle display of transmit fast path run-time message

  • CONFIG_RTE_LIBRTE_DPAA2_DEBUG_TX_FREE (default n)

    Toggle display of transmit fast path buffer free run-time message

Driver compilation and testing

Refer to the document compiling and testing a PMD for a NIC for details.

  1. Running testpmd:

    Follow instructions available in the document compiling and testing a PMD for a NIC to run testpmd.

    Example output:

    ./arm64-dpaa2-linuxapp-gcc/testpmd -c 0xff -n 1 \
      -- -i --portmask=0x3 --nb-cores=1 --no-flush-rx
    
    .....
    EAL: Registered [pci] bus.
    EAL: Registered [fslmc] bus.
    EAL: Detected 8 lcore(s)
    EAL: Probing VFIO support...
    EAL: VFIO support initialized
    .....
    PMD: DPAA2: Processing Container = dprc.2
    EAL: fslmc: DPRC contains = 51 devices
    EAL: fslmc: Bus scan completed
    .....
    Configuring Port 0 (socket 0)
    Port 0: 00:00:00:00:00:01
    Configuring Port 1 (socket 0)
    Port 1: 00:00:00:00:00:02
    .....
    Checking link statuses...
    Port 0 Link Up - speed 10000 Mbps - full-duplex
    Port 1 Link Up - speed 10000 Mbps - full-duplex
    Done
    testpmd>
    

Limitations

Platform Requirement

DPAA2 drivers for DPDK can only work on NXP SoCs as listed in the Supported DPAA2 SoCs.

Maximum packet length

The DPAA2 SoC family support a maximum of a 10240 jumbo frame. The value is fixed and cannot be changed. So, even when the rxmode.max_rx_pkt_len member of struct rte_eth_conf is set to a value lower than 10240, frames up to 10240 bytes can still reach the host interface.

Other Limitations
  • RSS hash key cannot be modified.
  • RSS RETA cannot be configured.
  • Secondary process packet I/O is not supported.

Driver for VM Emulated Devices

The DPDK EM poll mode driver supports the following emulated devices:

  • qemu-kvm emulated Intel® 82540EM Gigabit Ethernet Controller (qemu e1000 device)
  • VMware* emulated Intel® 82545EM Gigabit Ethernet Controller
  • VMware emulated Intel® 8274L Gigabit Ethernet Controller.

Validated Hypervisors

The validated hypervisors are:

  • KVM (Kernel Virtual Machine) with Qemu, version 0.14.0
  • KVM (Kernel Virtual Machine) with Qemu, version 0.15.1
  • VMware ESXi 5.0, Update 1

Setting Up a KVM Virtual Machine

The following describes a target environment:

  • Host Operating System: Fedora 14
  • Hypervisor: KVM (Kernel Virtual Machine) with Qemu version, 0.14.0
  • Guest Operating System: Fedora 14
  • Linux Kernel Version: Refer to the DPDK Getting Started Guide
  • Target Applications: testpmd

The setup procedure is as follows:

  1. Download qemu-kvm-0.14.0 from http://sourceforge.net/projects/kvm/files/qemu-kvm/ and install it in the Host OS using the following steps:

    When using a recent kernel (2.6.25+) with kvm modules included:

    tar xzf qemu-kvm-release.tar.gz cd qemu-kvm-release
    ./configure --prefix=/usr/local/kvm
    make
    sudo make install
    sudo /sbin/modprobe kvm-intel
    

    When using an older kernel or a kernel from a distribution without the kvm modules, you must download (from the same link), compile and install the modules yourself:

    tar xjf kvm-kmod-release.tar.bz2
    cd kvm-kmod-release
    ./configure
    make
    sudo make install
    sudo /sbin/modprobe kvm-intel
    

    Note that qemu-kvm installs in the /usr/local/bin directory.

    For more details about KVM configuration and usage, please refer to: http://www.linux-kvm.org/page/HOWTO1.

  2. Create a Virtual Machine and install Fedora 14 on the Virtual Machine. This is referred to as the Guest Operating System (Guest OS).

  3. Start the Virtual Machine with at least one emulated e1000 device.

    Note

    The Qemu provides several choices for the emulated network device backend. Most commonly used is a TAP networking backend that uses a TAP networking device in the host. For more information about Qemu supported networking backends and different options for configuring networking at Qemu, please refer to:

    http://www.linux-kvm.org/page/Networking

    http://wiki.qemu.org/Documentation/Networking

    http://qemu.weilnetz.de/qemu-doc.html

    For example, to start a VM with two emulated e1000 devices, issue the following command:

    /usr/local/kvm/bin/qemu-system-x86_64 -cpu host -smp 4 -hda qemu1.raw -m 1024
    -net nic,model=e1000,vlan=1,macaddr=DE:AD:1E:00:00:01
    -net tap,vlan=1,ifname=tapvm01,script=no,downscript=no
    -net nic,model=e1000,vlan=2,macaddr=DE:AD:1E:00:00:02
    -net tap,vlan=2,ifname=tapvm02,script=no,downscript=no
    

    where:

    — -m = memory to assign

    — -smp = number of smp cores

    — -hda = virtual disk image

    This command starts a new virtual machine with two emulated 82540EM devices, backed up with two TAP networking host interfaces, tapvm01 and tapvm02.

    # ip tuntap show
    tapvm01: tap
    tapvm02: tap
    
  4. Configure your TAP networking interfaces using ip/ifconfig tools.

  5. Log in to the guest OS and check that the expected emulated devices exist:

    # lspci -d 8086:100e
    00:04.0 Ethernet controller: Intel Corporation 82540EM Gigabit Ethernet Controller (rev 03)
    00:05.0 Ethernet controller: Intel Corporation 82540EM Gigabit Ethernet Controller (rev 03)
    
  6. Install the DPDK and run testpmd.

Known Limitations of Emulated Devices

The following are known limitations:

  1. The Qemu e1000 RX path does not support multiple descriptors/buffers per packet. Therefore, rte_mbuf should be big enough to hold the whole packet. For example, to allow testpmd to receive jumbo frames, use the following:

    testpmd [options] – –mbuf-size=<your-max-packet-size>

  2. Qemu e1000 does not validate the checksum of incoming packets.

  3. Qemu e1000 only supports one interrupt source, so link and Rx interrupt should be exclusive.

  4. Qemu e1000 does not support interrupt auto-clear, application should disable interrupt immediately when woken up.

ENA Poll Mode Driver

The ENA PMD is a DPDK poll-mode driver for the Amazon Elastic Network Adapter (ENA) family.

Overview

The ENA driver exposes a lightweight management interface with a minimal set of memory mapped registers and an extendable command set through an Admin Queue.

The driver supports a wide range of ENA adapters, is link-speed independent (i.e., the same driver is used for 10GbE, 25GbE, 40GbE, etc.), and it negotiates and supports an extendable feature set.

ENA adapters allow high speed and low overhead Ethernet traffic processing by providing a dedicated Tx/Rx queue pair per CPU core.

The ENA driver supports industry standard TCP/IP offload features such as checksum offload and TCP transmit segmentation offload (TSO).

Receive-side scaling (RSS) is supported for multi-core scaling.

Some of the ENA devices support a working mode called Low-latency Queue (LLQ), which saves several more microseconds.

Management Interface

ENA management interface is exposed by means of:

  • Device Registers
  • Admin Queue (AQ) and Admin Completion Queue (ACQ)

ENA device memory-mapped PCIe space for registers (MMIO registers) are accessed only during driver initialization and are not involved in further normal device operation.

AQ is used for submitting management commands, and the results/responses are reported asynchronously through ACQ.

ENA introduces a very small set of management commands with room for vendor-specific extensions. Most of the management operations are framed in a generic Get/Set feature command.

The following admin queue commands are supported:

  • Create I/O submission queue
  • Create I/O completion queue
  • Destroy I/O submission queue
  • Destroy I/O completion queue
  • Get feature
  • Set feature
  • Get statistics

Refer to ena_admin_defs.h for the list of supported Get/Set Feature properties.

Data Path Interface

I/O operations are based on Tx and Rx Submission Queues (Tx SQ and Rx SQ correspondingly). Each SQ has a completion queue (CQ) associated with it.

The SQs and CQs are implemented as descriptor rings in contiguous physical memory.

Refer to ena_eth_io_defs.h for the detailed structure of the descriptor

The driver supports multi-queue for both Tx and Rx.

Configuration information

DPDK Configuration Parameters

The following configuration options are available for the ENA PMD:

  • CONFIG_RTE_LIBRTE_ENA_PMD (default y): Enables or disables inclusion of the ENA PMD driver in the DPDK compilation.
  • CONFIG_RTE_LIBRTE_ENA_DEBUG_INIT (default y): Enables or disables debug logging of device initialization within the ENA PMD driver.
  • CONFIG_RTE_LIBRTE_ENA_DEBUG_RX (default n): Enables or disables debug logging of RX logic within the ENA PMD driver.
  • CONFIG_RTE_LIBRTE_ENA_DEBUG_TX (default n): Enables or disables debug logging of TX logic within the ENA PMD driver.
  • CONFIG_RTE_LIBRTE_ENA_COM_DEBUG (default n): Enables or disables debug logging of low level tx/rx logic in ena_com(base) within the ENA PMD driver.

ENA Configuration Parameters

  • Number of Queues

    This is the requested number of queues upon initialization, however, the actual number of receive and transmit queues to be created will be the minimum between the maximal number supported by the device and number of queues requested.

  • Size of Queues

    This is the requested size of receive/transmit queues, while the actual size will be the minimum between the requested size and the maximal receive/transmit supported by the device.

Building DPDK

See the DPDK Getting Started Guide for Linux for instructions on how to build DPDK.

By default the ENA PMD library will be built into the DPDK library.

For configuring and using UIO and VFIO frameworks, please also refer the documentation that comes with DPDK suite.

Supported ENA adapters

Current ENA PMD supports the following ENA adapters including:

  • 1d0f:ec20 - ENA VF
  • 1d0f:ec21 - ENA VF with LLQ support

Supported Operating Systems

Any Linux distribution fulfilling the conditions described in System Requirements section of the DPDK documentation or refer to DPDK Release Notes.

Supported features

  • Jumbo frames up to 9K
  • Port Hardware Statistics
  • IPv4/TCP/UDP checksum offload
  • TSO offload
  • Multiple receive and transmit queues
  • RSS
  • Low Latency Queue for Tx

Unsupported features

The features supported by the device and not yet supported by this PMD include:

  • Asynchronous Event Notification Queue (AENQ)

Prerequisites

  1. Prepare the system as recommended by DPDK suite. This includes environment variables, hugepages configuration, tool-chains and configuration
  2. Insert igb_uio kernel module using the command ‘modprobe igb_uio’
  3. Bind the intended ENA device to igb_uio module

At this point the system should be ready to run DPDK applications. Once the application runs to completion, the ENA can be detached from igb_uio if necessary.

Usage example

Follow instructions available in the document compiling and testing a PMD for a NIC to launch testpmd with Amazon ENA devices managed by librte_pmd_ena.

Example output:

[...]
EAL: PCI device 0000:02:00.1 on NUMA socket -1
EAL:   probe driver: 1d0f:ec20 rte_ena_pmd
EAL:   PCI memory mapped at 0x7f9b6c400000
PMD: eth_ena_dev_init(): Initializing 0:2:0.1
Interactive-mode selected
Configuring Port 0 (socket 0)
Port 0: 00:00:00:11:00:01
Checking link statuses...
Port 0 Link Up - speed 10000 Mbps - full-duplex
Done
testpmd>

ENIC Poll Mode Driver

ENIC PMD is the DPDK poll-mode driver for the Cisco System Inc. VIC Ethernet NICs. These adapters are also referred to as vNICs below. If you are running or would like to run DPDK software applications on Cisco UCS servers using Cisco VIC adapters the following documentation is relevant.

How to obtain ENIC PMD integrated DPDK

ENIC PMD support is integrated into the DPDK suite. dpdk-<version>.tar.gz should be downloaded from http://dpdk.org

Configuration information

  • DPDK Configuration Parameters

    The following configuration options are available for the ENIC PMD:

    • CONFIG_RTE_LIBRTE_ENIC_PMD (default y): Enables or disables inclusion of the ENIC PMD driver in the DPDK compilation.
    • CONFIG_RTE_LIBRTE_ENIC_DEBUG (default n): Enables or disables debug logging within the ENIC PMD driver.
  • vNIC Configuration Parameters

    • Number of Queues

      The maximum number of receive queues (RQs), work queues (WQs) and completion queues (CQs) are configurable on a per vNIC basis through the Cisco UCS Manager (CIMC or UCSM).

      These values should be configured as follows:

      • The number of WQs should be greater or equal to the value of the expected nb_tx_q parameter in the call to the rte_eth_dev_configure()
      • The number of RQs configured in the vNIC should be greater or equal to twice the value of the expected nb_rx_q parameter in the call to rte_eth_dev_configure(). With the addition of Rx scatter, a pair of RQs on the vnic is needed for each receive queue used by DPDK, even if Rx scatter is not being used. Having a vNIC with only 1 RQ is not a valid configuration, and will fail with an error message.
      • The number of CQs should set so that there is one CQ for each WQ, and one CQ for each pair of RQs.

      For example: If the application requires 3 Rx queues, and 3 Tx queues, the vNIC should be configured to have at least 3 WQs, 6 RQs (3 pairs), and 6 CQs (3 for use by WQs + 3 for use by the 3 pairs of RQs).

    • Size of Queues

      Likewise, the number of receive and transmit descriptors are configurable on a per vNIC bases via the UCS Manager and should be greater than or equal to the nb_rx_desc and nb_tx_desc parameters expected to be used in the calls to rte_eth_rx_queue_setup() and rte_eth_tx_queue_setup() respectively. An application requesting more than the set size will be limited to that size.

      Unless there is a lack of resources due to creating many vNICs, it is recommended that the WQ and RQ sizes be set to the maximum. This gives the application the greatest amount of flexibility in its queue configuration.

      • Note: Since the introduction of Rx scatter, for performance reasons, this PMD uses two RQs on the vNIC per receive queue in DPDK. One RQ holds descriptors for the start of a packet the second RQ holds the descriptors for the rest of the fragments of a packet. This means that the nb_rx_desc parameter to rte_eth_rx_queue_setup() can be a greater than 4096. The exact amount will depend on the size of the mbufs being used for receives, and the MTU size.

        For example: If the mbuf size is 2048, and the MTU is 9000, then receiving a full size packet will take 5 descriptors, 1 from the start of packet queue, and 4 from the second queue. Assuming that the RQ size was set to the maximum of 4096, then the application can specify up to 1024 + 4096 as the nb_rx_desc parameter to rte_eth_rx_queue_setup().

    • Interrupts

      Only one interrupt per vNIC interface should be configured in the UCS manager regardless of the number receive/transmit queues. The ENIC PMD uses this interrupt to get information about link status and errors in the fast path.

Flow director support

Advanced filtering support was added to 1300 series VIC firmware starting with version 2.0.13 for C-series UCS servers and version 3.1.2 for UCSM managed blade servers. In order to enable advanced filtering the ‘Advanced filter’ radio button should be enabled via CIMC or UCSM followed by a reboot of the server.

With advanced filters, perfect matching of all fields of IPv4, IPv6 headers as well as TCP, UDP and SCTP L4 headers is available through flow director. Masking of these fields for partial match is also supported.

Without advanced filter support, the flow director is limited to IPv4 perfect filtering of the 5-tuple with no masking of fields supported.

SR-IOV mode utilization

UCS blade servers configured with dynamic vNIC connection policies in UCS manager are capable of supporting assigned devices on virtual machines (VMs) through a KVM hypervisor. Assigned devices, also known as ‘passthrough’ devices, are SR-IOV virtual functions (VFs) on the host which are exposed to VM instances.

The Cisco Virtual Machine Fabric Extender (VM-FEX) gives the VM a dedicated interface on the Fabric Interconnect (FI). Layer 2 switching is done at the FI. This may eliminate the requirement for software switching on the host to route intra-host VM traffic.

Please refer to Creating a Dynamic vNIC Connection Policy for information on configuring SR-IOV Adapter policies using UCS manager.

Once the policies are in place and the host OS is rebooted, VFs should be visible on the host, E.g.:

# lspci | grep Cisco | grep Ethernet
0d:00.0 Ethernet controller: Cisco Systems Inc VIC Ethernet NIC (rev a2)
0d:00.1 Ethernet controller: Cisco Systems Inc VIC SR-IOV VF (rev a2)
0d:00.2 Ethernet controller: Cisco Systems Inc VIC SR-IOV VF (rev a2)
0d:00.3 Ethernet controller: Cisco Systems Inc VIC SR-IOV VF (rev a2)
0d:00.4 Ethernet controller: Cisco Systems Inc VIC SR-IOV VF (rev a2)
0d:00.5 Ethernet controller: Cisco Systems Inc VIC SR-IOV VF (rev a2)
0d:00.6 Ethernet controller: Cisco Systems Inc VIC SR-IOV VF (rev a2)
0d:00.7 Ethernet controller: Cisco Systems Inc VIC SR-IOV VF (rev a2)

Enable Intel IOMMU on the host and install KVM and libvirt. A VM instance should be created with an assigned device. When using libvirt, this configuration can be done within the domain (i.e. VM) config file. For example this entry maps host VF 0d:00:01 into the VM.

<interface type='hostdev' managed='yes'>
  <mac address='52:54:00:ac:ff:b6'/>
  <source>
    <address type='pci' domain='0x0000' bus='0x0d' slot='0x00' function='0x1'/>
  </source>

Alternatively, the configuration can be done in a separate file using the network keyword. These methods are described in the libvirt documentation for Network XML format.

When the VM instance is started, the ENIC KVM driver will bind the host VF to vfio, complete provisioning on the FI and bring up the link.

Note

It is not possible to use a VF directly from the host because it is not fully provisioned until the hypervisor brings up the VM that it is assigned to.

In the VM instance, the VF will now be visible. E.g., here the VF 00:04.0 is seen on the VM instance and should be available for binding to a DPDK.

# lspci | grep Ether
00:04.0 Ethernet controller: Cisco Systems Inc VIC SR-IOV VF (rev a2)

Follow the normal DPDK install procedure, binding the VF to either igb_uio or vfio in non-IOMMU mode.

Please see Limitations for limitations in the use of SR-IOV.

Generic Flow API support

Generic Flow API is supported. The baseline support is:

  • 1200 series VICs

    5-tuple exact Flow support for 1200 series adapters. This allows:

    • Attributes: ingress
    • Items: ipv4, ipv6, udp, tcp (must exactly match src/dst IP addresses and ports and all must be specified).
    • Actions: queue and void
    • Selectors: ‘is’
  • 1300 series VICS with Advanced filters disabled

    With advanced filters disabled, an IPv4 or IPv6 item must be specified in the pattern.

    • Attributes: ingress
    • Items: eth, ipv4, ipv6, udp, tcp, vxlan, inner eth, ipv4, ipv6, udp, tcp
    • Actions: queue and void
    • Selectors: ‘is’, ‘spec’ and ‘mask’. ‘last’ is not supported
    • In total, up to 64 bytes of mask is allowed across all haeders
  • 1300 series VICS with Advanced filters enabled

    • Attributes: ingress
    • Items: eth, ipv4, ipv6, udp, tcp, vxlan, inner eth, ipv4, ipv6, udp, tcp
    • Actions: queue, mark, flag and void
    • Selectors: ‘is’, ‘spec’ and ‘mask’. ‘last’ is not supported
    • In total, up to 64 bytes of mask is allowed across all haeders

More features may be added in future firmware and new versions of the VIC. Please refer to the release notes.

Limitations

  • VLAN 0 Priority Tagging

    If a vNIC is configured in TRUNK mode by the UCS manager, the adapter will priority tag egress packets according to 802.1Q if they were not already VLAN tagged by software. If the adapter is connected to a properly configured switch, there will be no unexpected behavior.

    In test setups where an Ethernet port of a Cisco adapter in TRUNK mode is connected point-to-point to another adapter port or connected though a router instead of a switch, all ingress packets will be VLAN tagged. Programs such as l3fwd which do not account for VLAN tags in packets will misbehave. The solution is to enable VLAN stripping on ingress. The follow code fragment is example of how to accomplish this:

vlan_offload = rte_eth_dev_get_vlan_offload(port);
vlan_offload |= ETH_VLAN_STRIP_OFFLOAD;
rte_eth_dev_set_vlan_offload(port, vlan_offload);
  • Limited flow director support on 1200 series and 1300 series Cisco VIC adapters with old firmware. Please see Flow director support.
  • Flow director features are not supported on generation 1 Cisco VIC adapters (M81KR and P81E)
  • SR-IOV
    • KVM hypervisor support only. VMware has not been tested.
    • Requires VM-FEX, and so is only available on UCS managed servers connected to Fabric Interconnects. It is not on standalone C-Series servers.
    • VF devices are not usable directly from the host. They can only be used as assigned devices on VM instances.
    • Currently, unbind of the ENIC kernel mode driver ‘enic.ko’ on the VM instance may hang. As a workaround, enic.ko should blacklisted or removed from the boot process.
    • pci_generic cannot be used as the uio module in the VM. igb_uio or vfio in non-IOMMU mode can be used.
    • The number of RQs in UCSM dynamic vNIC configurations must be at least 2.
    • The number of SR-IOV devices is limited to 256. Components on target system might limit this number to fewer than 256.
  • Flow API
    • The number of filters that can be specified with the Generic Flow API is dependent on how many header fields are being masked. Use ‘flow create’ in a loop to determine how many filters your VIC will support (not more than 1000 for 1300 series VICs). Filter are checked for matching in the order they were added. Since there currently is no grouping or priority support, ‘catch-all’ filters should be added last.

How to build the suite

The build instructions for the DPDK suite should be followed. By default the ENIC PMD library will be built into the DPDK library.

Refer to the document compiling and testing a PMD for a NIC for details.

By default the ENIC PMD library will be built into the DPDK library.

For configuring and using UIO and VFIO frameworks, please refer to the documentation that comes with DPDK suite.

Supported Cisco VIC adapters

ENIC PMD supports all recent generations of Cisco VIC adapters including:

  • VIC 1280
  • VIC 1240
  • VIC 1225
  • VIC 1285
  • VIC 1225T
  • VIC 1227
  • VIC 1227T
  • VIC 1380
  • VIC 1340
  • VIC 1385
  • VIC 1387

Supported Operating Systems

Any Linux distribution fulfilling the conditions described in Dependencies section of DPDK documentation.

Supported features

  • Unicast, multicast and broadcast transmission and reception
  • Receive queue polling
  • Port Hardware Statistics
  • Hardware VLAN acceleration
  • IP checksum offload
  • Receive side VLAN stripping
  • Multiple receive and transmit queues
  • Flow Director ADD, UPDATE, DELETE, STATS operation support IPv4 and IPv6
  • Promiscuous mode
  • Setting RX VLAN (supported via UCSM/CIMC only)
  • VLAN filtering (supported via UCSM/CIMC only)
  • Execution of application by unprivileged system users
  • IPV4, IPV6 and TCP RSS hashing
  • Scattered Rx
  • MTU update
  • SR-IOV on UCS managed servers connected to Fabric Interconnects.
  • Flow API

Known bugs and unsupported features in this release

  • Signature or flex byte based flow direction
  • Drop feature of flow direction
  • VLAN based flow direction
  • non-IPV4 flow direction
  • Setting of extended VLAN
  • UDP RSS hashing
  • MTU update only works if Scattered Rx mode is disabled

Prerequisites

  • Prepare the system as recommended by DPDK suite. This includes environment variables, hugepages configuration, tool-chains and configuration
  • Insert vfio-pci kernel module using the command ‘modprobe vfio-pci’ if the user wants to use VFIO framework
  • Insert uio kernel module using the command ‘modprobe uio’ if the user wants to use UIO framework
  • DPDK suite should be configured based on the user’s decision to use VFIO or UIO framework
  • If the vNIC device(s) to be used is bound to the kernel mode Ethernet driver use ‘ifconfig’ to bring the interface down. The dpdk-devbind.py tool can then be used to unbind the device’s bus id from the ENIC kernel mode driver.
  • Bind the intended vNIC to vfio-pci in case the user wants ENIC PMD to use VFIO framework using dpdk-devbind.py.
  • Bind the intended vNIC to igb_uio in case the user wants ENIC PMD to use UIO framework using dpdk-devbind.py.

At this point the system should be ready to run DPDK applications. Once the application runs to completion, the vNIC can be detached from vfio-pci or igb_uio if necessary.

Root privilege is required to bind and unbind vNICs to/from VFIO/UIO. VFIO framework helps an unprivileged user to run the applications. For an unprivileged user to run the applications on DPDK and ENIC PMD, it may be necessary to increase the maximum locked memory of the user. The following command could be used to do this.

sudo sh -c "ulimit -l <value in Kilo Bytes>"

The value depends on the memory configuration of the application, DPDK and PMD. Typically, the limit has to be raised to higher than 2GB. e.g., 2621440

The compilation of any unused drivers can be disabled using the configuration file in config/ directory (e.g., config/common_linuxapp). This would help in bringing down the time taken for building the libraries and the initialization time of the application.

Contact Information

Any questions or bugs should be reported to DPDK community and to the ENIC PMD maintainers:

FM10K Poll Mode Driver

The FM10K poll mode driver library provides support for the Intel FM10000 (FM10K) family of 40GbE/100GbE adapters.

FTAG Based Forwarding of FM10K

FTAG Based Forwarding is a unique feature of FM10K. The FM10K family of NICs support the addition of a Fabric Tag (FTAG) to carry special information. The FTAG is placed at the beginning of the frame, it contains information such as where the packet comes from and goes, and the vlan tag. In FTAG based forwarding mode, the switch logic forwards packets according to glort (global resource tag) information, rather than the mac and vlan table. Currently this feature works only on PF.

To enable this feature, the user should pass a devargs parameter to the eal like “-w 84:00.0,enable_ftag=1”, and the application should make sure an appropriate FTAG is inserted for every frame on TX side.

Vector PMD for FM10K

Vector PMD (vPMD) uses Intel® SIMD instructions to optimize packet I/O. It improves load/store bandwidth efficiency of L1 data cache by using a wider SSE/AVX ‘’register (1)’‘. The wider register gives space to hold multiple packet buffers so as to save on the number of instructions when bulk processing packets.

There is no change to the PMD API. The RX/TX handlers are the only two entries for vPMD packet I/O. They are transparently registered at runtime RX/TX execution if all required conditions are met.

  1. To date, only an SSE version of FM10K vPMD is available. To ensure that vPMD is in the binary code, set CONFIG_RTE_LIBRTE_FM10K_INC_VECTOR=y in the configure file.

Some constraints apply as pre-conditions for specific optimizations on bulk packet transfers. The following sections explain RX and TX constraints in the vPMD.

RX Constraints
Prerequisites and Pre-conditions

For Vector RX it is assumed that the number of descriptor rings will be a power of 2. With this pre-condition, the ring pointer can easily scroll back to the head after hitting the tail without a conditional check. In addition Vector RX can use this assumption to do a bit mask using ring_size - 1.

Features not Supported by Vector RX PMD

Some features are not supported when trying to increase the throughput in vPMD. They are:

  • IEEE1588
  • Flow director
  • Header split
  • RX checksum offload

Other features are supported using optional MACRO configuration. They include:

  • HW VLAN strip
  • L3/L4 packet type

To enable via RX_OLFLAGS use RTE_LIBRTE_FM10K_RX_OLFLAGS_ENABLE=y.

To guarantee the constraint, the following configuration flags in dev_conf.rxmode will be checked:

  • hw_vlan_extend
  • hw_ip_checksum
  • header_split
  • fdir_conf->mode
RX Burst Size

As vPMD is focused on high throughput, it processes 4 packets at a time. So it assumes that the RX burst should be greater than 4 packets per burst. It returns zero if using nb_pkt < 4 in the receive handler. If nb_pkt is not a multiple of 4, a floor alignment will be applied.

TX Constraint
Features not Supported by TX Vector PMD

TX vPMD only works when txq_flags is set to FM10K_SIMPLE_TX_FLAG. This means that it does not support TX multi-segment, VLAN offload or TX csum offload. The following MACROs are used for these three features:

  • ETH_TXQ_FLAGS_NOMULTSEGS
  • ETH_TXQ_FLAGS_NOVLANOFFL
  • ETH_TXQ_FLAGS_NOXSUMSCTP
  • ETH_TXQ_FLAGS_NOXSUMUDP
  • ETH_TXQ_FLAGS_NOXSUMTCP

Limitations

Switch manager

The Intel FM10000 family of NICs integrate a hardware switch and multiple host interfaces. The FM10000 PMD driver only manages host interfaces. For the switch component another switch driver has to be loaded prior to to the FM10000 PMD driver. The switch driver can be acquired from Intel support. Only Testpoint is validated with DPDK, the latest version that has been validated with DPDK is 4.1.6.

Support for Switch Restart

For FM10000 multi host based design a DPDK app running in the VM or host needs to be aware of the switch’s state since it may undergo a quit-restart. When the switch goes down the DPDK app will receive a LSC event indicating link status down, and the app should stop the worker threads that are polling on the Rx/Tx queues. When switch comes up, a LSC event indicating LINK_UP is sent to the app, which can then restart the FM10000 port to resume network processing.

CRC striping

The FM10000 family of NICs strip the CRC for every packets coming into the host interface. So, CRC will be stripped even when the rxmode.hw_strip_crc member is set to 0 in struct rte_eth_conf.

Maximum packet length

The FM10000 family of NICS support a maximum of a 15K jumbo frame. The value is fixed and cannot be changed. So, even when the rxmode.max_rx_pkt_len member of struct rte_eth_conf is set to a value lower than 15364, frames up to 15364 bytes can still reach the host interface.

Statistic Polling Frequency

The FM10000 NICs expose a set of statistics via the PCI BARs. These statistics are read from the hardware registers when rte_eth_stats_get() or rte_eth_xstats_get() is called. The packet counting registers are 32 bits while the byte counting registers are 48 bits. As a result, the statistics must be polled regularly in order to ensure the consistency of the returned reads.

Given the PCIe Gen3 x8, about 50Gbps of traffic can occur. With 64 byte packets this gives almost 100 million packets/second, causing 32 bit integer overflow after approx 40 seconds. To ensure these overflows are detected and accounted for in the statistics, it is necessary to read statistic regularly. It is suggested to read stats every 20 seconds, which will ensure the statistics are accurate.

Interrupt mode

The FM10000 family of NICS need one separate interrupt for mailbox. So only drivers which support multiple interrupt vectors e.g. vfio-pci can work for fm10k interrupt mode.

I40E Poll Mode Driver

The I40E PMD (librte_pmd_i40e) provides poll mode driver support for the Intel X710/XL710/X722 10/40 Gbps family of adapters.

Features

Features of the I40E PMD are:

  • Multiple queues for TX and RX
  • Receiver Side Scaling (RSS)
  • MAC/VLAN filtering
  • Packet type information
  • Flow director
  • Cloud filter
  • Checksum offload
  • VLAN/QinQ stripping and inserting
  • TSO offload
  • Promiscuous mode
  • Multicast mode
  • Port hardware statistics
  • Jumbo frames
  • Link state information
  • Link flow control
  • Mirror on port, VLAN and VSI
  • Interrupt mode for RX
  • Scattered and gather for TX and RX
  • Vector Poll mode driver
  • DCB
  • VMDQ
  • SR-IOV VF
  • Hot plug
  • IEEE1588/802.1AS timestamping
  • VF Daemon (VFD) - EXPERIMENTAL

Prerequisites

Pre-Installation Configuration

Config File Options

The following options can be modified in the config file. Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_LIBRTE_I40E_PMD (default y)

    Toggle compilation of the librte_pmd_i40e driver.

  • CONFIG_RTE_LIBRTE_I40E_DEBUG_* (default n)

    Toggle display of generic debugging messages.

  • CONFIG_RTE_LIBRTE_I40E_RX_ALLOW_BULK_ALLOC (default y)

    Toggle bulk allocation for RX.

  • CONFIG_RTE_LIBRTE_I40E_INC_VECTOR (default n)

    Toggle the use of Vector PMD instead of normal RX/TX path. To enable vPMD for RX, bulk allocation for Rx must be allowed.

  • CONFIG_RTE_LIBRTE_I40E_16BYTE_RX_DESC (default n)

    Toggle to use a 16-byte RX descriptor, by default the RX descriptor is 32 byte.

  • CONFIG_RTE_LIBRTE_I40E_QUEUE_NUM_PER_PF (default 64)

    Number of queues reserved for PF.

  • CONFIG_RTE_LIBRTE_I40E_QUEUE_NUM_PER_VF (default 4)

    Number of queues reserved for each SR-IOV VF.

  • CONFIG_RTE_LIBRTE_I40E_QUEUE_NUM_PER_VM (default 4)

    Number of queues reserved for each VMDQ Pool.

  • CONFIG_RTE_LIBRTE_I40E_ITR_INTERVAL (default -1)

    Interrupt Throttling interval.

Driver compilation and testing

Refer to the document compiling and testing a PMD for a NIC for details.

SR-IOV: Prerequisites and sample Application Notes

  1. Load the kernel module:

    modprobe i40e
    

    Check the output in dmesg:

    i40e 0000:83:00.1 ens802f0: renamed from eth0
    
  2. Bring up the PF ports:

    ifconfig ens802f0 up
    
  3. Create VF device(s):

    Echo the number of VFs to be created into the sriov_numvfs sysfs entry of the parent PF.

    Example:

    echo 2 > /sys/devices/pci0000:00/0000:00:03.0/0000:81:00.0/sriov_numvfs
    
  4. Assign VF MAC address:

    Assign MAC address to the VF using iproute2 utility. The syntax is:

    ip link set <PF netdev id> vf <VF id> mac <macaddr>
    

    Example:

    ip link set ens802f0 vf 0 mac a0:b0:c0:d0:e0:f0
    
  5. Assign VF to VM, and bring up the VM. Please see the documentation for the I40E/IXGBE/IGB Virtual Function Driver.

  6. Running testpmd:

    Follow instructions available in the document compiling and testing a PMD for a NIC to run testpmd.

    Example output:

    ...
    EAL: PCI device 0000:83:00.0 on NUMA socket 1
    EAL: probe driver: 8086:1572 rte_i40e_pmd
    EAL: PCI memory mapped at 0x7f7f80000000
    EAL: PCI memory mapped at 0x7f7f80800000
    PMD: eth_i40e_dev_init(): FW 5.0 API 1.5 NVM 05.00.02 eetrack 8000208a
    Interactive-mode selected
    Configuring Port 0 (socket 0)
    ...
    
    PMD: i40e_dev_rx_queue_setup(): Rx Burst Bulk Alloc Preconditions are
    satisfied.Rx Burst Bulk Alloc function will be used on port=0, queue=0.
    
    ...
    Port 0: 68:05:CA:26:85:84
    Checking link statuses...
    Port 0 Link Up - speed 10000 Mbps - full-duplex
    Done
    
    testpmd>
    

Sample Application Notes

Vlan filter

Vlan filter only works when Promiscuous mode is off.

To start testpmd, and add vlan 10 to port 0:

./app/testpmd -l 0-15 -n 4 -- -i --forward-mode=mac
...

testpmd> set promisc 0 off
testpmd> rx_vlan add 10 0
Flow Director

The Flow Director works in receive mode to identify specific flows or sets of flows and route them to specific queues. The Flow Director filters can match the different fields for different type of packet: flow type, specific input set per flow type and the flexible payload.

The default input set of each flow type is:

ipv4-other : src_ip_address, dst_ip_address
ipv4-frag  : src_ip_address, dst_ip_address
ipv4-tcp   : src_ip_address, dst_ip_address, src_port, dst_port
ipv4-udp   : src_ip_address, dst_ip_address, src_port, dst_port
ipv4-sctp  : src_ip_address, dst_ip_address, src_port, dst_port,
             verification_tag
ipv6-other : src_ip_address, dst_ip_address
ipv6-frag  : src_ip_address, dst_ip_address
ipv6-tcp   : src_ip_address, dst_ip_address, src_port, dst_port
ipv6-udp   : src_ip_address, dst_ip_address, src_port, dst_port
ipv6-sctp  : src_ip_address, dst_ip_address, src_port, dst_port,
             verification_tag
l2_payload : ether_type

The flex payload is selected from offset 0 to 15 of packet’s payload by default, while it is masked out from matching.

Start testpmd with --disable-rss and --pkt-filter-mode=perfect:

./app/testpmd -l 0-15 -n 4 -- -i --disable-rss --pkt-filter-mode=perfect \
              --rxq=8 --txq=8 --nb-cores=8 --nb-ports=1

Add a rule to direct ipv4-udp packet whose dst_ip=2.2.2.5, src_ip=2.2.2.3, src_port=32, dst_port=32 to queue 1:

testpmd> flow_director_filter 0 mode IP add flow ipv4-udp  \
         src 2.2.2.3 32 dst 2.2.2.5 32 vlan 0 flexbytes () \
         fwd pf queue 1 fd_id 1

Check the flow director status:

testpmd> show port fdir 0

######################## FDIR infos for port 0      ####################
  MODE:   PERFECT
  SUPPORTED FLOW TYPE:  ipv4-frag ipv4-tcp ipv4-udp ipv4-sctp ipv4-other
                        ipv6-frag ipv6-tcp ipv6-udp ipv6-sctp ipv6-other
                        l2_payload
  FLEX PAYLOAD INFO:
  max_len:       16          payload_limit: 480
  payload_unit:  2           payload_seg:   3
  bitmask_unit:  2           bitmask_num:   2
  MASK:
    vlan_tci: 0x0000,
    src_ipv4: 0x00000000,
    dst_ipv4: 0x00000000,
    src_port: 0x0000,
    dst_port: 0x0000
    src_ipv6: 0x00000000,0x00000000,0x00000000,0x00000000,
    dst_ipv6: 0x00000000,0x00000000,0x00000000,0x00000000
  FLEX PAYLOAD SRC OFFSET:
    L2_PAYLOAD:    0      1      2      3      4      5      6  ...
    L3_PAYLOAD:    0      1      2      3      4      5      6  ...
    L4_PAYLOAD:    0      1      2      3      4      5      6  ...
  FLEX MASK CFG:
    ipv4-udp:    00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
    ipv4-tcp:    00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
    ipv4-sctp:   00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
    ipv4-other:  00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
    ipv4-frag:   00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
    ipv6-udp:    00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
    ipv6-tcp:    00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
    ipv6-sctp:   00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
    ipv6-other:  00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
    ipv6-frag:   00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
    l2_payload:  00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
  guarant_count: 1           best_count:    0
  guarant_space: 512         best_space:    7168
  collision:     0           free:          0
  maxhash:       0           maxlen:        0
  add:           0           remove:        0
  f_add:         0           f_remove:      0

Delete all flow director rules on a port:

testpmd> flush_flow_director 0
Floating VEB

The Intel® Ethernet Controller X710 and XL710 Family support a feature called “Floating VEB”.

A Virtual Ethernet Bridge (VEB) is an IEEE Edge Virtual Bridging (EVB) term for functionality that allows local switching between virtual endpoints within a physical endpoint and also with an external bridge/network.

A “Floating” VEB doesn’t have an uplink connection to the outside world so all switching is done internally and remains within the host. As such, this feature provides security benefits.

In addition, a Floating VEB overcomes a limitation of normal VEBs where they cannot forward packets when the physical link is down. Floating VEBs don’t need to connect to the NIC port so they can still forward traffic from VF to VF even when the physical link is down.

Therefore, with this feature enabled VFs can be limited to communicating with each other but not an outside network, and they can do so even when there is no physical uplink on the associated NIC port.

To enable this feature, the user should pass a devargs parameter to the EAL, for example:

-w 84:00.0,enable_floating_veb=1

In this configuration the PMD will use the floating VEB feature for all the VFs created by this PF device.

Alternatively, the user can specify which VFs need to connect to this floating VEB using the floating_veb_list argument:

-w 84:00.0,enable_floating_veb=1,floating_veb_list=1;3-4

In this example VF1, VF3 and VF4 connect to the floating VEB, while other VFs connect to the normal VEB.

The current implementation only supports one floating VEB and one regular VEB. VFs can connect to a floating VEB or a regular VEB according to the configuration passed on the EAL command line.

The floating VEB functionality requires a NIC firmware version of 5.0 or greater.

Limitations or Known issues

MPLS packet classification on X710/XL710

For firmware versions prior to 5.0, MPLS packets are not recognized by the NIC. The L2 Payload flow type in flow director can be used to classify MPLS packet by using a command in testpmd like:

testpmd> flow_director_filter 0 mode IP add flow l2_payload ether
0x8847 flexbytes () fwd pf queue <N> fd_id <M>

With the NIC firmware version 5.0 or greater, some limited MPLS support is added: Native MPLS (MPLS in Ethernet) skip is implemented, while no new packet type, no classification or offload are possible. With this change, L2 Payload flow type in flow director cannot be used to classify MPLS packet as with previous firmware versions. Meanwhile, the Ethertype filter can be used to classify MPLS packet by using a command in testpmd like:

testpmd> ethertype_filter 0 add mac_ignr 00:00:00:00:00:00 ethertype
0x8847 fwd queue <M>
16 Byte RX Descriptor setting on DPDK VF

Currently the VF’s RX descriptor mode is decided by PF. There’s no PF-VF interface for VF to request the RX descriptor mode, also no interface to notify VF its own RX descriptor mode. For all available versions of the i40e driver, these drivers don’t support 16 byte RX descriptor. If the Linux i40e kernel driver is used as host driver, while DPDK i40e PMD is used as the VF driver, DPDK cannot choose 16 byte receive descriptor. The reason is that the RX descriptor is already set to 32 byte by the i40e kernel driver. That is to say, user should keep CONFIG_RTE_LIBRTE_I40E_16BYTE_RX_DESC=n in config file. In the future, if the Linux i40e driver supports 16 byte RX descriptor, user should make sure the DPDK VF uses the same RX descriptor mode, 16 byte or 32 byte, as the PF driver.

The same rule for DPDK PF + DPDK VF. The PF and VF should use the same RX descriptor mode. Or the VF RX will not work.

Receive packets with Ethertype 0x88A8

Due to the FW limitation, PF can receive packets with Ethertype 0x88A8 only when floating VEB is disabled.

Incorrect Rx statistics when packet is oversize

When a packet is over maximum frame size, the packet is dropped. However the Rx statistics, when calling rte_eth_stats_get incorrectly shows it as received.

VF & TC max bandwidth setting

The per VF max bandwidth and per TC max bandwidth cannot be enabled in parallel. The dehavior is different when handling per VF and per TC max bandwidth setting. When enabling per VF max bandwidth, SW will check if per TC max bandwidth is enabled. If so, return failure. When enabling per TC max bandwidth, SW will check if per VF max bandwidth is enabled. If so, disable per VF max bandwidth and continue with per TC max bandwidth setting.

TC TX scheduling mode setting

There’re 2 TX scheduling modes for TCs, round robin and strict priority mode. If a TC is set to strict priority mode, it can consume unlimited bandwidth. It means if APP has set the max bandwidth for that TC, it comes to no effect. It’s suggested to set the strict priority mode for a TC that is latency sensitive but no consuming much bandwidth.

VF performance is impacted by PCI extended tag setting

To reach maximum NIC performance in the VF the PCI extended tag must be enabled. The DPDK I40E PF driver will set this feature during initialization, but the kernel PF driver does not. So when running traffic on a VF which is managed by the kernel PF driver, a significant NIC performance downgrade has been observed (for 64 byte packets, there is about 25% linerate downgrade for a 25G device and about 35% for a 40G device).

For kernel version >= 4.11, the kernel’s PCI driver will enable the extended tag if it detects that the device supports it. So by default, this is not an issue. For kernels <= 4.11 or when the PCI extended tag is disabled it can be enabled using the steps below.

  1. Get the current value of the PCI configure register:

    setpci -s <XX:XX.X> a8.w
    
  2. Set bit 8:

    value = value | 0x100
    
  3. Set the PCI configure register with new value:

    setpci -s <XX:XX.X> a8.w=<value>
    
Vlan strip of VF

The VF vlan strip function is only supported in the i40e kernel driver >= 2.1.26.

High Performance of Small Packets on 40G NIC

As there might be firmware fixes for performance enhancement in latest version of firmware image, the firmware update might be needed for getting high performance. Check with the local Intel’s Network Division application engineers for firmware updates. Users should consult the release notes specific to a DPDK release to identify the validated firmware version for a NIC using the i40e driver.

Use 16 Bytes RX Descriptor Size

As i40e PMD supports both 16 and 32 bytes RX descriptor sizes, and 16 bytes size can provide helps to high performance of small packets. Configuration of CONFIG_RTE_LIBRTE_I40E_16BYTE_RX_DESC in config files can be changed to use 16 bytes size RX descriptors.

High Performance and per Packet Latency Tradeoff

Due to the hardware design, the interrupt signal inside NIC is needed for per packet descriptor write-back. The minimum interval of interrupts could be set at compile time by CONFIG_RTE_LIBRTE_I40E_ITR_INTERVAL in configuration files. Though there is a default configuration, the interval could be tuned by the users with that configuration item depends on what the user cares about more, performance or per packet latency.

Example of getting best performance with l3fwd example

The following is an example of running the DPDK l3fwd sample application to get high performance with an Intel server platform and Intel XL710 NICs.

The example scenario is to get best performance with two Intel XL710 40GbE ports. See Fig. 15.1 for the performance test setup.

_images/intel_perf_test_setup.svg

Performance Test Setup

  1. Add two Intel XL710 NICs to the platform, and use one port per card to get best performance. The reason for using two NICs is to overcome a PCIe Gen3’s limitation since it cannot provide 80G bandwidth for two 40G ports, but two different PCIe Gen3 x8 slot can. Refer to the sample NICs output above, then we can select 82:00.0 and 85:00.0 as test ports:

    82:00.0 Ethernet [0200]: Intel XL710 for 40GbE QSFP+ [8086:1583]
    85:00.0 Ethernet [0200]: Intel XL710 for 40GbE QSFP+ [8086:1583]
    
  2. Connect the ports to the traffic generator. For high speed testing, it’s best to use a hardware traffic generator.

  3. Check the PCI devices numa node (socket id) and get the cores number on the exact socket id. In this case, 82:00.0 and 85:00.0 are both in socket 1, and the cores on socket 1 in the referenced platform are 18-35 and 54-71. Note: Don’t use 2 logical cores on the same core (e.g core18 has 2 logical cores, core18 and core54), instead, use 2 logical cores from different cores (e.g core18 and core19).

  4. Bind these two ports to igb_uio.

  5. As to XL710 40G port, we need at least two queue pairs to achieve best performance, then two queues per port will be required, and each queue pair will need a dedicated CPU core for receiving/transmitting packets.

  6. The DPDK sample application l3fwd will be used for performance testing, with using two ports for bi-directional forwarding. Compile the l3fwd sample with the default lpm mode.

  7. The command line of running l3fwd would be something like the following:

    ./l3fwd -l 18-21 -n 4 -w 82:00.0 -w 85:00.0 \
            -- -p 0x3 --config '(0,0,18),(0,1,19),(1,0,20),(1,1,21)'
    

    This means that the application uses core 18 for port 0, queue pair 0 forwarding, core 19 for port 0, queue pair 1 forwarding, core 20 for port 1, queue pair 0 forwarding, and core 21 for port 1, queue pair 1 forwarding.

  8. Configure the traffic at a traffic generator.

    • Start creating a stream on packet generator.
    • Set the Ethernet II type to 0x0800.

IXGBE Driver

Vector PMD for IXGBE

Vector PMD uses Intel® SIMD instructions to optimize packet I/O. It improves load/store bandwidth efficiency of L1 data cache by using a wider SSE/AVX register 1 (1). The wider register gives space to hold multiple packet buffers so as to save instruction number when processing bulk of packets.

There is no change to PMD API. The RX/TX handler are the only two entries for vPMD packet I/O. They are transparently registered at runtime RX/TX execution if all condition checks pass.

  1. To date, only an SSE version of IX GBE vPMD is available. To ensure that vPMD is in the binary code, ensure that the option CONFIG_RTE_IXGBE_INC_VECTOR=y is in the configure file.

Some constraints apply as pre-conditions for specific optimizations on bulk packet transfers. The following sections explain RX and TX constraints in the vPMD.

RX Constraints
Prerequisites and Pre-conditions

The following prerequisites apply:

  • To enable vPMD to work for RX, bulk allocation for Rx must be allowed.

Ensure that the following pre-conditions are satisfied:

  • rxq->rx_free_thresh >= RTE_PMD_IXGBE_RX_MAX_BURST
  • rxq->rx_free_thresh < rxq->nb_rx_desc
  • (rxq->nb_rx_desc % rxq->rx_free_thresh) == 0
  • rxq->nb_rx_desc < (IXGBE_MAX_RING_DESC - RTE_PMD_IXGBE_RX_MAX_BURST)

These conditions are checked in the code.

Scattered packets are not supported in this mode. If an incoming packet is greater than the maximum acceptable length of one “mbuf” data size (by default, the size is 2 KB), vPMD for RX would be disabled.

By default, IXGBE_MAX_RING_DESC is set to 4096 and RTE_PMD_IXGBE_RX_MAX_BURST is set to 32.

Feature not Supported by RX Vector PMD

Some features are not supported when trying to increase the throughput in vPMD. They are:

  • IEEE1588
  • FDIR
  • Header split
  • RX checksum off load

Other features are supported using optional MACRO configuration. They include:

  • HW VLAN strip
  • HW extend dual VLAN

To guarantee the constraint, configuration flags in dev_conf.rxmode will be checked:

  • hw_vlan_strip
  • hw_vlan_extend
  • hw_ip_checksum
  • header_split
  • dev_conf

fdir_conf->mode will also be checked.

RX Burst Size

As vPMD is focused on high throughput, it assumes that the RX burst size is equal to or greater than 32 per burst. It returns zero if using nb_pkt < 32 as the expected packet number in the receive handler.

TX Constraint
Prerequisite

The only prerequisite is related to tx_rs_thresh. The tx_rs_thresh value must be greater than or equal to RTE_PMD_IXGBE_TX_MAX_BURST, but less or equal to RTE_IXGBE_TX_MAX_FREE_BUF_SZ. Consequently, by default the tx_rs_thresh value is in the range 32 to 64.

Feature not Supported by TX Vector PMD

TX vPMD only works when txq_flags is set to IXGBE_SIMPLE_FLAGS.

This means that it does not support TX multi-segment, VLAN offload and TX csum offload. The following MACROs are used for these three features:

  • ETH_TXQ_FLAGS_NOMULTSEGS
  • ETH_TXQ_FLAGS_NOVLANOFFL
  • ETH_TXQ_FLAGS_NOXSUMSCTP
  • ETH_TXQ_FLAGS_NOXSUMUDP
  • ETH_TXQ_FLAGS_NOXSUMTCP

Application Programming Interface

In DPDK release v16.11 an API for ixgbe specific functions has been added to the ixgbe PMD. The declarations for the API functions are in the header rte_pmd_ixgbe.h.

Sample Application Notes

l3fwd

When running l3fwd with vPMD, there is one thing to note. In the configuration, ensure that port_conf.rxmode.hw_ip_checksum=0. Otherwise, by default, RX vPMD is disabled.

load_balancer

As in the case of l3fwd, set configure port_conf.rxmode.hw_ip_checksum=0 to enable vPMD. In addition, for improved performance, use -bsz “(32,32),(64,64),(32,32)” in load_balancer to avoid using the default burst size of 144.

Limitations or Known issues

Malicious Driver Detection not Supported

The Intel x550 series NICs support a feature called MDD (Malicious Driver Detection) which checks the behavior of the VF driver. If this feature is enabled, the VF must use the advanced context descriptor correctly and set the CC (Check Context) bit. DPDK PF doesn’t support MDD, but kernel PF does. We may hit problem in this scenario kernel PF + DPDK VF. If user enables MDD in kernel PF, DPDK VF will not work. Because kernel PF thinks the VF is malicious. But actually it’s not. The only reason is the VF doesn’t act as MDD required. There’s significant performance impact to support MDD. DPDK should check if the advanced context descriptor should be set and set it. And DPDK has to ask the info about the header length from the upper layer, because parsing the packet itself is not acceptable. So, it’s too expensive to support MDD. When using kernel PF + DPDK VF on x550, please make sure to use a kernel PF driver that disables MDD or can disable MDD.

Some kernel drivers already disable MDD by default while some kernels can use the command insmod ixgbe.ko MDD=0,0 to disable MDD. Each “0” in the command refers to a port. For example, if there are 6 ixgbe ports, the command should be changed to insmod ixgbe.ko MDD=0,0,0,0,0,0.

Statistics

The statistics of ixgbe hardware must be polled regularly in order for it to remain consistent. Running a DPDK application without polling the statistics will cause registers on hardware to count to the maximum value, and “stick” at that value.

In order to avoid statistic registers every reaching the maximum value, read the statistics from the hardware using rte_eth_stats_get() or rte_eth_xstats_get().

The maximum time between statistics polls that ensures consistent results can be calculated as follows:

max_read_interval = UINT_MAX / max_packets_per_second
max_read_interval = 4294967295 / 14880952
max_read_interval = 288.6218096127183 (seconds)
max_read_interval = ~4 mins 48 sec.

In order to ensure valid results, it is recommended to poll every 4 minutes.

MTU setting

Although the user can set the MTU separately on PF and VF ports, the ixgbe NIC only supports one global MTU per physical port. So when the user sets different MTUs on PF and VF ports in one physical port, the real MTU for all these PF and VF ports is the largest value set. This behavior is based on the kernel driver behavior.

VF MAC address setting

On ixgbe, the concept of “pool” can be used for different things depending on the mode. In VMDq mode, “pool” means a VMDq pool. In IOV mode, “pool” means a VF.

There is no RTE API to add a VF’s MAC address from the PF. On ixgbe, the rte_eth_dev_mac_addr_add() function can be used to add a VF’s MAC address, as a workaround.

Inline crypto processing support

Inline IPsec processing is supported for RTE_SECURITY_ACTION_TYPE_INLINE_CRYPTO mode for ESP packets only:

  • ESP authentication only: AES-128-GMAC (128-bit key)
  • ESP encryption and authentication: AES-128-GCM (128-bit key)

IPsec Security Gateway Sample Application supports inline IPsec processing for ixgbe PMD.

For more details see the IPsec Security Gateway Sample Application and Security library documentation.

Supported Chipsets and NICs

  • Intel 82599EB 10 Gigabit Ethernet Controller
  • Intel 82598EB 10 Gigabit Ethernet Controller
  • Intel 82599ES 10 Gigabit Ethernet Controller
  • Intel 82599EN 10 Gigabit Ethernet Controller
  • Intel Ethernet Controller X540-AT2
  • Intel Ethernet Controller X550-BT2
  • Intel Ethernet Controller X550-AT2
  • Intel Ethernet Controller X550-AT
  • Intel Ethernet Converged Network Adapter X520-SR1
  • Intel Ethernet Converged Network Adapter X520-SR2
  • Intel Ethernet Converged Network Adapter X520-LR1
  • Intel Ethernet Converged Network Adapter X520-DA1
  • Intel Ethernet Converged Network Adapter X520-DA2
  • Intel Ethernet Converged Network Adapter X520-DA4
  • Intel Ethernet Converged Network Adapter X520-QDA1
  • Intel Ethernet Converged Network Adapter X520-T2
  • Intel 10 Gigabit AF DA Dual Port Server Adapter
  • Intel 10 Gigabit AT Server Adapter
  • Intel 10 Gigabit AT2 Server Adapter
  • Intel 10 Gigabit CX4 Dual Port Server Adapter
  • Intel 10 Gigabit XF LR Server Adapter
  • Intel 10 Gigabit XF SR Dual Port Server Adapter
  • Intel 10 Gigabit XF SR Server Adapter
  • Intel Ethernet Converged Network Adapter X540-T1
  • Intel Ethernet Converged Network Adapter X540-T2
  • Intel Ethernet Converged Network Adapter X550-T1
  • Intel Ethernet Converged Network Adapter X550-T2

I40E/IXGBE/IGB Virtual Function Driver

Supported Intel® Ethernet Controllers (see the DPDK Release Notes for details) support the following modes of operation in a virtualized environment:

  • SR-IOV mode: Involves direct assignment of part of the port resources to different guest operating systems using the PCI-SIG Single Root I/O Virtualization (SR IOV) standard, also known as “native mode” or “pass-through” mode. In this chapter, this mode is referred to as IOV mode.
  • VMDq mode: Involves central management of the networking resources by an IO Virtual Machine (IOVM) or a Virtual Machine Monitor (VMM), also known as software switch acceleration mode. In this chapter, this mode is referred to as the Next Generation VMDq mode.

SR-IOV Mode Utilization in a DPDK Environment

The DPDK uses the SR-IOV feature for hardware-based I/O sharing in IOV mode. Therefore, it is possible to partition SR-IOV capability on Ethernet controller NIC resources logically and expose them to a virtual machine as a separate PCI function called a “Virtual Function”. Refer to Fig. 17.15.

Therefore, a NIC is logically distributed among multiple virtual machines (as shown in Fig. 17.15), while still having global data in common to share with the Physical Function and other Virtual Functions. The DPDK fm10kvf, i40evf, igbvf or ixgbevf as a Poll Mode Driver (PMD) serves for the Intel® 82576 Gigabit Ethernet Controller, Intel® Ethernet Controller I350 family, Intel® 82599 10 Gigabit Ethernet Controller NIC, Intel® Fortville 10/40 Gigabit Ethernet Controller NIC’s virtual PCI function, or PCIe host-interface of the Intel Ethernet Switch FM10000 Series. Meanwhile the DPDK Poll Mode Driver (PMD) also supports “Physical Function” of such NIC’s on the host.

The DPDK PF/VF Poll Mode Driver (PMD) supports the Layer 2 switch on Intel® 82576 Gigabit Ethernet Controller, Intel® Ethernet Controller I350 family, Intel® 82599 10 Gigabit Ethernet Controller, and Intel® Fortville 10/40 Gigabit Ethernet Controller NICs so that guest can choose it for inter virtual machine traffic in SR-IOV mode.

For more detail on SR-IOV, please refer to the following documents:

_images/single_port_nic.png

Virtualization for a Single Port NIC in SR-IOV Mode

Physical and Virtual Function Infrastructure

The following describes the Physical Function and Virtual Functions infrastructure for the supported Ethernet Controller NICs.

Virtual Functions operate under the respective Physical Function on the same NIC Port and therefore have no access to the global NIC resources that are shared between other functions for the same NIC port.

A Virtual Function has basic access to the queue resources and control structures of the queues assigned to it. For global resource access, a Virtual Function has to send a request to the Physical Function for that port, and the Physical Function operates on the global resources on behalf of the Virtual Function. For this out-of-band communication, an SR-IOV enabled NIC provides a memory buffer for each Virtual Function, which is called a “Mailbox”.

The PCIE host-interface of Intel Ethernet Switch FM10000 Series VF infrastructure

In a virtualized environment, the programmer can enable a maximum of 64 Virtual Functions (VF) globally per PCIE host-interface of the Intel Ethernet Switch FM10000 Series device. Each VF can have a maximum of 16 queue pairs. The Physical Function in host could be only configured by the Linux* fm10k driver (in the case of the Linux Kernel-based Virtual Machine [KVM]), DPDK PMD PF driver doesn’t support it yet.

For example,

  • Using Linux* fm10k driver:

    rmmod fm10k (To remove the fm10k module)
    insmod fm0k.ko max_vfs=2,2 (To enable two Virtual Functions per port)
    

Virtual Function enumeration is performed in the following sequence by the Linux* pci driver for a dual-port NIC. When you enable the four Virtual Functions with the above command, the four enabled functions have a Function# represented by (Bus#, Device#, Function#) in sequence starting from 0 to 3. However:

  • Virtual Functions 0 and 2 belong to Physical Function 0
  • Virtual Functions 1 and 3 belong to Physical Function 1

Note

The above is an important consideration to take into account when targeting specific packets to a selected port.

Intel® X710/XL710 Gigabit Ethernet Controller VF Infrastructure

In a virtualized environment, the programmer can enable a maximum of 128 Virtual Functions (VF) globally per Intel® X710/XL710 Gigabit Ethernet Controller NIC device. The number of queue pairs of each VF can be configured by CONFIG_RTE_LIBRTE_I40E_QUEUE_NUM_PER_VF in config file. The Physical Function in host could be either configured by the Linux* i40e driver (in the case of the Linux Kernel-based Virtual Machine [KVM]) or by DPDK PMD PF driver. When using both DPDK PMD PF/VF drivers, the whole NIC will be taken over by DPDK based application.

For example,

  • Using Linux* i40e driver:

    rmmod i40e (To remove the i40e module)
    insmod i40e.ko max_vfs=2,2 (To enable two Virtual Functions per port)
    
  • Using the DPDK PMD PF i40e driver:

    Kernel Params: iommu=pt, intel_iommu=on

    modprobe uio
    insmod igb_uio
    ./dpdk-devbind.py -b igb_uio bb:ss.f
    echo 2 > /sys/bus/pci/devices/0000\:bb\:ss.f/max_vfs (To enable two VFs on a specific PCI device)
    

    Launch the DPDK testpmd/example or your own host daemon application using the DPDK PMD library.

Virtual Function enumeration is performed in the following sequence by the Linux* pci driver for a dual-port NIC. When you enable the four Virtual Functions with the above command, the four enabled functions have a Function# represented by (Bus#, Device#, Function#) in sequence starting from 0 to 3. However:

  • Virtual Functions 0 and 2 belong to Physical Function 0
  • Virtual Functions 1 and 3 belong to Physical Function 1

Note

The above is an important consideration to take into account when targeting specific packets to a selected port.

For Intel® X710/XL710 Gigabit Ethernet Controller, queues are in pairs. One queue pair means one receive queue and one transmit queue. The default number of queue pairs per VF is 4, and can be 16 in maximum.

Intel® 82599 10 Gigabit Ethernet Controller VF Infrastructure

The programmer can enable a maximum of 63 Virtual Functions and there must be one Physical Function per Intel® 82599 10 Gigabit Ethernet Controller NIC port. The reason for this is that the device allows for a maximum of 128 queues per port and a virtual/physical function has to have at least one queue pair (RX/TX). The current implementation of the DPDK ixgbevf driver supports a single queue pair (RX/TX) per Virtual Function. The Physical Function in host could be either configured by the Linux* ixgbe driver (in the case of the Linux Kernel-based Virtual Machine [KVM]) or by DPDK PMD PF driver. When using both DPDK PMD PF/VF drivers, the whole NIC will be taken over by DPDK based application.

For example,

  • Using Linux* ixgbe driver:

    rmmod ixgbe (To remove the ixgbe module)
    insmod ixgbe max_vfs=2,2 (To enable two Virtual Functions per port)
    
  • Using the DPDK PMD PF ixgbe driver:

    Kernel Params: iommu=pt, intel_iommu=on

    modprobe uio
    insmod igb_uio
    ./dpdk-devbind.py -b igb_uio bb:ss.f
    echo 2 > /sys/bus/pci/devices/0000\:bb\:ss.f/max_vfs (To enable two VFs on a specific PCI device)
    

    Launch the DPDK testpmd/example or your own host daemon application using the DPDK PMD library.

  • Using the DPDK PMD PF ixgbe driver to enable VF RSS:

    Same steps as above to install the modules of uio, igb_uio, specify max_vfs for PCI device, and launch the DPDK testpmd/example or your own host daemon application using the DPDK PMD library.

    The available queue number (at most 4) per VF depends on the total number of pool, which is determined by the max number of VF at PF initialization stage and the number of queue specified in config:

    • If the max number of VFs (max_vfs) is set in the range of 1 to 32:

      If the number of Rx queues is specified as 4 (--rxq=4 in testpmd), then there are totally 32 pools (ETH_32_POOLS), and each VF could have 4 Rx queues;

      If the number of Rx queues is specified as 2 (--rxq=2 in testpmd), then there are totally 32 pools (ETH_32_POOLS), and each VF could have 2 Rx queues;

    • If the max number of VFs (max_vfs) is in the range of 33 to 64:

      If the number of Rx queues in specified as 4 (--rxq=4 in testpmd), then error message is expected as rxq is not correct at this case;

      If the number of rxq is 2 (--rxq=2 in testpmd), then there is totally 64 pools (ETH_64_POOLS), and each VF have 2 Rx queues;

    On host, to enable VF RSS functionality, rx mq mode should be set as ETH_MQ_RX_VMDQ_RSS or ETH_MQ_RX_RSS mode, and SRIOV mode should be activated (max_vfs >= 1). It also needs config VF RSS information like hash function, RSS key, RSS key length.

Note

The limitation for VF RSS on Intel® 82599 10 Gigabit Ethernet Controller is: The hash and key are shared among PF and all VF, the RETA table with 128 entries is also shared among PF and all VF; So it could not to provide a method to query the hash and reta content per VF on guest, while, if possible, please query them on host for the shared RETA information.

Virtual Function enumeration is performed in the following sequence by the Linux* pci driver for a dual-port NIC. When you enable the four Virtual Functions with the above command, the four enabled functions have a Function# represented by (Bus#, Device#, Function#) in sequence starting from 0 to 3. However:

  • Virtual Functions 0 and 2 belong to Physical Function 0
  • Virtual Functions 1 and 3 belong to Physical Function 1

Note

The above is an important consideration to take into account when targeting specific packets to a selected port.

Intel® 82576 Gigabit Ethernet Controller and Intel® Ethernet Controller I350 Family VF Infrastructure

In a virtualized environment, an Intel® 82576 Gigabit Ethernet Controller serves up to eight virtual machines (VMs). The controller has 16 TX and 16 RX queues. They are generally referred to (or thought of) as queue pairs (one TX and one RX queue). This gives the controller 16 queue pairs.

A pool is a group of queue pairs for assignment to the same VF, used for transmit and receive operations. The controller has eight pools, with each pool containing two queue pairs, that is, two TX and two RX queues assigned to each VF.

In a virtualized environment, an Intel® Ethernet Controller I350 family device serves up to eight virtual machines (VMs) per port. The eight queues can be accessed by eight different VMs if configured correctly (the i350 has 4x1GbE ports each with 8T X and 8 RX queues), that means, one Transmit and one Receive queue assigned to each VF.

For example,

  • Using Linux* igb driver:

    rmmod igb (To remove the igb module)
    insmod igb max_vfs=2,2 (To enable two Virtual Functions per port)
    
  • Using DPDK PMD PF igb driver:

    Kernel Params: iommu=pt, intel_iommu=on modprobe uio

    insmod igb_uio
    ./dpdk-devbind.py -b igb_uio bb:ss.f
    echo 2 > /sys/bus/pci/devices/0000\:bb\:ss.f/max_vfs (To enable two VFs on a specific pci device)
    

    Launch DPDK testpmd/example or your own host daemon application using the DPDK PMD library.

Virtual Function enumeration is performed in the following sequence by the Linux* pci driver for a four-port NIC. When you enable the four Virtual Functions with the above command, the four enabled functions have a Function# represented by (Bus#, Device#, Function#) in sequence, starting from 0 to 7. However:

  • Virtual Functions 0 and 4 belong to Physical Function 0
  • Virtual Functions 1 and 5 belong to Physical Function 1
  • Virtual Functions 2 and 6 belong to Physical Function 2
  • Virtual Functions 3 and 7 belong to Physical Function 3

Note

The above is an important consideration to take into account when targeting specific packets to a selected port.

Validated Hypervisors

The validated hypervisor is:

  • KVM (Kernel Virtual Machine) with Qemu, version 0.14.0

However, the hypervisor is bypassed to configure the Virtual Function devices using the Mailbox interface, the solution is hypervisor-agnostic. Xen* and VMware* (when SR- IOV is supported) will also be able to support the DPDK with Virtual Function driver support.

Expected Guest Operating System in Virtual Machine

The expected guest operating systems in a virtualized environment are:

  • Fedora* 14 (64-bit)
  • Ubuntu* 10.04 (64-bit)

For supported kernel versions, refer to the DPDK Release Notes.

Setting Up a KVM Virtual Machine Monitor

The following describes a target environment:

  • Host Operating System: Fedora 14
  • Hypervisor: KVM (Kernel Virtual Machine) with Qemu version 0.14.0
  • Guest Operating System: Fedora 14
  • Linux Kernel Version: Refer to the DPDK Getting Started Guide
  • Target Applications: l2fwd, l3fwd-vf

The setup procedure is as follows:

  1. Before booting the Host OS, open BIOS setup and enable Intel® VT features.

  2. While booting the Host OS kernel, pass the intel_iommu=on kernel command line argument using GRUB. When using DPDK PF driver on host, pass the iommu=pt kernel command line argument in GRUB.

  3. Download qemu-kvm-0.14.0 from http://sourceforge.net/projects/kvm/files/qemu-kvm/ and install it in the Host OS using the following steps:

    When using a recent kernel (2.6.25+) with kvm modules included:

    tar xzf qemu-kvm-release.tar.gz
    cd qemu-kvm-release
    ./configure --prefix=/usr/local/kvm
    make
    sudo make install
    sudo /sbin/modprobe kvm-intel
    

    When using an older kernel, or a kernel from a distribution without the kvm modules, you must download (from the same link), compile and install the modules yourself:

    tar xjf kvm-kmod-release.tar.bz2
    cd kvm-kmod-release
    ./configure
    make
    sudo make install
    sudo /sbin/modprobe kvm-intel
    

    qemu-kvm installs in the /usr/local/bin directory.

    For more details about KVM configuration and usage, please refer to:

    http://www.linux-kvm.org/page/HOWTO1.

  4. Create a Virtual Machine and install Fedora 14 on the Virtual Machine. This is referred to as the Guest Operating System (Guest OS).

  5. Download and install the latest ixgbe driver from:

    http://downloadcenter.intel.com/Detail_Desc.aspx?agr=Y&amp;DwnldID=14687

  6. In the Host OS

    When using Linux kernel ixgbe driver, unload the Linux ixgbe driver and reload it with the max_vfs=2,2 argument:

    rmmod ixgbe
    modprobe ixgbe max_vfs=2,2
    

    When using DPDK PMD PF driver, insert DPDK kernel module igb_uio and set the number of VF by sysfs max_vfs:

    modprobe uio
    insmod igb_uio
    ./dpdk-devbind.py -b igb_uio 02:00.0 02:00.1 0e:00.0 0e:00.1
    echo 2 > /sys/bus/pci/devices/0000\:02\:00.0/max_vfs
    echo 2 > /sys/bus/pci/devices/0000\:02\:00.1/max_vfs
    echo 2 > /sys/bus/pci/devices/0000\:0e\:00.0/max_vfs
    echo 2 > /sys/bus/pci/devices/0000\:0e\:00.1/max_vfs
    

    Note

    You need to explicitly specify number of vfs for each port, for example, in the command above, it creates two vfs for the first two ixgbe ports.

    Let say we have a machine with four physical ixgbe ports:

    0000:02:00.0

    0000:02:00.1

    0000:0e:00.0

    0000:0e:00.1

    The command above creates two vfs for device 0000:02:00.0:

    ls -alrt /sys/bus/pci/devices/0000\:02\:00.0/virt*
    lrwxrwxrwx. 1 root root 0 Apr 13 05:40 /sys/bus/pci/devices/0000:02:00.0/virtfn1 -> ../0000:02:10.2
    lrwxrwxrwx. 1 root root 0 Apr 13 05:40 /sys/bus/pci/devices/0000:02:00.0/virtfn0 -> ../0000:02:10.0
    

    It also creates two vfs for device 0000:02:00.1:

    ls -alrt /sys/bus/pci/devices/0000\:02\:00.1/virt*
    lrwxrwxrwx. 1 root root 0 Apr 13 05:51 /sys/bus/pci/devices/0000:02:00.1/virtfn1 -> ../0000:02:10.3
    lrwxrwxrwx. 1 root root 0 Apr 13 05:51 /sys/bus/pci/devices/0000:02:00.1/virtfn0 -> ../0000:02:10.1
    
  7. List the PCI devices connected and notice that the Host OS shows two Physical Functions (traditional ports) and four Virtual Functions (two for each port). This is the result of the previous step.

  8. Insert the pci_stub module to hold the PCI devices that are freed from the default driver using the following command (see http://www.linux-kvm.org/page/How_to_assign_devices_with_VT-d_in_KVM Section 4 for more information):

    sudo /sbin/modprobe pci-stub
    

    Unbind the default driver from the PCI devices representing the Virtual Functions. A script to perform this action is as follows:

    echo "8086 10ed" > /sys/bus/pci/drivers/pci-stub/new_id
    echo 0000:08:10.0 > /sys/bus/pci/devices/0000:08:10.0/driver/unbind
    echo 0000:08:10.0 > /sys/bus/pci/drivers/pci-stub/bind
    

    where, 0000:08:10.0 belongs to the Virtual Function visible in the Host OS.

  9. Now, start the Virtual Machine by running the following command:

    /usr/local/kvm/bin/qemu-system-x86_64 -m 4096 -smp 4 -boot c -hda lucid.qcow2 -device pci-assign,host=08:10.0
    

    where:

    — -m = memory to assign

    —-smp = number of smp cores

    — -boot = boot option

    —-hda = virtual disk image

    — -device = device to attach

    Note

    — The pci-assign,host=08:10.0 value indicates that you want to attach a PCI device to a Virtual Machine and the respective (Bus:Device.Function) numbers should be passed for the Virtual Function to be attached.

    — qemu-kvm-0.14.0 allows a maximum of four PCI devices assigned to a VM, but this is qemu-kvm version dependent since qemu-kvm-0.14.1 allows a maximum of five PCI devices.

    — qemu-system-x86_64 also has a -cpu command line option that is used to select the cpu_model to emulate in a Virtual Machine. Therefore, it can be used as:

    /usr/local/kvm/bin/qemu-system-x86_64 -cpu ?
    
    (to list all available cpu_models)
    
    /usr/local/kvm/bin/qemu-system-x86_64 -m 4096 -cpu host -smp 4 -boot c -hda lucid.qcow2 -device pci-assign,host=08:10.0
    
    (to use the same cpu_model equivalent to the host cpu)
    

    For more information, please refer to: http://wiki.qemu.org/Features/CPUModels.

  10. Install and run DPDK host app to take over the Physical Function. Eg.

    make install T=x86_64-native-linuxapp-gcc
    ./x86_64-native-linuxapp-gcc/app/testpmd -l 0-3 -n 4 -- -i
    
  11. Finally, access the Guest OS using vncviewer with the localhost:5900 port and check the lspci command output in the Guest OS. The virtual functions will be listed as available for use.

  12. Configure and install the DPDK with an x86_64-native-linuxapp-gcc configuration on the Guest OS as normal, that is, there is no change to the normal installation procedure.

    make config T=x86_64-native-linuxapp-gcc O=x86_64-native-linuxapp-gcc
    cd x86_64-native-linuxapp-gcc
    make
    

Note

If you are unable to compile the DPDK and you are getting “error: CPU you selected does not support x86-64 instruction set”, power off the Guest OS and start the virtual machine with the correct -cpu option in the qemu- system-x86_64 command as shown in step 9. You must select the best x86_64 cpu_model to emulate or you can select host option if available.

Note

Run the DPDK l2fwd sample application in the Guest OS with Hugepages enabled. For the expected benchmark performance, you must pin the cores from the Guest OS to the Host OS (taskset can be used to do this) and you must also look at the PCI Bus layout on the board to ensure you are not running the traffic over the QPI Interface.

Note

  • The Virtual Machine Manager (the Fedora package name is virt-manager) is a utility for virtual machine management that can also be used to create, start, stop and delete virtual machines. If this option is used, step 2 and 6 in the instructions provided will be different.
  • virsh, a command line utility for virtual machine management, can also be used to bind and unbind devices to a virtual machine in Ubuntu. If this option is used, step 6 in the instructions provided will be different.
  • The Virtual Machine Monitor (see Fig. 17.16) is equivalent to a Host OS with KVM installed as described in the instructions.
_images/perf_benchmark.png

Performance Benchmark Setup

DPDK SR-IOV PMD PF/VF Driver Usage Model

Fast Host-based Packet Processing

Software Defined Network (SDN) trends are demanding fast host-based packet handling. In a virtualization environment, the DPDK VF PMD driver performs the same throughput result as a non-VT native environment.

With such host instance fast packet processing, lots of services such as filtering, QoS, DPI can be offloaded on the host fast path.

Fig. 17.17 shows the scenario where some VMs directly communicate externally via a VFs, while others connect to a virtual switch and share the same uplink bandwidth.

_images/fast_pkt_proc.png

Fast Host-based Packet Processing

SR-IOV (PF/VF) Approach for Inter-VM Communication

Inter-VM data communication is one of the traffic bottle necks in virtualization platforms. SR-IOV device assignment helps a VM to attach the real device, taking advantage of the bridge in the NIC. So VF-to-VF traffic within the same physical port (VM0<->VM1) have hardware acceleration. However, when VF crosses physical ports (VM0<->VM2), there is no such hardware bridge. In this case, the DPDK PMD PF driver provides host forwarding between such VMs.

Fig. 17.18 shows an example. In this case an update of the MAC address lookup tables in both the NIC and host DPDK application is required.

In the NIC, writing the destination of a MAC address belongs to another cross device VM to the PF specific pool. So when a packet comes in, its destination MAC address will match and forward to the host DPDK PMD application.

In the host DPDK application, the behavior is similar to L2 forwarding, that is, the packet is forwarded to the correct PF pool. The SR-IOV NIC switch forwards the packet to a specific VM according to the MAC destination address which belongs to the destination VF on the VM.

_images/inter_vm_comms.png

Inter-VM Communication

KNI Poll Mode Driver

KNI PMD is wrapper to the librte_kni library.

This PMD enables using KNI without having a KNI specific application, any forwarding application can use PMD interface for KNI.

Sending packets to any DPDK controlled interface or sending to the Linux networking stack will be transparent to the DPDK application.

To create a KNI device net_kni# device name should be used, and this will create kni# Linux virtual network interface.

There is no physical device backend for the virtual KNI device.

Packets sent to the KNI Linux interface will be received by the DPDK application, and DPDK application may forward packets to a physical NIC or to a virtual device (like another KNI interface or PCAP interface).

To forward any traffic from physical NIC to the Linux networking stack, an application should control a physical port and create one virtual KNI port, and forward between two.

Using this PMD requires KNI kernel module be inserted.

Usage

EAL --vdev argument can be used to create KNI device instance, like:

testpmd --vdev=net_kni0 --vdev=net_kn1 -- -i

Above command will create kni0 and kni1 Linux network interfaces, those interfaces can be controlled by standard Linux tools.

When testpmd forwarding starts, any packets sent to kni0 interface forwarded to the kni1 interface and vice versa.

There is no hard limit on number of interfaces that can be created.

Default interface configuration

librte_kni can create Linux network interfaces with different features, feature set controlled by a configuration struct, and KNI PMD uses a fixed configuration:

Interface name: kni#
force bind kernel thread to a core : NO
mbuf size: MAX_PACKET_SZ

KNI control path is not supported with the PMD, since there is no physical backend device by default.

PMD arguments

no_request_thread, by default PMD creates a phtread for each KNI interface to handle Linux network interface control commands, like ifconfig kni0 up

With no_request_thread option, pthread is not created and control commands not handled by PMD.

By default request thread is enabled. And this argument should not be used most of the time, unless this PMD used with customized DPDK application to handle requests itself.

Argument usage:

testpmd --vdev "net_kni0,no_request_thread=1" -- -i

PMD log messages

If KNI kernel module (rte_kni.ko) not inserted, following error log printed:

"KNI: KNI subsystem has not been initialized. Invoke rte_kni_init() first"

PMD testing

It is possible to test PMD quickly using KNI kernel module loopback feature:

  • Insert KNI kernel module with loopback support:

    insmod build/kmod/rte_kni.ko lo_mode=lo_mode_fifo_skb
    
  • Start testpmd with no physical device but two KNI virtual devices:

    ./testpmd --vdev net_kni0 --vdev net_kni1 -- -i
    
    ...
    Configuring Port 0 (socket 0)
    KNI: pci: 00:00:00       c580:b8
    Port 0: 1A:4A:5B:7C:A2:8C
    Configuring Port 1 (socket 0)
    KNI: pci: 00:00:00       600:b9
    Port 1: AE:95:21:07:93:DD
    Checking link statuses...
    Port 0 Link Up - speed 10000 Mbps - full-duplex
    Port 1 Link Up - speed 10000 Mbps - full-duplex
    Done
    testpmd>
    
  • Observe Linux interfaces

    $ ifconfig kni0 && ifconfig kni1
    kni0: flags=4098<BROADCAST,MULTICAST>  mtu 1500
            ether ae:8e:79:8e:9b:c8  txqueuelen 1000  (Ethernet)
            RX packets 0  bytes 0 (0.0 B)
            RX errors 0  dropped 0  overruns 0  frame 0
            TX packets 0  bytes 0 (0.0 B)
            TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0
    
    kni1: flags=4098<BROADCAST,MULTICAST>  mtu 1500
            ether 9e:76:43:53:3e:9b  txqueuelen 1000  (Ethernet)
            RX packets 0  bytes 0 (0.0 B)
            RX errors 0  dropped 0  overruns 0  frame 0
            TX packets 0  bytes 0 (0.0 B)
            TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0
    
  • Start forwarding with tx_first:

    testpmd> start tx_first
    
  • Quit and check forwarding stats:

    testpmd> quit
    Telling cores to stop...
    Waiting for lcores to finish...
    
    ---------------------- Forward statistics for port 0  ----------------------
    RX-packets: 35637905       RX-dropped: 0             RX-total: 35637905
    TX-packets: 35637947       TX-dropped: 0             TX-total: 35637947
    ----------------------------------------------------------------------------
    
    ---------------------- Forward statistics for port 1  ----------------------
    RX-packets: 35637915       RX-dropped: 0             RX-total: 35637915
    TX-packets: 35637937       TX-dropped: 0             TX-total: 35637937
    ----------------------------------------------------------------------------
    
    +++++++++++++++ Accumulated forward statistics for all ports+++++++++++++++
    RX-packets: 71275820       RX-dropped: 0             RX-total: 71275820
    TX-packets: 71275884       TX-dropped: 0             TX-total: 71275884
    ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
    

LiquidIO VF Poll Mode Driver

The LiquidIO VF PMD library (librte_pmd_lio) provides poll mode driver support for Cavium LiquidIO® II server adapter VFs. PF management and VF creation can be done using kernel driver.

More information can be found at Cavium Official Website.

Supported LiquidIO Adapters

  • LiquidIO II CN2350 210SV/225SV
  • LiquidIO II CN2350 210SVPT
  • LiquidIO II CN2360 210SV/225SV
  • LiquidIO II CN2360 210SVPT

Pre-Installation Configuration

The following options can be modified in the config file. Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_LIBRTE_LIO_PMD (default y)

    Toggle compilation of LiquidIO PMD.

  • CONFIG_RTE_LIBRTE_LIO_DEBUG_DRIVER (default n)

    Toggle display of generic debugging messages.

  • CONFIG_RTE_LIBRTE_LIO_DEBUG_INIT (default n)

    Toggle display of initialization related messages.

  • CONFIG_RTE_LIBRTE_LIO_DEBUG_RX (default n)

    Toggle display of receive fast path run-time messages.

  • CONFIG_RTE_LIBRTE_LIO_DEBUG_TX (default n)

    Toggle display of transmit fast path run-time messages.

  • CONFIG_RTE_LIBRTE_LIO_DEBUG_MBOX (default n)

    Toggle display of mailbox messages.

  • CONFIG_RTE_LIBRTE_LIO_DEBUG_REGS (default n)

    Toggle display of register reads and writes.

SR-IOV: Prerequisites and Sample Application Notes

This section provides instructions to configure SR-IOV with Linux OS.

  1. Verify SR-IOV and ARI capabilities are enabled on the adapter using lspci:

    lspci -s <slot> -vvv
    

    Example output:

    [...]
    Capabilities: [148 v1] Alternative Routing-ID Interpretation (ARI)
    [...]
    Capabilities: [178 v1] Single Root I/O Virtualization (SR-IOV)
    [...]
    Kernel driver in use: LiquidIO
    
  2. Load the kernel module:

    modprobe liquidio
    
  3. Bring up the PF ports:

    ifconfig p4p1 up
    ifconfig p4p2 up
    
  4. Change PF MTU if required:

    ifconfig p4p1 mtu 9000
    ifconfig p4p2 mtu 9000
    
  5. Create VF device(s):

    Echo number of VFs to be created into "sriov_numvfs" sysfs entry of the parent PF.

    echo 1 > /sys/bus/pci/devices/0000:03:00.0/sriov_numvfs
    echo 1 > /sys/bus/pci/devices/0000:03:00.1/sriov_numvfs
    
  6. Assign VF MAC address:

    Assign MAC address to the VF using iproute2 utility. The syntax is:

    ip link set <PF iface> vf <VF id> mac <macaddr>
    

    Example output:

    ip link set p4p1 vf 0 mac F2:A8:1B:5E:B4:66
    
  7. Assign VF(s) to VM.

    The VF devices may be passed through to the guest VM using qemu or virt-manager or virsh etc.

    Example qemu guest launch command:

    ./qemu-system-x86_64 -name lio-vm -machine accel=kvm \
    -cpu host -m 4096 -smp 4 \
    -drive file=<disk_file>,if=none,id=disk1,format=<type> \
    -device virtio-blk-pci,scsi=off,drive=disk1,id=virtio-disk1,bootindex=1 \
    -device vfio-pci,host=03:00.3 -device vfio-pci,host=03:08.3
    
  8. Running testpmd

    Refer to the document compiling and testing a PMD for a NIC to run testpmd application.

    Note

    Use igb_uio instead of vfio-pci in VM.

    Example output:

    [...]
    EAL: PCI device 0000:03:00.3 on NUMA socket 0
    EAL:   probe driver: 177d:9712 net_liovf
    EAL:   using IOMMU type 1 (Type 1)
    PMD: net_liovf[03:00.3]INFO: DEVICE : CN23XX VF
    EAL: PCI device 0000:03:08.3 on NUMA socket 0
    EAL:   probe driver: 177d:9712 net_liovf
    PMD: net_liovf[03:08.3]INFO: DEVICE : CN23XX VF
    Interactive-mode selected
    USER1: create a new mbuf pool <mbuf_pool_socket_0>: n=171456, size=2176, socket=0
    Configuring Port 0 (socket 0)
    PMD: net_liovf[03:00.3]INFO: Starting port 0
    Port 0: F2:A8:1B:5E:B4:66
    Configuring Port 1 (socket 0)
    PMD: net_liovf[03:08.3]INFO: Starting port 1
    Port 1: 32:76:CC:EE:56:D7
    Checking link statuses...
    Port 0 Link Up - speed 10000 Mbps - full-duplex
    Port 1 Link Up - speed 10000 Mbps - full-duplex
    Done
    testpmd>
    
  9. Enabling VF promiscuous mode

    One VF per PF can be marked as trusted for promiscuous mode.

    ip link set dev <PF iface> vf <VF id> trust on
    

Limitations

VF MTU

VF MTU is limited by PF MTU. Raise PF value before configuring VF for larger packet size.

VLAN offload

Tx VLAN insertion is not supported and consequently VLAN offload feature is marked partial.

Ring size

Number of descriptors for Rx/Tx ring should be in the range 128 to 512.

CRC striping

LiquidIO adapters strip ethernet FCS of every packet coming to the host interface. So, CRC will be stripped even when the rxmode.hw_strip_crc member is set to 0 in struct rte_eth_conf.

MLX4 poll mode driver library

The MLX4 poll mode driver library (librte_pmd_mlx4) implements support for Mellanox ConnectX-3 and Mellanox ConnectX-3 Pro 10/40 Gbps adapters as well as their virtual functions (VF) in SR-IOV context.

Information and documentation about this family of adapters can be found on the Mellanox website. Help is also provided by the Mellanox community.

There is also a section dedicated to this poll mode driver.

Note

Due to external dependencies, this driver is disabled by default. It must be enabled manually by setting CONFIG_RTE_LIBRTE_MLX4_PMD=y and recompiling DPDK.

Implementation details

Most Mellanox ConnectX-3 devices provide two ports but expose a single PCI bus address, thus unlike most drivers, librte_pmd_mlx4 registers itself as a PCI driver that allocates one Ethernet device per detected port.

For this reason, one cannot white/blacklist a single port without also white/blacklisting the others on the same device.

Besides its dependency on libibverbs (that implies libmlx4 and associated kernel support), librte_pmd_mlx4 relies heavily on system calls for control operations such as querying/updating the MTU and flow control parameters.

For security reasons and robustness, this driver only deals with virtual memory addresses. The way resources allocations are handled by the kernel combined with hardware specifications that allow it to handle virtual memory addresses directly ensure that DPDK applications cannot access random physical memory (or memory that does not belong to the current process).

This capability allows the PMD to coexist with kernel network interfaces which remain functional, although they stop receiving unicast packets as long as they share the same MAC address.

Compiling librte_pmd_mlx4 causes DPDK to be linked against libibverbs.

Configuration

Compilation options

These options can be modified in the .config file.

  • CONFIG_RTE_LIBRTE_MLX4_PMD (default n)

    Toggle compilation of librte_pmd_mlx4 itself.

  • CONFIG_RTE_LIBRTE_MLX4_DEBUG (default n)

    Toggle debugging code and stricter compilation flags. Enabling this option adds additional run-time checks and debugging messages at the cost of lower performance.

  • CONFIG_RTE_LIBRTE_MLX4_DEBUG_BROKEN_VERBS (default n)

    Mellanox OFED versions earlier than 4.2 may return false errors from Verbs object destruction APIs after the device is plugged out. Enabling this option replaces assertion checks that cause the program to abort with harmless debugging messages as a workaround. Relevant only when CONFIG_RTE_LIBRTE_MLX4_DEBUG is enabled.

  • CONFIG_RTE_LIBRTE_MLX4_TX_MP_CACHE (default 8)

    Maximum number of cached memory pools (MPs) per TX queue. Each MP from which buffers are to be transmitted must be associated to memory regions (MRs). This is a slow operation that must be cached.

    This value is always 1 for RX queues since they use a single MP.

Run-time configuration
  • librte_pmd_mlx4 brings kernel network interfaces up during initialization because it is affected by their state. Forcing them down prevents packets reception.

  • ethtool operations on related kernel interfaces also affect the PMD.

  • port parameter [int]

    This parameter provides a physical port to probe and can be specified multiple times for additional ports. All ports are probed by default if left unspecified.

Kernel module parameters

The mlx4_core kernel module has several parameters that affect the behavior and/or the performance of librte_pmd_mlx4. Some of them are described below.

  • num_vfs (integer or triplet, optionally prefixed by device address strings)

    Create the given number of VFs on the specified devices.

  • log_num_mgm_entry_size (integer)

    Device-managed flow steering (DMFS) is required by DPDK applications. It is enabled by using a negative value, the last four bits of which have a special meaning.

    • -1: force device-managed flow steering (DMFS).
    • -7: configure optimized steering mode to improve performance with the following limitation: VLAN filtering is not supported with this mode. This is the recommended mode in case VLAN filter is not needed.

Prerequisites

This driver relies on external libraries and kernel drivers for resources allocations and initialization. The following dependencies are not part of DPDK and must be installed separately:

  • libibverbs (provided by rdma-core package)

    User space verbs framework used by librte_pmd_mlx4. This library provides a generic interface between the kernel and low-level user space drivers such as libmlx4.

    It allows slow and privileged operations (context initialization, hardware resources allocations) to be managed by the kernel and fast operations to never leave user space.

  • libmlx4 (provided by rdma-core package)

    Low-level user space driver library for Mellanox ConnectX-3 devices, it is automatically loaded by libibverbs.

    This library basically implements send/receive calls to the hardware queues.

  • Kernel modules

    They provide the kernel-side verbs API and low level device drivers that manage actual hardware initialization and resources sharing with user space processes.

    Unlike most other PMDs, these modules must remain loaded and bound to their devices:

    • mlx4_core: hardware driver managing Mellanox ConnectX-3 devices.
    • mlx4_en: Ethernet device driver that provides kernel network interfaces.
    • mlx4_ib: InifiniBand device driver.
    • ib_uverbs: user space driver for verbs (entry point for libibverbs).
  • Firmware update

    Mellanox OFED releases include firmware updates for ConnectX-3 adapters.

    Because each release provides new features, these updates must be applied to match the kernel modules and libraries they come with.

Note

Both libraries are BSD and GPL licensed. Linux kernel modules are GPL licensed.

Depending on system constraints and user preferences either RDMA core library with a recent enough Linux kernel release (recommended) or Mellanox OFED, which provides compatibility with older releases.

Mellanox OFED as a fallback
  • Mellanox OFED version: 4.2.
  • firmware version: 2.42.5000 and above.

Note

Several versions of Mellanox OFED are available. Installing the version this DPDK release was developed and tested against is strongly recommended. Please check the prerequisites.

Installing Mellanox OFED
  1. Download latest Mellanox OFED.

  2. Install the required libraries and kernel modules either by installing only the required set, or by installing the entire Mellanox OFED:

    For bare metal use:

    ./mlnxofedinstall --dpdk --upstream-libs
    

    For SR-IOV hypervisors use:

    ./mlnxofedinstall --dpdk --upstream-libs --enable-sriov --hypervisor
    

    For SR-IOV virtual machine use:

    ./mlnxofedinstall --dpdk --upstream-libs --guest
    
  3. Verify the firmware is the correct one:

    ibv_devinfo
    
  4. Set all ports links to Ethernet, follow instructions on the screen:

    connectx_port_config
    
  5. Continue with section 2 of the Quick Start Guide.

Supported NICs

  • Mellanox(R) ConnectX(R)-3 Pro 40G MCX354A-FCC_Ax (2*40G)

Quick Start Guide

  1. Set all ports links to Ethernet

    PCI=<NIC PCI address>
    echo eth > "/sys/bus/pci/devices/$PCI/mlx4_port0"
    echo eth > "/sys/bus/pci/devices/$PCI/mlx4_port1"
    

    Note

    If using Mellanox OFED one can permanently set the port link to Ethernet using connectx_port_config tool provided by it. Mellanox OFED as a fallback:

  1. In case of bare metal or hypervisor, configure optimized steering mode by adding the following line to /etc/modprobe.d/mlx4_core.conf:

    options mlx4_core log_num_mgm_entry_size=-7
    

    Note

    If VLAN filtering is used, set log_num_mgm_entry_size=-1. Performance degradation can occur on this case.

  2. Restart the driver:

    /etc/init.d/openibd restart
    

    or:

    service openibd restart
    
  3. Compile DPDK and you are ready to go. See instructions on Development Kit Build System

Performance tuning

  1. Verify the optimized steering mode is configured:
cat /sys/module/mlx4_core/parameters/log_num_mgm_entry_size
  1. Use the CPU near local NUMA node to which the PCIe adapter is connected, for better performance. For VMs, verify that the right CPU and NUMA node are pinned according to the above. Run:

    lstopo-no-graphics
    

    to identify the NUMA node to which the PCIe adapter is connected.

  2. If more than one adapter is used, and root complex capabilities allow to put both adapters on the same NUMA node without PCI bandwidth degradation, it is recommended to locate both adapters on the same NUMA node. This in order to forward packets from one to the other without NUMA performance penalty.

  3. Disable pause frames:

    ethtool -A <netdev> rx off tx off
    
  4. Verify IO non-posted prefetch is disabled by default. This can be checked via the BIOS configuration. Please contact you server provider for more information about the settings.

Note

On some machines, depends on the machine integrator, it is beneficial to set the PCI max read request parameter to 1K. This can be done in the following way:

To query the read request size use:

setpci -s <NIC PCI address> 68.w

If the output is different than 3XXX, set it by:

setpci -s <NIC PCI address> 68.w=3XXX

The XXX can be different on different systems. Make sure to configure according to the setpci output.

Usage example

This section demonstrates how to launch testpmd with Mellanox ConnectX-3 devices managed by librte_pmd_mlx4.

  1. Load the kernel modules:

    modprobe -a ib_uverbs mlx4_en mlx4_core mlx4_ib
    

    Alternatively if MLNX_OFED is fully installed, the following script can be run:

    /etc/init.d/openibd restart
    

    Note

    User space I/O kernel modules (uio and igb_uio) are not used and do not have to be loaded.

  2. Make sure Ethernet interfaces are in working order and linked to kernel verbs. Related sysfs entries should be present:

    ls -d /sys/class/net/*/device/infiniband_verbs/uverbs* | cut -d / -f 5
    

    Example output:

    eth2
    eth3
    eth4
    eth5
    
  3. Optionally, retrieve their PCI bus addresses for whitelisting:

    {
        for intf in eth2 eth3 eth4 eth5;
        do
            (cd "/sys/class/net/${intf}/device/" && pwd -P);
        done;
    } |
    sed -n 's,.*/\(.*\),-w \1,p'
    

    Example output:

    -w 0000:83:00.0
    -w 0000:83:00.0
    -w 0000:84:00.0
    -w 0000:84:00.0
    

    Note

    There are only two distinct PCI bus addresses because the Mellanox ConnectX-3 adapters installed on this system are dual port.

  4. Request huge pages:

    echo 1024 > /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages/nr_hugepages
    
  5. Start testpmd with basic parameters:

    testpmd -l 8-15 -n 4 -w 0000:83:00.0 -w 0000:84:00.0 -- --rxq=2 --txq=2 -i
    

    Example output:

    [...]
    EAL: PCI device 0000:83:00.0 on NUMA socket 1
    EAL:   probe driver: 15b3:1007 librte_pmd_mlx4
    PMD: librte_pmd_mlx4: PCI information matches, using device "mlx4_0" (VF: false)
    PMD: librte_pmd_mlx4: 2 port(s) detected
    PMD: librte_pmd_mlx4: port 1 MAC address is 00:02:c9:b5:b7:50
    PMD: librte_pmd_mlx4: port 2 MAC address is 00:02:c9:b5:b7:51
    EAL: PCI device 0000:84:00.0 on NUMA socket 1
    EAL:   probe driver: 15b3:1007 librte_pmd_mlx4
    PMD: librte_pmd_mlx4: PCI information matches, using device "mlx4_1" (VF: false)
    PMD: librte_pmd_mlx4: 2 port(s) detected
    PMD: librte_pmd_mlx4: port 1 MAC address is 00:02:c9:b5:ba:b0
    PMD: librte_pmd_mlx4: port 2 MAC address is 00:02:c9:b5:ba:b1
    Interactive-mode selected
    Configuring Port 0 (socket 0)
    PMD: librte_pmd_mlx4: 0x867d60: TX queues number update: 0 -> 2
    PMD: librte_pmd_mlx4: 0x867d60: RX queues number update: 0 -> 2
    Port 0: 00:02:C9:B5:B7:50
    Configuring Port 1 (socket 0)
    PMD: librte_pmd_mlx4: 0x867da0: TX queues number update: 0 -> 2
    PMD: librte_pmd_mlx4: 0x867da0: RX queues number update: 0 -> 2
    Port 1: 00:02:C9:B5:B7:51
    Configuring Port 2 (socket 0)
    PMD: librte_pmd_mlx4: 0x867de0: TX queues number update: 0 -> 2
    PMD: librte_pmd_mlx4: 0x867de0: RX queues number update: 0 -> 2
    Port 2: 00:02:C9:B5:BA:B0
    Configuring Port 3 (socket 0)
    PMD: librte_pmd_mlx4: 0x867e20: TX queues number update: 0 -> 2
    PMD: librte_pmd_mlx4: 0x867e20: RX queues number update: 0 -> 2
    Port 3: 00:02:C9:B5:BA:B1
    Checking link statuses...
    Port 0 Link Up - speed 10000 Mbps - full-duplex
    Port 1 Link Up - speed 40000 Mbps - full-duplex
    Port 2 Link Up - speed 10000 Mbps - full-duplex
    Port 3 Link Up - speed 40000 Mbps - full-duplex
    Done
    testpmd>
    

MLX5 poll mode driver

The MLX5 poll mode driver library (librte_pmd_mlx5) provides support for Mellanox ConnectX-4, Mellanox ConnectX-4 Lx and Mellanox ConnectX-5 families of 10/25/40/50/100 Gb/s adapters as well as their virtual functions (VF) in SR-IOV context.

Information and documentation about these adapters can be found on the Mellanox website. Help is also provided by the Mellanox community.

There is also a section dedicated to this poll mode driver.

Note

Due to external dependencies, this driver is disabled by default. It must be enabled manually by setting CONFIG_RTE_LIBRTE_MLX5_PMD=y and recompiling DPDK.

Implementation details

Besides its dependency on libibverbs (that implies libmlx5 and associated kernel support), librte_pmd_mlx5 relies heavily on system calls for control operations such as querying/updating the MTU and flow control parameters.

For security reasons and robustness, this driver only deals with virtual memory addresses. The way resources allocations are handled by the kernel combined with hardware specifications that allow it to handle virtual memory addresses directly ensure that DPDK applications cannot access random physical memory (or memory that does not belong to the current process).

This capability allows the PMD to coexist with kernel network interfaces which remain functional, although they stop receiving unicast packets as long as they share the same MAC address. This means legacy linux control tools (for example: ethtool, ifconfig and more) can operate on the same network interfaces that owned by the DPDK application.

Enabling librte_pmd_mlx5 causes DPDK applications to be linked against libibverbs.

Features

  • Multi arch support: x86_64, POWER8, ARMv8.
  • Multiple TX and RX queues.
  • Support for scattered TX and RX frames.
  • IPv4, IPv6, TCPv4, TCPv6, UDPv4 and UDPv6 RSS on any number of queues.
  • Several RSS hash keys, one for each flow type.
  • Configurable RETA table.
  • Support for multiple MAC addresses.
  • VLAN filtering.
  • RX VLAN stripping.
  • TX VLAN insertion.
  • RX CRC stripping configuration.
  • Promiscuous mode.
  • Multicast promiscuous mode.
  • Hardware checksum offloads.
  • Flow director (RTE_FDIR_MODE_PERFECT, RTE_FDIR_MODE_PERFECT_MAC_VLAN and RTE_ETH_FDIR_REJECT).
  • Flow API.
  • Multiple process.
  • KVM and VMware ESX SR-IOV modes are supported.
  • RSS hash result is supported.
  • Hardware TSO.
  • Hardware checksum TX offload for VXLAN and GRE.
  • RX interrupts.
  • Statistics query including Basic, Extended and per queue.
  • Rx HW timestamp.

Limitations

  • Inner RSS for VXLAN frames is not supported yet.

  • Port statistics through software counters only. Flow statistics are supported by hardware counters.

  • Hardware checksum RX offloads for VXLAN inner header are not supported yet.

  • Forked secondary process not supported.

  • Flow pattern without any specific vlan will match for vlan packets as well:

    When VLAN spec is not specified in the pattern, the matching rule will be created with VLAN as a wild card. Meaning, the flow rule:

    flow create 0 ingress pattern eth / vlan vid is 3 / ipv4 / end ...
    

    Will only match vlan packets with vid=3. and the flow rules:

    flow create 0 ingress pattern eth / ipv4 / end ...
    

    Or:

    flow create 0 ingress pattern eth / vlan / ipv4 / end ...
    

    Will match any ipv4 packet (VLAN included).

  • A multi segment packet must have less than 6 segments in case the Tx burst function is set to multi-packet send or Enhanced multi-packet send. Otherwise it must have less than 50 segments.

  • Count action for RTE flow is only supported in Mellanox OFED 4.2.

  • Flows with a VXLAN Network Identifier equal (or ends to be equal) to 0 are not supported.

  • VXLAN TSO and checksum offloads are not supported on VM.

Configuration

Compilation options

These options can be modified in the .config file.

  • CONFIG_RTE_LIBRTE_MLX5_PMD (default n)

    Toggle compilation of librte_pmd_mlx5 itself.

  • CONFIG_RTE_LIBRTE_MLX5_DEBUG (default n)

    Toggle debugging code and stricter compilation flags. Enabling this option adds additional run-time checks and debugging messages at the cost of lower performance.

  • CONFIG_RTE_LIBRTE_MLX5_TX_MP_CACHE (default 8)

    Maximum number of cached memory pools (MPs) per TX queue. Each MP from which buffers are to be transmitted must be associated to memory regions (MRs). This is a slow operation that must be cached.

    This value is always 1 for RX queues since they use a single MP.

Environment variables
  • MLX5_PMD_ENABLE_PADDING

    Enables HW packet padding in PCI bus transactions.

    When packet size is cache aligned and CRC stripping is enabled, 4 fewer bytes are written to the PCI bus. Enabling padding makes such packets aligned again.

    In cases where PCI bandwidth is the bottleneck, padding can improve performance by 10%.

    This is disabled by default since this can also decrease performance for unaligned packet sizes.

  • MLX5_SHUT_UP_BF

    Configures HW Tx doorbell register as IO-mapped.

    By default, the HW Tx doorbell is configured as a write-combining register. The register would be flushed to HW usually when the write-combining buffer becomes full, but it depends on CPU design.

    Except for vectorized Tx burst routines, a write memory barrier is enforced after updating the register so that the update can be immediately visible to HW.

    When vectorized Tx burst is called, the barrier is set only if the burst size is not aligned to MLX5_VPMD_TX_MAX_BURST. However, setting this environmental variable will bring better latency even though the maximum throughput can slightly decline.

Run-time configuration
  • librte_pmd_mlx5 brings kernel network interfaces up during initialization because it is affected by their state. Forcing them down prevents packets reception.

  • ethtool operations on related kernel interfaces also affect the PMD.

  • rxq_cqe_comp_en parameter [int]

    A nonzero value enables the compression of CQE on RX side. This feature allows to save PCI bandwidth and improve performance. Enabled by default.

    Supported on:

    • x86_64 with ConnectX-4, ConnectX-4 LX and ConnectX-5.
    • POWER8 and ARMv8 with ConnectX-4 LX and ConnectX-5.
  • txq_inline parameter [int]

    Amount of data to be inlined during TX operations. Improves latency. Can improve PPS performance when PCI back pressure is detected and may be useful for scenarios involving heavy traffic on many queues.

    Because additional software logic is necessary to handle this mode, this option should be used with care, as it can lower performance when back pressure is not expected.

  • txqs_min_inline parameter [int]

    Enable inline send only when the number of TX queues is greater or equal to this value.

    This option should be used in combination with txq_inline above.

    On ConnectX-4, ConnectX-4 LX and ConnectX-5 without Enhanced MPW:

    • Disabled by default.
    • In case txq_inline is set recommendation is 4.

    On ConnectX-5 with Enhanced MPW:

    • Set to 8 by default.
  • txq_mpw_en parameter [int]

    A nonzero value enables multi-packet send (MPS) for ConnectX-4 Lx and enhanced multi-packet send (Enhanced MPS) for ConnectX-5. MPS allows the TX burst function to pack up multiple packets in a single descriptor session in order to save PCI bandwidth and improve performance at the cost of a slightly higher CPU usage. When txq_inline is set along with txq_mpw_en, TX burst function tries to copy entire packet data on to TX descriptor instead of including pointer of packet only if there is enough room remained in the descriptor. txq_inline sets per-descriptor space for either pointers or inlined packets. In addition, Enhanced MPS supports hybrid mode - mixing inlined packets and pointers in the same descriptor.

    This option cannot be used in conjunction with tso below. When tso is set, txq_mpw_en is disabled.

    It is currently only supported on the ConnectX-4 Lx and ConnectX-5 families of adapters. Enabled by default.

  • txq_mpw_hdr_dseg_en parameter [int]

    A nonzero value enables including two pointers in the first block of TX descriptor. This can be used to lessen CPU load for memory copy.

    Effective only when Enhanced MPS is supported. Disabled by default.

  • txq_max_inline_len parameter [int]

    Maximum size of packet to be inlined. This limits the size of packet to be inlined. If the size of a packet is larger than configured value, the packet isn’t inlined even though there’s enough space remained in the descriptor. Instead, the packet is included with pointer.

    Effective only when Enhanced MPS is supported. The default value is 256.

  • tso parameter [int]

    A nonzero value enables hardware TSO. When hardware TSO is enabled, packets marked with TCP segmentation offload will be divided into segments by the hardware. Disabled by default.

  • tx_vec_en parameter [int]

    A nonzero value enables Tx vector on ConnectX-5 only NIC if the number of global Tx queues on the port is lesser than MLX5_VPMD_MIN_TXQS.

    Enabled by default on ConnectX-5.

  • rx_vec_en parameter [int]

    A nonzero value enables Rx vector if the port is not configured in multi-segment otherwise this parameter is ignored.

    Enabled by default.

Prerequisites

This driver relies on external libraries and kernel drivers for resources allocations and initialization. The following dependencies are not part of DPDK and must be installed separately:

  • libibverbs

    User space Verbs framework used by librte_pmd_mlx5. This library provides a generic interface between the kernel and low-level user space drivers such as libmlx5.

    It allows slow and privileged operations (context initialization, hardware resources allocations) to be managed by the kernel and fast operations to never leave user space.

  • libmlx5

    Low-level user space driver library for Mellanox ConnectX-4/ConnectX-5 devices, it is automatically loaded by libibverbs.

    This library basically implements send/receive calls to the hardware queues.

  • Kernel modules

    They provide the kernel-side Verbs API and low level device drivers that manage actual hardware initialization and resources sharing with user space processes.

    Unlike most other PMDs, these modules must remain loaded and bound to their devices:

    • mlx5_core: hardware driver managing Mellanox ConnectX-4/ConnectX-5 devices and related Ethernet kernel network devices.
    • mlx5_ib: InifiniBand device driver.
    • ib_uverbs: user space driver for Verbs (entry point for libibverbs).
  • Firmware update

    Mellanox OFED releases include firmware updates for ConnectX-4/ConnectX-5 adapters.

    Because each release provides new features, these updates must be applied to match the kernel modules and libraries they come with.

Note

Both libraries are BSD and GPL licensed. Linux kernel modules are GPL licensed.

Installation

Either RDMA Core library with a recent enough Linux kernel release (recommended) or Mellanox OFED, which provides compatibility with older releases.

RMDA Core with Linux Kernel
Mellanox OFED
  • Mellanox OFED version: 4.2.
  • firmware version:
    • ConnectX-4: 12.21.1000 and above.
    • ConnectX-4 Lx: 14.21.1000 and above.
    • ConnectX-5: 16.21.1000 and above.
    • ConnectX-5 Ex: 16.21.1000 and above.

While these libraries and kernel modules are available on OpenFabrics Alliance’s website and provided by package managers on most distributions, this PMD requires Ethernet extensions that may not be supported at the moment (this is a work in progress).

Mellanox OFED includes the necessary support and should be used in the meantime. For DPDK, only libibverbs, libmlx5, mlnx-ofed-kernel packages and firmware updates are required from that distribution.

Note

Several versions of Mellanox OFED are available. Installing the version this DPDK release was developed and tested against is strongly recommended. Please check the prerequisites.

Supported NICs

  • Mellanox(R) ConnectX(R)-4 10G MCX4111A-XCAT (1x10G)
  • Mellanox(R) ConnectX(R)-4 10G MCX4121A-XCAT (2x10G)
  • Mellanox(R) ConnectX(R)-4 25G MCX4111A-ACAT (1x25G)
  • Mellanox(R) ConnectX(R)-4 25G MCX4121A-ACAT (2x25G)
  • Mellanox(R) ConnectX(R)-4 40G MCX4131A-BCAT (1x40G)
  • Mellanox(R) ConnectX(R)-4 40G MCX413A-BCAT (1x40G)
  • Mellanox(R) ConnectX(R)-4 40G MCX415A-BCAT (1x40G)
  • Mellanox(R) ConnectX(R)-4 50G MCX4131A-GCAT (1x50G)
  • Mellanox(R) ConnectX(R)-4 50G MCX413A-GCAT (1x50G)
  • Mellanox(R) ConnectX(R)-4 50G MCX414A-BCAT (2x50G)
  • Mellanox(R) ConnectX(R)-4 50G MCX415A-GCAT (2x50G)
  • Mellanox(R) ConnectX(R)-4 50G MCX416A-BCAT (2x50G)
  • Mellanox(R) ConnectX(R)-4 50G MCX416A-GCAT (2x50G)
  • Mellanox(R) ConnectX(R)-4 50G MCX415A-CCAT (1x100G)
  • Mellanox(R) ConnectX(R)-4 100G MCX416A-CCAT (2x100G)
  • Mellanox(R) ConnectX(R)-4 Lx 10G MCX4121A-XCAT (2x10G)
  • Mellanox(R) ConnectX(R)-4 Lx 25G MCX4121A-ACAT (2x25G)
  • Mellanox(R) ConnectX(R)-5 100G MCX556A-ECAT (2x100G)
  • Mellanox(R) ConnectX(R)-5 Ex EN 100G MCX516A-CDAT (2x100G)

Quick Start Guide on OFED

  1. Download latest Mellanox OFED. For more info check the prerequisites.

  2. Install the required libraries and kernel modules either by installing only the required set, or by installing the entire Mellanox OFED:

    ./mlnxofedinstall --upstream-libs --dpdk
    
  3. Verify the firmware is the correct one:

    ibv_devinfo
    
  4. Verify all ports links are set to Ethernet:

    mlxconfig -d <mst device> query | grep LINK_TYPE
    LINK_TYPE_P1                        ETH(2)
    LINK_TYPE_P2                        ETH(2)
    

    Link types may have to be configured to Ethernet:

    mlxconfig -d <mst device> set LINK_TYPE_P1/2=1/2/3
    
    * LINK_TYPE_P1=<1|2|3> , 1=Infiniband 2=Ethernet 3=VPI(auto-sense)
    

    For hypervisors verify SR-IOV is enabled on the NIC:

    mlxconfig -d <mst device> query | grep SRIOV_EN
    SRIOV_EN                            True(1)
    

    If needed, set enable the set the relevant fields:

    mlxconfig -d <mst device> set SRIOV_EN=1 NUM_OF_VFS=16
    mlxfwreset -d <mst device> reset
    
  5. Restart the driver:

    /etc/init.d/openibd restart
    

    or:

    service openibd restart
    

    If link type was changed, firmware must be reset as well:

    mlxfwreset -d <mst device> reset
    

    For hypervisors, after reset write the sysfs number of virtual functions needed for the PF.

    To dynamically instantiate a given number of virtual functions (VFs):

    echo [num_vfs] > /sys/class/infiniband/mlx5_0/device/sriov_numvfs
    
  6. Compile DPDK and you are ready to go. See instructions on Development Kit Build System

Performance tuning

  1. Configure aggressive CQE Zipping for maximum performance:
mlxconfig -d <mst device> s CQE_COMPRESSION=1

To set it back to the default CQE Zipping mode use:

mlxconfig -d <mst device> s CQE_COMPRESSION=0
  1. In case of virtualization:

    • Make sure that hypervisor kernel is 3.16 or newer.
    • Configure boot with iommu=pt.
    • Use 1G huge pages.
    • Make sure to allocate a VM on huge pages.
    • Make sure to set CPU pinning.
  2. Use the CPU near local NUMA node to which the PCIe adapter is connected, for better performance. For VMs, verify that the right CPU and NUMA node are pinned according to the above. Run:

    lstopo-no-graphics
    

    to identify the NUMA node to which the PCIe adapter is connected.

  3. If more than one adapter is used, and root complex capabilities allow to put both adapters on the same NUMA node without PCI bandwidth degradation, it is recommended to locate both adapters on the same NUMA node. This in order to forward packets from one to the other without NUMA performance penalty.

  4. Disable pause frames:

    ethtool -A <netdev> rx off tx off
    
  5. Verify IO non-posted prefetch is disabled by default. This can be checked via the BIOS configuration. Please contact you server provider for more information about the settings.

Note

On some machines, depends on the machine integrator, it is beneficial to set the PCI max read request parameter to 1K. This can be done in the following way:

To query the read request size use:

setpci -s <NIC PCI address> 68.w

If the output is different than 3XXX, set it by:

setpci -s <NIC PCI address> 68.w=3XXX

The XXX can be different on different systems. Make sure to configure according to the setpci output.

Notes for testpmd

Compared to librte_pmd_mlx4 that implements a single RSS configuration per port, librte_pmd_mlx5 supports per-protocol RSS configuration.

Since testpmd defaults to IP RSS mode and there is currently no command-line parameter to enable additional protocols (UDP and TCP as well as IP), the following commands must be entered from its CLI to get the same behavior as librte_pmd_mlx4:

> port stop all
> port config all rss all
> port start all

Usage example

This section demonstrates how to launch testpmd with Mellanox ConnectX-4/ConnectX-5 devices managed by librte_pmd_mlx5.

  1. Load the kernel modules:

    modprobe -a ib_uverbs mlx5_core mlx5_ib
    

    Alternatively if MLNX_OFED is fully installed, the following script can be run:

    /etc/init.d/openibd restart
    

    Note

    User space I/O kernel modules (uio and igb_uio) are not used and do not have to be loaded.

  2. Make sure Ethernet interfaces are in working order and linked to kernel verbs. Related sysfs entries should be present:

    ls -d /sys/class/net/*/device/infiniband_verbs/uverbs* | cut -d / -f 5
    

    Example output:

    eth30
    eth31
    eth32
    eth33
    
  3. Optionally, retrieve their PCI bus addresses for whitelisting:

    {
        for intf in eth2 eth3 eth4 eth5;
        do
            (cd "/sys/class/net/${intf}/device/" && pwd -P);
        done;
    } |
    sed -n 's,.*/\(.*\),-w \1,p'
    

    Example output:

    -w 0000:05:00.1
    -w 0000:06:00.0
    -w 0000:06:00.1
    -w 0000:05:00.0
    
  4. Request huge pages:

    echo 1024 > /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages/nr_hugepages
    
  5. Start testpmd with basic parameters:

    testpmd -l 8-15 -n 4 -w 05:00.0 -w 05:00.1 -w 06:00.0 -w 06:00.1 -- --rxq=2 --txq=2 -i
    

    Example output:

    [...]
    EAL: PCI device 0000:05:00.0 on NUMA socket 0
    EAL:   probe driver: 15b3:1013 librte_pmd_mlx5
    PMD: librte_pmd_mlx5: PCI information matches, using device "mlx5_0" (VF: false)
    PMD: librte_pmd_mlx5: 1 port(s) detected
    PMD: librte_pmd_mlx5: port 1 MAC address is e4:1d:2d:e7:0c:fe
    EAL: PCI device 0000:05:00.1 on NUMA socket 0
    EAL:   probe driver: 15b3:1013 librte_pmd_mlx5
    PMD: librte_pmd_mlx5: PCI information matches, using device "mlx5_1" (VF: false)
    PMD: librte_pmd_mlx5: 1 port(s) detected
    PMD: librte_pmd_mlx5: port 1 MAC address is e4:1d:2d:e7:0c:ff
    EAL: PCI device 0000:06:00.0 on NUMA socket 0
    EAL:   probe driver: 15b3:1013 librte_pmd_mlx5
    PMD: librte_pmd_mlx5: PCI information matches, using device "mlx5_2" (VF: false)
    PMD: librte_pmd_mlx5: 1 port(s) detected
    PMD: librte_pmd_mlx5: port 1 MAC address is e4:1d:2d:e7:0c:fa
    EAL: PCI device 0000:06:00.1 on NUMA socket 0
    EAL:   probe driver: 15b3:1013 librte_pmd_mlx5
    PMD: librte_pmd_mlx5: PCI information matches, using device "mlx5_3" (VF: false)
    PMD: librte_pmd_mlx5: 1 port(s) detected
    PMD: librte_pmd_mlx5: port 1 MAC address is e4:1d:2d:e7:0c:fb
    Interactive-mode selected
    Configuring Port 0 (socket 0)
    PMD: librte_pmd_mlx5: 0x8cba80: TX queues number update: 0 -> 2
    PMD: librte_pmd_mlx5: 0x8cba80: RX queues number update: 0 -> 2
    Port 0: E4:1D:2D:E7:0C:FE
    Configuring Port 1 (socket 0)
    PMD: librte_pmd_mlx5: 0x8ccac8: TX queues number update: 0 -> 2
    PMD: librte_pmd_mlx5: 0x8ccac8: RX queues number update: 0 -> 2
    Port 1: E4:1D:2D:E7:0C:FF
    Configuring Port 2 (socket 0)
    PMD: librte_pmd_mlx5: 0x8cdb10: TX queues number update: 0 -> 2
    PMD: librte_pmd_mlx5: 0x8cdb10: RX queues number update: 0 -> 2
    Port 2: E4:1D:2D:E7:0C:FA
    Configuring Port 3 (socket 0)
    PMD: librte_pmd_mlx5: 0x8ceb58: TX queues number update: 0 -> 2
    PMD: librte_pmd_mlx5: 0x8ceb58: RX queues number update: 0 -> 2
    Port 3: E4:1D:2D:E7:0C:FB
    Checking link statuses...
    Port 0 Link Up - speed 40000 Mbps - full-duplex
    Port 1 Link Up - speed 40000 Mbps - full-duplex
    Port 2 Link Up - speed 10000 Mbps - full-duplex
    Port 3 Link Up - speed 10000 Mbps - full-duplex
    Done
    testpmd>
    

MRVL Poll Mode Driver

The MRVL PMD (librte_pmd_mrvl) provides poll mode driver support for the Marvell PPv2 (Packet Processor v2) 1/10 Gbps adapter.

Detailed information about SoCs that use PPv2 can be obtained here:

Note

Due to external dependencies, this driver is disabled by default. It must be enabled manually by setting relevant configuration option manually. Please refer to Config File Options section for further details.

Features

Features of the MRVL PMD are:

  • Speed capabilities
  • Link status
  • Queue start/stop
  • MTU update
  • Jumbo frame
  • Promiscuous mode
  • Allmulticast mode
  • Unicast MAC filter
  • Multicast MAC filter
  • RSS hash
  • VLAN filter
  • CRC offload
  • L3 checksum offload
  • L4 checksum offload
  • Packet type parsing
  • Basic stats
  • QoS

Limitations

  • Number of lcores is limited to 9 by MUSDK internal design. If more lcores need to be allocated, locking will have to be considered. Number of available lcores can be changed via MRVL_MUSDK_HIFS_RESERVED define in mrvl_ethdev.c source file.
  • Flushing vlans added for filtering is not possible due to MUSDK missing functionality. Current workaround is to reset board so that PPv2 has a chance to start in a sane state.

Prerequisites

  • Custom Linux Kernel sources available here.

  • Out of tree mvpp2x_sysfs kernel module sources available here.

  • MUSDK (Marvell User-Space SDK) sources available here.

    MUSDK is a light-weight library that provides direct access to Marvell’s PPv2 (Packet Processor v2). Alternatively prebuilt MUSDK library can be requested from Marvell Extranet. Once approval has been granted, library can be found by typing musdk in the search box.

    MUSDK must be configured with the following features:

    --enable-bpool-dma=64
    
  • DPDK environment

    Follow the DPDK Getting Started Guide for Linux to setup DPDK environment.

Config File Options

The following options can be modified in the config file.

  • CONFIG_RTE_LIBRTE_MRVL_PMD (default n)

    Toggle compilation of the librte_pmd_mrvl driver.

  • CONFIG_RTE_MRVL_MUSDK_DMA_MEMSIZE (default 41943040)

    Size in bytes of the contiguous memory region that MUSDK will allocate for run-time DMA-able data buffers.

QoS Configuration

QoS configuration is done through external configuration file. Path to the file must be given as cfg in driver’s vdev parameter list.

Configuration syntax
[port <portnum> default]
default_tc = <default_tc>
qos_mode = <qos_priority>

[port <portnum> tc <traffic_class>]
rxq = <rx_queue_list>
pcp = <pcp_list>
dscp = <dscp_list>

[port <portnum> tc <traffic_class>]
rxq = <rx_queue_list>
pcp = <pcp_list>
dscp = <dscp_list>

Where:

  • <portnum>: DPDK Port number (0..n).
  • <default_tc>: Default traffic class (e.g. 0)
  • <qos_priority>: QoS priority for mapping (ip, vlan, ip/vlan or vlan/ip).
  • <traffic_class>: Traffic Class to be configured.
  • <rx_queue_list>: List of DPDK RX queues (e.g. 0 1 3-4)
  • <pcp_list>: List of PCP values to handle in particular TC (e.g. 0 1 3-4 7).
  • <dscp_list>: List of DSCP values to handle in particular TC (e.g. 0-12 32-48 63).

Setting PCP/DSCP values for the default TC is not required. All PCP/DSCP values not assigned explicitly to particular TC will be handled by the default TC.

Configuration file example
[port 0 default]
default_tc = 0
qos_mode = ip

[port 0 tc 0]
rxq = 0 1

[port 0 tc 1]
rxq = 2
pcp = 5 6 7
dscp = 26-38

[port 1 default]
default_tc = 0
qos_mode = vlan/ip

[port 1 tc 0]
rxq = 0

[port 1 tc 1]
rxq = 1 2
pcp = 5 6 7
dscp = 26-38
Usage example
./testpmd --vdev=eth_mrvl,iface=eth0,iface=eth2,cfg=/home/user/mrvl.conf \
  -c 7 -- -i -a --disable-hw-vlan-strip --rxq=2

Building DPDK

Driver needs precompiled MUSDK library during compilation. Please consult doc/musdk_get_started.txt for the detailed build instructions.

Before the DPDK build process the environmental variable LIBMUSDK_PATH with the path to the MUSDK installation directory needs to be exported.

Usage Example

MRVL PMD requires extra out of tree kernel modules to function properly. musdk_uio and mv_pp_uio sources are part of the MUSDK. Please consult doc/musdk_get_started.txt for the detailed build instructions. For mvpp2x_sysfs please consult Documentation/pp22_sysfs.txt for the detailed build instructions.

insmod musdk_uio.ko
insmod mv_pp_uio.ko
insmod mvpp2x_sysfs.ko

Additionally interfaces used by DPDK application need to be put up:

ip link set eth0 up
ip link set eth1 up

In order to run testpmd example application following command can be used:

./testpmd --vdev=eth_mrvl,iface=eth0,iface=eth2 -c 7 -- \
  --burst=128 --txd=2048 --rxd=1024 --rxq=2 --txq=2  --nb-cores=2 \
  -i -a --disable-hw-vlan-strip --rss-udp

NFP poll mode driver library

Netronome’s sixth generation of flow processors pack 216 programmable cores and over 100 hardware accelerators that uniquely combine packet, flow, security and content processing in a single device that scales up to 400 Gbps.

This document explains how to use DPDK with the Netronome Poll Mode Driver (PMD) supporting Netronome’s Network Flow Processor 6xxx (NFP-6xxx) and Netronome’s Flow Processor 4xxx (NFP-4xxx).

NFP is a SRIOV capable device and the PMD driver supports the physical function (PF) and virtual functions (VFs).

Dependencies

Before using the Netronome’s DPDK PMD some NFP configuration, which is not related to DPDK, is required. The system requires installation of Netronome’s BSP (Board Support Package) along with some specific NFP firmware application. Netronome’s NSP ABI version should be 0.20 or higher.

If you have a NFP device you should already have the code and documentation for doing all this configuration. Contact support@netronome.com to obtain the latest available firmware.

The NFP Linux netdev kernel driver for VFs is part of vanilla kernel since kernel version 4.5, and support for the PF since kernel version 4.11. Support for older kernels can be obtained on Github at https://github.com/Netronome/nfp-drv-kmods along with build instructions.

NFP PMD needs to be used along with UIO igb_uio or VFIO (vfio-pci) Linux kernel driver.

Building the software

Netronome’s PMD code is provided in the drivers/net/nfp directory. Although NFP PMD has Netronome´s BSP dependencies, it is possible to compile it along with other DPDK PMDs even if no BSP was installed before. Of course, a DPDK app will require such a BSP installed for using the NFP PMD, along with a specific NFP firmware application.

Default PMD configuration is at common_linuxapp configuration file:

  • CONFIG_RTE_LIBRTE_NFP_PMD=y

Once DPDK is built all the DPDK apps and examples include support for the NFP PMD.

Driver compilation and testing

Refer to the document compiling and testing a PMD for a NIC for details.

Using the PF

NFP PMD has support for using the NFP PF as another DPDK port, but it does not have any functionality for controlling VFs. In fact, it is not possible to use the PMD with the VFs if the PF is being used by DPDK, that is, with the NFP PF bound to igb_uio or vfio-pci kernel drivers. Future DPDK version will have a PMD able to work with the PF and VFs at the same time and with the PF implementing VF management along with other PF-only functionalities/offloads.

The PMD PF has extra work to do which will delay the DPDK app initialization like checking if a firmware is already available in the device, uploading the firmware if necessary, and configure the Link state properly when starting or stopping a PF port. Note that firmware upload is not always necessary which is the main delay for NFP PF PMD initialization.

Depending on the Netronome product installed in the system, firmware files should be available under /lib/firmware/netronome. DPDK PMD supporting the PF requires a specific link, /lib/firmware/netronome/nic_dpdk_default.nffw, which should be created automatically with Netronome’s Agilio products installation.

PF multiport support

Some NFP cards support several physical ports with just one single PCI device. DPDK core is designed with the 1:1 relationship between PCI devices and DPDK ports, so NFP PMD PF support requires handling the multiport case specifically. During NFP PF initialization, the PMD will extract the information about the number of PF ports from the firmware and will create as many DPDK ports as needed.

Because the unusual relationship between a single PCI device and several DPDK ports, there are some limitations when using more than one PF DPDK ports: there is no support for RX interrupts and it is not possible either to use those PF ports with the device hotplug functionality.

System configuration

  1. Enable SR-IOV on the NFP device: The current NFP PMD supports the PF and the VFs on a NFP device. However, it is not possible to work with both at the same time because the VFs require the PF being bound to the NFP PF Linux netdev driver. Make sure you are working with a kernel with NFP PF support or get the drivers from the above Github repository and follow the instructions for building and installing it.

    Virtual Functions need to be enabled before they can be used with the PMD. Before enabling the VFs it is useful to obtain information about the current NFP PCI device detected by the system:

    lspci -d19ee:
    

    Now, for example, configure two virtual functions on a NFP-6xxx device whose PCI system identity is “0000:03:00.0”:

    echo 2 > /sys/bus/pci/devices/0000:03:00.0/sriov_numvfs
    

    The result of this command may be shown using lspci again:

    lspci -d19ee: -k
    

    Two new PCI devices should appear in the output of the above command. The -k option shows the device driver, if any, that devices are bound to. Depending on the modules loaded at this point the new PCI devices may be bound to nfp_netvf driver.

OCTEONTX Poll Mode driver

The OCTEONTX ETHDEV PMD (librte_pmd_octeontx) provides poll mode ethdev driver support for the inbuilt network device found in the Cavium OCTEONTX SoC family as well as their virtual functions (VF) in SR-IOV context.

More information can be found at Cavium, Inc Official Website.

Features

Features of the OCTEONTX Ethdev PMD are:

  • Packet type information
  • Promiscuous mode
  • Port hardware statistics
  • Jumbo frames
  • Link state information
  • SR-IOV VF
  • Multiple queues for TX
  • Lock-free Tx queue
  • HW offloaded ethdev Rx queue to eventdev event queue packet injection

Supported OCTEONTX SoCs

  • CN83xx

Unsupported features

The features supported by the device and not yet supported by this PMD include:

  • Receive Side Scaling (RSS)
  • Scattered and gather for TX and RX
  • Ingress classification support
  • Egress hierarchical scheduling, traffic shaping, and marking

Prerequisites

See OCTEONTX Board Support Package for setup information.

Pre-Installation Configuration

Config File Options

The following options can be modified in the config file. Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_LIBRTE_OCTEONTX_PMD (default y)

    Toggle compilation of the librte_pmd_octeontx driver.

  • CONFIG_RTE_LIBRTE_OCTEONTX_DEBUG_DRIVER (default n)

    Toggle display of generic debugging messages

  • CONFIG_RTE_LIBRTE_OCTEONTX_DEBUG_INIT (default n)

    Toggle display of initialization related messages.

  • CONFIG_RTE_LIBRTE_OCTEONTX_DEBUG_RX (default n)

    Toggle display of receive path message

  • CONFIG_RTE_LIBRTE_OCTEONTX_DEBUG_TX (default n)

    Toggle display of transmit path message

  • CONFIG_RTE_LIBRTE_OCTEONTX_DEBUG_MBOX (default n)

    Toggle display of mbox related message

Driver compilation and testing

Refer to the document compiling and testing a PMD for a NIC for details.

To compile the OCTEONTX PMD for Linux arm64 gcc target, run the following make command:

cd <DPDK-source-directory>
make config T=arm64-thunderx-linuxapp-gcc install
  1. Running testpmd:

    Follow instructions available in the document compiling and testing a PMD for a NIC to run testpmd.

    Example output:

    ./arm64-thunderx-linuxapp-gcc/app/testpmd -c 700 \
              --base-virtaddr=0x100000000000 \
              --mbuf-pool-ops-name="octeontx_fpavf" \
              --vdev='event_octeontx' \
              --vdev='eth_octeontx,nr_port=2' \
              -- --rxq=1 --txq=1 --nb-core=2 --total-num-mbufs=16384 \
              --disable-hw-vlan-filter -i
    .....
    EAL: Detected 24 lcore(s)
    EAL: Probing VFIO support...
    EAL: VFIO support initialized
    .....
    EAL: PCI device 0000:07:00.1 on NUMA socket 0
    EAL:   probe driver: 177d:a04b octeontx_ssovf
    .....
    EAL: PCI device 0001:02:00.7 on NUMA socket 0
    EAL:   probe driver: 177d:a0dd octeontx_pkivf
    .....
    EAL: PCI device 0001:03:01.0 on NUMA socket 0
    EAL:   probe driver: 177d:a049 octeontx_pkovf
    .....
    PMD: octeontx_probe(): created ethdev eth_octeontx for port 0
    PMD: octeontx_probe(): created ethdev eth_octeontx for port 1
    .....
    Configuring Port 0 (socket 0)
    Port 0: 00:0F:B7:11:94:46
    Configuring Port 1 (socket 0)
    Port 1: 00:0F:B7:11:94:47
    .....
    Checking link statuses...
    Port 0 Link Up - speed 40000 Mbps - full-duplex
    Port 1 Link Up - speed 40000 Mbps - full-duplex
    Done
    testpmd>
    

Initialization

The octeontx ethdev pmd is exposed as a vdev device which consists of a set of PKI and PKO PCIe VF devices. On EAL initialization, PKI/PKO PCIe VF devices will be probed and then the vdev device can be created from the application code, or from the EAL command line based on the number of probed/bound PKI/PKO PCIe VF device to DPDK by

  • Invoking rte_vdev_init("eth_octeontx") from the application
  • Using --vdev="eth_octeontx" in the EAL options, which will call rte_vdev_init() internally
Device arguments

Each ethdev port is mapped to a physical port(LMAC), Application can specify the number of interesting ports with nr_ports argument.

Dependency

eth_octeontx pmd is depend on event_octeontx eventdev device and octeontx_fpavf external mempool handler.

Example:

./your_dpdk_application --mbuf-pool-ops-name="octeontx_fpavf" \
            --vdev='event_octeontx' \
            --vdev="eth_octeontx,nr_port=2"

Limitations

octeontx_fpavf external mempool handler dependency

The OCTEONTX SoC family NIC has inbuilt HW assisted external mempool manager. This driver will only work with octeontx_fpavf external mempool handler as it is the most performance effective way for packet allocation and Tx buffer recycling on OCTEONTX SoC platform.

CRC striping

The OCTEONTX SoC family NICs strip the CRC for every packets coming into the host interface. So, CRC will be stripped even when the rxmode.hw_strip_crc member is set to 0 in struct rte_eth_conf.

Maximum packet length

The OCTEONTX SoC family NICs support a maximum of a 32K jumbo frame. The value is fixed and cannot be changed. So, even when the rxmode.max_rx_pkt_len member of struct rte_eth_conf is set to a value lower than 32k, frames up to 32k bytes can still reach the host interface.

QEDE Poll Mode Driver

The QEDE poll mode driver library (librte_pmd_qede) implements support for QLogic FastLinQ QL4xxxx 10G/25G/40G/50G/100G Intelligent Ethernet Adapters (IEA) and Converged Network Adapters (CNA) family of adapters as well as SR-IOV virtual functions (VF). It is supported on several standard Linux distros like RHEL7.x, SLES12.x and Ubuntu. It is compile-tested under FreeBSD OS.

More information can be found at QLogic Corporation’s Website.

Supported Features

  • Unicast/Multicast filtering
  • Promiscuous mode
  • Allmulti mode
  • Port hardware statistics
  • Jumbo frames
  • Multiple MAC address
  • MTU change
  • Default pause flow control
  • Multiprocess aware
  • Scatter-Gather
  • Multiple Rx/Tx queues
  • RSS (with RETA/hash table/key)
  • TSS
  • Stateless checksum offloads (IPv4/IPv6/TCP/UDP)
  • LRO/TSO
  • VLAN offload - Filtering and stripping
  • N-tuple filter and flow director (limited support)
  • NPAR (NIC Partitioning)
  • SR-IOV VF
  • VXLAN tunneling offload
  • MPLSoUDP Tx tunnel offload

Non-supported Features

  • SR-IOV PF
  • GENEVE and NVGRE Tunneling offloads

Supported QLogic Adapters

  • QLogic FastLinQ QL4xxxx 10G/25G/40G/50G/100G Intelligent Ethernet Adapters (IEA) and Converged Network Adapters (CNA)

Prerequisites

  • Requires storm firmware version 8.30.12.0. Firmware may be available inbox in certain newer Linux distros under the standard directory E.g. /lib/firmware/qed/qed_init_values-8.30.12.0.bin If the required firmware files are not available then download it from QLogic Driver Download Center. For downloading firmware file, select adapter category, model and DPDK Poll Mode Driver.
  • Requires management firmware (MFW) version 8.30.x.x or higher to be flashed on to the adapter. If the required management firmware is not available then download from QLogic Driver Download Center. For downloading firmware upgrade utility, select adapter category, model and Linux distro. To flash the management firmware refer to the instructions in the QLogic Firmware Upgrade Utility Readme document.
  • SR-IOV requires Linux PF driver version 8.20.x.x or higher. If the required PF driver is not available then download it from QLogic Driver Download Center. For downloading PF driver, select adapter category, model and Linux distro.
Performance note
  • For better performance, it is recommended to use 4K or higher RX/TX rings.
Config File Options

The following options can be modified in the .config file. Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_LIBRTE_QEDE_PMD (default y)

    Toggle compilation of QEDE PMD driver.

  • CONFIG_RTE_LIBRTE_QEDE_DEBUG_INFO (default n)

    Toggle display of generic debugging messages.

  • CONFIG_RTE_LIBRTE_QEDE_DEBUG_DRIVER (default n)

    Toggle display of ecore related messages.

  • CONFIG_RTE_LIBRTE_QEDE_DEBUG_TX (default n)

    Toggle display of transmit fast path run-time messages.

  • CONFIG_RTE_LIBRTE_QEDE_DEBUG_RX (default n)

    Toggle display of receive fast path run-time messages.

  • CONFIG_RTE_LIBRTE_QEDE_VF_TX_SWITCH (default “y”)

    A knob to control per-VF Tx switching feature.

  • CONFIG_RTE_LIBRTE_QEDE_FW (default “”)

    Gives absolute path of firmware file. Eg: "/lib/firmware/qed/qed_init_values-8.30.12.0.bin" Empty string indicates driver will pick up the firmware file from the default location /lib/firmware/qed. CAUTION this option is more for custom firmware, it is not recommended for use under normal condition.

Driver compilation and testing

Refer to the document compiling and testing a PMD for a NIC for details.

SR-IOV: Prerequisites and Sample Application Notes

This section provides instructions to configure SR-IOV with Linux OS.

Note: librte_pmd_qede will be used to bind to SR-IOV VF device and Linux native kernel driver (qede) will function as SR-IOV PF driver. Requires PF driver to be 8.10.x.x or higher.

  1. Verify SR-IOV and ARI capability is enabled on the adapter using lspci:

    lspci -s <slot> -vvv
    

    Example output:

    [...]
    Capabilities: [1b8 v1] Alternative Routing-ID Interpretation (ARI)
    [...]
    Capabilities: [1c0 v1] Single Root I/O Virtualization (SR-IOV)
    [...]
    Kernel driver in use: igb_uio
    
  2. Load the kernel module:

    modprobe qede
    

    Example output:

    systemd-udevd[4848]: renamed network interface eth0 to ens5f0
    systemd-udevd[4848]: renamed network interface eth1 to ens5f1
    
  3. Bring up the PF ports:

    ifconfig ens5f0 up
    ifconfig ens5f1 up
    
  4. Create VF device(s):

    Echo the number of VFs to be created into "sriov_numvfs" sysfs entry of the parent PF.

    Example output:

    echo 2 > /sys/devices/pci0000:00/0000:00:03.0/0000:81:00.0/sriov_numvfs
    
  5. Assign VF MAC address:

    Assign MAC address to the VF using iproute2 utility. The syntax is:

    ip link set <PF iface> vf <VF id> mac <macaddr>
    

    Example output:

    ip link set ens5f0 vf 0 mac 52:54:00:2f:9d:e8
    
  6. PCI Passthrough:

    The VF devices may be passed through to the guest VM using virt-manager or virsh. QEDE PMD should be used to bind the VF devices in the guest VM using the instructions from Driver compilation and testing section above.

  7. Running testpmd (Enable QEDE_DEBUG_INFO=y to view informational messages):

    Refer to the document compiling and testing a PMD for a NIC to run testpmd application.

    Example output:

      testpmd -l 0,4-11 -n 4 -- -i --nb-cores=8 --portmask=0xf --rxd=4096 \
      --txd=4096 --txfreet=4068 --enable-rx-cksum --rxq=4 --txq=4 \
      --rss-ip --rss-udp
    
      [...]
    
    EAL: PCI device 0000:84:00.0 on NUMA socket 1
    EAL:   probe driver: 1077:1634 rte_qede_pmd
    EAL:   Not managed by a supported kernel driver, skipped
    EAL: PCI device 0000:84:00.1 on NUMA socket 1
    EAL:   probe driver: 1077:1634 rte_qede_pmd
    EAL:   Not managed by a supported kernel driver, skipped
    EAL: PCI device 0000:88:00.0 on NUMA socket 1
    EAL:   probe driver: 1077:1656 rte_qede_pmd
    EAL:   PCI memory mapped at 0x7f738b200000
    EAL:   PCI memory mapped at 0x7f738b280000
    EAL:   PCI memory mapped at 0x7f738b300000
    PMD: Chip details : BB1
    PMD: Driver version : QEDE PMD 8.7.9.0_1.0.0
    PMD: Firmware version : 8.7.7.0
    PMD: Management firmware version : 8.7.8.0
    PMD: Firmware file : /lib/firmware/qed/qed_init_values_zipped-8.7.7.0.bin
    [QEDE PMD: (84:00.0:dpdk-port-0)]qede_common_dev_init:macaddr \
                                                        00:0e:1e:d2:09:9c
      [...]
    [QEDE PMD: (84:00.0:dpdk-port-0)]qede_tx_queue_setup:txq 0 num_desc 4096 \
                                                tx_free_thresh 4068 socket 0
    [QEDE PMD: (84:00.0:dpdk-port-0)]qede_tx_queue_setup:txq 1 num_desc 4096 \
                                                tx_free_thresh 4068 socket 0
    [QEDE PMD: (84:00.0:dpdk-port-0)]qede_tx_queue_setup:txq 2 num_desc 4096 \
                                                 tx_free_thresh 4068 socket 0
    [QEDE PMD: (84:00.0:dpdk-port-0)]qede_tx_queue_setup:txq 3 num_desc 4096 \
                                                 tx_free_thresh 4068 socket 0
    [QEDE PMD: (84:00.0:dpdk-port-0)]qede_rx_queue_setup:rxq 0 num_desc 4096 \
                                                rx_buf_size=2148 socket 0
    [QEDE PMD: (84:00.0:dpdk-port-0)]qede_rx_queue_setup:rxq 1 num_desc 4096 \
                                                rx_buf_size=2148 socket 0
    [QEDE PMD: (84:00.0:dpdk-port-0)]qede_rx_queue_setup:rxq 2 num_desc 4096 \
                                                rx_buf_size=2148 socket 0
    [QEDE PMD: (84:00.0:dpdk-port-0)]qede_rx_queue_setup:rxq 3 num_desc 4096 \
                                                rx_buf_size=2148 socket 0
    [QEDE PMD: (84:00.0:dpdk-port-0)]qede_dev_start:port 0
    [QEDE PMD: (84:00.0:dpdk-port-0)]qede_dev_start:link status: down
      [...]
    Checking link statuses...
    Port 0 Link Up - speed 25000 Mbps - full-duplex
    Port 1 Link Up - speed 25000 Mbps - full-duplex
    Port 2 Link Up - speed 25000 Mbps - full-duplex
    Port 3 Link Up - speed 25000 Mbps - full-duplex
    Done
    testpmd>
    

Solarflare libefx-based Poll Mode Driver

The SFC EFX PMD (librte_pmd_sfc_efx) provides poll mode driver support for Solarflare SFN7xxx and SFN8xxx family of 10/40 Gbps adapters. SFC EFX PMD has support for the latest Linux and FreeBSD operating systems.

More information can be found at Solarflare Communications website.

Features

SFC EFX PMD has support for:

  • Multiple transmit and receive queues
  • Link state information including link status change interrupt
  • IPv4/IPv6 TCP/UDP transmit checksum offload
  • Port hardware statistics
  • Extended statistics (see Solarflare Server Adapter User’s Guide for the statistics description)
  • Basic flow control
  • MTU update
  • Jumbo frames up to 9K
  • Promiscuous mode
  • Allmulticast mode
  • TCP segmentation offload (TSO)
  • Multicast MAC filter
  • IPv4/IPv6 TCP/UDP receive checksum offload
  • Received packet type information
  • Receive side scaling (RSS)
  • RSS hash
  • Scattered Rx DMA for packet that are larger that a single Rx descriptor
  • Deferred receive and transmit queue start
  • Transmit VLAN insertion (if running firmware variant supports it)
  • Flow API

Non-supported Features

The features not yet supported include:

  • Receive queue interupts
  • Priority-based flow control
  • Loopback
  • Configurable RX CRC stripping (always stripped)
  • Header split on receive
  • VLAN filtering
  • VLAN stripping
  • LRO

Limitations

Due to requirements on receive buffer alignment and usage of the receive buffer for the auxiliary packet information provided by the NIC up to extra 269 (14 bytes prefix plus up to 255 bytes for end padding) bytes may be required in the receive buffer. It should be taken into account when mbuf pool for receive is created.

Flow API support

Supported attributes:

  • Ingress

Supported pattern items:

  • VOID
  • ETH (exact match of source/destination addresses, individual/group match of destination address, EtherType)
  • VLAN (exact match of VID, double-tagging is supported)
  • IPV4 (exact match of source/destination addresses, IP transport protocol)
  • IPV6 (exact match of source/destination addresses, IP transport protocol)
  • TCP (exact match of source/destination ports)
  • UDP (exact match of source/destination ports)

Supported actions:

  • VOID
  • QUEUE
  • RSS

Validating flow rules depends on the firmware variant.

Ethernet destinaton individual/group match

Ethernet item supports I/G matching, if only the corresponding bit is set in the mask of destination address. If destinaton address in the spec is multicast, it matches all multicast (and broadcast) packets, oherwise it matches unicast packets that are not filtered by other flow rules.

Supported NICs

  • Solarflare Flareon [Ultra] Server Adapters:

    • Solarflare SFN8522 Dual Port SFP+ Server Adapter
    • Solarflare SFN8542 Dual Port QSFP+ Server Adapter
    • Solarflare SFN7002F Dual Port SFP+ Server Adapter
    • Solarflare SFN7004F Quad Port SFP+ Server Adapter
    • Solarflare SFN7042Q Dual Port QSFP+ Server Adapter
    • Solarflare SFN7122F Dual Port SFP+ Server Adapter
    • Solarflare SFN7124F Quad Port SFP+ Server Adapter
    • Solarflare SFN7142Q Dual Port QSFP+ Server Adapter
    • Solarflare SFN7322F Precision Time Synchronization Server Adapter

Prerequisites

  • Requires firmware version:

    • SFN7xxx: 4.7.1.1001 or higher
    • SFN8xxx: 6.0.2.1004 or higher

Visit Solarflare Support Downloads to get Solarflare Utilities (either Linux or FreeBSD) with the latest firmware. Follow instructions from Solarflare Server Adapter User’s Guide to update firmware and configure the adapter.

Pre-Installation Configuration

Config File Options

The following options can be modified in the .config file. Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_LIBRTE_SFC_EFX_PMD (default y)

    Enable compilation of Solarflare libefx-based poll-mode driver.

  • CONFIG_RTE_LIBRTE_SFC_EFX_DEBUG (default n)

    Enable compilation of the extra run-time consistency checks.

Per-Device Parameters

The following per-device parameters can be passed via EAL PCI device whitelist option like “-w 02:00.0,arg1=value1,...”.

Case-insensitive 1/y/yes/on or 0/n/no/off may be used to specify boolean parameters value.

  • rx_datapath [auto|efx|ef10] (default auto)

    Choose receive datapath implementation. auto allows the driver itself to make a choice based on firmware features available and required by the datapath implementation. efx chooses libefx-based datapath which supports Rx scatter. ef10 chooses EF10 (SFN7xxx, SFN8xxx) native datapath which is more efficient than libefx-based and provides richer packet type classification, but lacks Rx scatter support.

  • tx_datapath [auto|efx|ef10|ef10_simple] (default auto)

    Choose transmit datapath implementation. auto allows the driver itself to make a choice based on firmware features available and required by the datapath implementation. efx chooses libefx-based datapath which supports VLAN insertion (full-feature firmware variant only), TSO and multi-segment mbufs. Mbuf segments may come from different mempools, and mbuf reference counters are treated responsibly. ef10 chooses EF10 (SFN7xxx, SFN8xxx) native datapath which is more efficient than libefx-based but has no VLAN insertion and TSO support yet. Mbuf segments may come from different mempools, and mbuf reference counters are treated responsibly. ef10_simple chooses EF10 (SFN7xxx, SFN8xxx) native datapath which is even more faster then ef10 but does not support multi-segment mbufs, disallows multiple mempools and neglects mbuf reference counters.

  • perf_profile [auto|throughput|low-latency] (default throughput)

    Choose hardware tunning to be optimized for either throughput or low-latency. auto allows NIC firmware to make a choice based on installed licences and firmware variant configured using sfboot.

  • debug_init [bool] (default n)

    Enable extra logging during device initialization and startup.

  • mcdi_logging [bool] (default n)

    Enable extra logging of the communication with the NIC’s management CPU. The logging is done using RTE_LOG() with INFO level and PMD type. The format is consumed by the Solarflare netlogdecode cross-platform tool.

  • stats_update_period_ms [long] (default 1000)

    Adjust period in milliseconds to update port hardware statistics. The accepted range is 0 to 65535. The value of 0 may be used to disable periodic statistics update. One should note that it’s only possible to set an arbitrary value on SFN8xxx provided that firmware version is 6.2.1.1033 or higher, otherwise any positive value will select a fixed update period of 1000 milliseconds

SZEDATA2 poll mode driver library

The SZEDATA2 poll mode driver library implements support for the Netcope FPGA Boards (NFB-*), FPGA-based programmable NICs. The SZEDATA2 PMD uses interface provided by the libsze2 library to communicate with the NFB cards over the sze2 layer.

More information about the NFB cards and used technology (Netcope Development Kit) can be found on the Netcope Technologies website.

Note

This driver has external dependencies. Therefore it is disabled in default configuration files. It can be enabled by setting CONFIG_RTE_LIBRTE_PMD_SZEDATA2=y and recompiling.

Note

Currently the driver is supported only on x86_64 architectures. Only x86_64 versions of the external libraries are provided.

Prerequisites

This PMD requires kernel modules which are responsible for initialization and allocation of resources needed for sze2 layer function. Communication between PMD and kernel modules is mediated by libsze2 library. These kernel modules and library are not part of DPDK and must be installed separately:

  • libsze2 library

    The library provides API for initialization of sze2 transfers, receiving and transmitting data segments.

  • Kernel modules

    • combov3
    • szedata2_cv3

    Kernel modules manage initialization of hardware, allocation and sharing of resources for user space applications.

Information about getting the dependencies can be found here.

Configuration

These configuration options can be modified before compilation in the .config file:

  • CONFIG_RTE_LIBRTE_PMD_SZEDATA2 default value: n

    Value y enables compilation of szedata2 PMD.

  • CONFIG_RTE_LIBRTE_PMD_SZEDATA2_AS default value: 0

    This option defines type of firmware address space and must be set according to the used card and mode. Currently supported values are:

    • 0 - for cards (modes):

      • NFB-100G1 (100G1)
    • 1 - for cards (modes):

      • NFB-100G2Q (100G1)
    • 2 - for cards (modes):

      • NFB-40G2 (40G2)
      • NFB-100G2C (100G2)
      • NFB-100G2Q (40G2)
    • 3 - for cards (modes):

      • NFB-40G2 (10G8)
      • NFB-100G2Q (10G8)
    • 4 - for cards (modes):

      • NFB-100G1 (10G10)
    • 5 - for experimental firmwares and future use

Using the SZEDATA2 PMD

From DPDK version 16.04 the type of SZEDATA2 PMD is changed to PMD_PDEV. SZEDATA2 device is automatically recognized during EAL initialization. No special command line options are needed.

Kernel modules have to be loaded before running the DPDK application.

Example of usage

Read packets from 0. and 1. receive channel and write them to 0. and 1. transmit channel:

$RTE_TARGET/app/testpmd -l 0-3 -n 2 \
-- --port-topology=chained --rxq=2 --txq=2 --nb-cores=2 -i -a

Example output:

[...]
EAL: PCI device 0000:06:00.0 on NUMA socket -1
EAL:   probe driver: 1b26:c1c1 rte_szedata2_pmd
PMD: Initializing szedata2 device (0000:06:00.0)
PMD: SZEDATA2 path: /dev/szedataII0
PMD: Available DMA channels RX: 8 TX: 8
PMD: resource0 phys_addr = 0xe8000000 len = 134217728 virt addr = 7f48f8000000
PMD: szedata2 device (0000:06:00.0) successfully initialized
Interactive-mode selected
Auto-start selected
Configuring Port 0 (socket 0)
Port 0: 00:11:17:00:00:00
Checking link statuses...
Port 0 Link Up - speed 10000 Mbps - full-duplex
Done
Start automatic packet forwarding
  io packet forwarding - CRC stripping disabled - packets/burst=32
  nb forwarding cores=2 - nb forwarding ports=1
  RX queues=2 - RX desc=128 - RX free threshold=0
  RX threshold registers: pthresh=0 hthresh=0 wthresh=0
  TX queues=2 - TX desc=512 - TX free threshold=0
  TX threshold registers: pthresh=0 hthresh=0 wthresh=0
  TX RS bit threshold=0 - TXQ flags=0x0
testpmd>

Tun/Tap Poll Mode Driver

The rte_eth_tap.c PMD creates a device using TUN/TAP interfaces on the local host. The PMD allows for DPDK and the host to communicate using a raw device interface on the host and in the DPDK application.

The device created is a TAP device, which sends/receives packet in a raw format with a L2 header. The usage for a TAP PMD is for connectivity to the local host using a TAP interface. When the TAP PMD is initialized it will create a number of tap devices in the host accessed via ifconfig -a or ip command. The commands can be used to assign and query the virtual like device.

These TAP interfaces can be used with Wireshark or tcpdump or Pktgen-DPDK along with being able to be used as a network connection to the DPDK application. The method enable one or more interfaces is to use the --vdev=net_tap0 option on the DPDK application command line. Each --vdev=net_tap1 option given will create an interface named dtap0, dtap1, and so on.

The interface name can be changed by adding the iface=foo0, for example:

--vdev=net_tap0,iface=foo0 --vdev=net_tap1,iface=foo1, ...

Also the speed of the interface can be changed from 10G to whatever number needed, but the interface does not enforce that speed, for example:

--vdev=net_tap0,iface=foo0,speed=25000

Normally the PMD will generate a random MAC address, but when testing or with a static configuration the developer may need a fixed MAC address style. Using the option mac=fixed you can create a fixed known MAC address:

--vdev=net_tap0,mac=fixed

The MAC address will have a fixed value with the last octet incrementing by one for each interface string containing mac=fixed. The MAC address is formatted as 00:’d’:’t’:’a’:’p’:[00-FF]. Convert the characters to hex and you get the actual MAC address: 00:64:74:61:70:[00-FF].

It is possible to specify a remote netdevice to capture packets from by adding remote=foo1, for example:

--vdev=net_tap,iface=tap0,remote=foo1

If a remote is set, the tap MAC address will be set to match the remote one just after netdevice creation. Using TC rules, traffic from the remote netdevice will be redirected to the tap. If the tap is in promiscuous mode, then all packets will be redirected. In allmulti mode, all multicast packets will be redirected.

Using the remote feature is especially useful for capturing traffic from a netdevice that has no support in the DPDK. It is possible to add explicit rte_flow rules on the tap PMD to capture specific traffic (see next section for examples).

After the DPDK application is started you can send and receive packets on the interface using the standard rx_burst/tx_burst APIs in DPDK. From the host point of view you can use any host tool like tcpdump, Wireshark, ping, Pktgen and others to communicate with the DPDK application. The DPDK application may not understand network protocols like IPv4/6, UDP or TCP unless the application has been written to understand these protocols.

If you need the interface as a real network interface meaning running and has a valid IP address then you can do this with the following commands:

sudo ip link set dtap0 up; sudo ip addr add 192.168.0.250/24 dev dtap0
sudo ip link set dtap1 up; sudo ip addr add 192.168.1.250/24 dev dtap1

Please change the IP addresses as you see fit.

If routing is enabled on the host you can also communicate with the DPDK App over the internet via a standard socket layer application as long as you account for the protocol handing in the application.

If you have a Network Stack in your DPDK application or something like it you can utilize that stack to handle the network protocols. Plus you would be able to address the interface using an IP address assigned to the internal interface.

Flow API support

The tap PMD supports major flow API pattern items and actions, when running on linux kernels above 4.2 (“Flower” classifier required). The kernel support can be checked with this command:

zcat /proc/config.gz | ( grep 'CLS_FLOWER=' || echo 'not supported' ) |
tee -a /dev/stderr | grep -q '=m' &&
lsmod | ( grep cls_flower || echo 'try modprobe cls_flower' )

Supported items:

  • eth: src and dst (with variable masks), and eth_type (0xffff mask).
  • vlan: vid, pcp, tpid, but not eid. (requires kernel 4.9)
  • ipv4/6: src and dst (with variable masks), and ip_proto (0xffff mask).
  • udp/tcp: src and dst port (0xffff) mask.

Supported actions:

  • DROP
  • QUEUE
  • PASSTHRU

It is generally not possible to provide a “last” item. However, if the “last” item, once masked, is identical to the masked spec, then it is supported.

Only IPv4/6 and MAC addresses can use a variable mask. All other items need a full mask (exact match).

As rules are translated to TC, it is possible to show them with something like:

tc -s filter show dev tap1 parent 1:
Examples of testpmd flow rules

Drop packets for destination IP 192.168.0.1:

testpmd> flow create 0 priority 1 ingress pattern eth / ipv4 dst is 1.1.1.1 \
         / end actions drop / end

Ensure packets from a given MAC address are received on a queue 2:

testpmd> flow create 0 priority 2 ingress pattern eth src is 06:05:04:03:02:01 \
         / end actions queue index 2 / end

Drop UDP packets in vlan 3:

testpmd> flow create 0 priority 3 ingress pattern eth / vlan vid is 3 / \
         ipv4 proto is 17 / end actions drop / end

Example

The following is a simple example of using the TUN/TAP PMD with the Pktgen packet generator. It requires that the socat utility is installed on the test system.

Build DPDK, then pull down Pktgen and build pktgen using the DPDK SDK/Target used to build the dpdk you pulled down.

Run pktgen from the pktgen directory in a terminal with a commandline like the following:

sudo ./app/app/x86_64-native-linuxapp-gcc/app/pktgen -l 1-5 -n 4        \
 --proc-type auto --log-level 8 --socket-mem 512,512 --file-prefix pg   \
 --vdev=net_tap0 --vdev=net_tap1 -b 05:00.0 -b 05:00.1                  \
 -b 04:00.0 -b 04:00.1 -b 04:00.2 -b 04:00.3                            \
 -b 81:00.0 -b 81:00.1 -b 81:00.2 -b 81:00.3                            \
 -b 82:00.0 -b 83:00.0 -- -T -P -m [2:3].0 -m [4:5].1                   \
 -f themes/black-yellow.theme

Verify with ifconfig -a command in a different xterm window, should have a dtap0 and dtap1 interfaces created.

Next set the links for the two interfaces to up via the commands below:

sudo ip link set dtap0 up; sudo ip addr add 192.168.0.250/24 dev dtap0
sudo ip link set dtap1 up; sudo ip addr add 192.168.1.250/24 dev dtap1

Then use socat to create a loopback for the two interfaces:

sudo socat interface:dtap0 interface:dtap1

Then on the Pktgen command line interface you can start sending packets using the commands start 0 and start 1 or you can start both at the same time with start all. The command str is an alias for start all and stp is an alias for stop all.

While running you should see the 64 byte counters increasing to verify the traffic is being looped back. You can use set all size XXX to change the size of the packets after you stop the traffic. Use pktgen help command to see a list of all commands. You can also use the -f option to load commands at startup in command line or Lua script in pktgen.

ThunderX NICVF Poll Mode Driver

The ThunderX NICVF PMD (librte_pmd_thunderx_nicvf) provides poll mode driver support for the inbuilt NIC found in the Cavium ThunderX SoC family as well as their virtual functions (VF) in SR-IOV context.

More information can be found at Cavium, Inc Official Website.

Features

Features of the ThunderX PMD are:

  • Multiple queues for TX and RX
  • Receive Side Scaling (RSS)
  • Packet type information
  • Checksum offload
  • Promiscuous mode
  • Multicast mode
  • Port hardware statistics
  • Jumbo frames
  • Link state information
  • Scattered and gather for TX and RX
  • VLAN stripping
  • SR-IOV VF
  • NUMA support
  • Multi queue set support (up to 96 queues (12 queue sets)) per port

Supported ThunderX SoCs

  • CN88xx
  • CN81xx
  • CN83xx

Prerequisites

Pre-Installation Configuration

Config File Options

The following options can be modified in the config file. Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_LIBRTE_THUNDERX_NICVF_PMD (default y)

    Toggle compilation of the librte_pmd_thunderx_nicvf driver.

  • CONFIG_RTE_LIBRTE_THUNDERX_NICVF_DEBUG_INIT (default n)

    Toggle display of initialization related messages.

  • CONFIG_RTE_LIBRTE_THUNDERX_NICVF_DEBUG_RX (default n)

    Toggle display of receive fast path run-time message

  • CONFIG_RTE_LIBRTE_THUNDERX_NICVF_DEBUG_TX (default n)

    Toggle display of transmit fast path run-time message

  • CONFIG_RTE_LIBRTE_THUNDERX_NICVF_DEBUG_DRIVER (default n)

    Toggle display of generic debugging messages

  • CONFIG_RTE_LIBRTE_THUNDERX_NICVF_DEBUG_MBOX (default n)

    Toggle display of PF mailbox related run-time check messages

Driver compilation and testing

Refer to the document compiling and testing a PMD for a NIC for details.

To compile the ThunderX NICVF PMD for Linux arm64 gcc, use arm64-thunderx-linuxapp-gcc as target.

Linux

SR-IOV: Prerequisites and sample Application Notes

Current ThunderX NIC PF/VF kernel modules maps each physical Ethernet port automatically to virtual function (VF) and presented them as PCIe-like SR-IOV device. This section provides instructions to configure SR-IOV with Linux OS.

  1. Verify PF devices capabilities using lspci:

    lspci -vvv
    

    Example output:

    0002:01:00.0 Ethernet controller: Cavium Networks Device a01e (rev 01)
    ...
    Capabilities: [100 v1] Alternative Routing-ID Interpretation (ARI)
    ...
    Capabilities: [180 v1] Single Root I/O Virtualization (SR-IOV)
    ...
    Kernel driver in use: thunder-nic
    ...
    

    Note

    Unless thunder-nic driver is in use make sure your kernel config includes CONFIG_THUNDER_NIC_PF setting.

  2. Verify VF devices capabilities and drivers using lspci:

    lspci -vvv
    

    Example output:

    0002:01:00.1 Ethernet controller: Cavium Networks Device 0011 (rev 01)
    ...
    Capabilities: [100 v1] Alternative Routing-ID Interpretation (ARI)
    ...
    Kernel driver in use: thunder-nicvf
    ...
    
    0002:01:00.2 Ethernet controller: Cavium Networks Device 0011 (rev 01)
    ...
    Capabilities: [100 v1] Alternative Routing-ID Interpretation (ARI)
    ...
    Kernel driver in use: thunder-nicvf
    ...
    

    Note

    Unless thunder-nicvf driver is in use make sure your kernel config includes CONFIG_THUNDER_NIC_VF setting.

  3. Pass VF device to VM context (PCIe Passthrough):

    The VF devices may be passed through to the guest VM using qemu or virt-manager or virsh etc.

    Example qemu guest launch command:

    sudo qemu-system-aarch64 -name vm1 \
    -machine virt,gic_version=3,accel=kvm,usb=off \
    -cpu host -m 4096 \
    -smp 4,sockets=1,cores=8,threads=1 \
    -nographic -nodefaults \
    -kernel <kernel image> \
    -append "root=/dev/vda console=ttyAMA0 rw hugepagesz=512M hugepages=3" \
    -device vfio-pci,host=0002:01:00.1 \
    -drive file=<rootfs.ext3>,if=none,id=disk1,format=raw  \
    -device virtio-blk-device,scsi=off,drive=disk1,id=virtio-disk1,bootindex=1 \
    -netdev tap,id=net0,ifname=tap0,script=/etc/qemu-ifup_thunder \
    -device virtio-net-device,netdev=net0 \
    -serial stdio \
    -mem-path /dev/huge
    
  4. Enable VFIO-NOIOMMU mode (optional):

    echo 1 > /sys/module/vfio/parameters/enable_unsafe_noiommu_mode
    

    Note

    VFIO-NOIOMMU is required only when running in VM context and should not be enabled otherwise.

  5. Running testpmd:

    Follow instructions available in the document compiling and testing a PMD for a NIC to run testpmd.

    Example output:

    ./arm64-thunderx-linuxapp-gcc/app/testpmd -l 0-3 -n 4 -w 0002:01:00.2 \
      -- -i --disable-hw-vlan-filter --disable-crc-strip --no-flush-rx \
      --port-topology=loop
    
    ...
    
    PMD: rte_nicvf_pmd_init(): librte_pmd_thunderx nicvf version 1.0
    
    ...
    EAL:   probe driver: 177d:11 rte_nicvf_pmd
    EAL:   using IOMMU type 1 (Type 1)
    EAL:   PCI memory mapped at 0x3ffade50000
    EAL: Trying to map BAR 4 that contains the MSI-X table.
         Trying offsets: 0x40000000000:0x0000, 0x10000:0x1f0000
    EAL:   PCI memory mapped at 0x3ffadc60000
    PMD: nicvf_eth_dev_init(): nicvf: device (177d:11) 2:1:0:2
    PMD: nicvf_eth_dev_init(): node=0 vf=1 mode=tns-bypass sqs=false
         loopback_supported=true
    PMD: nicvf_eth_dev_init(): Port 0 (177d:11) mac=a6:c6:d9:17:78:01
    Interactive-mode selected
    Configuring Port 0 (socket 0)
    ...
    
    PMD: nicvf_dev_configure(): Configured ethdev port0 hwcap=0x0
    Port 0: A6:C6:D9:17:78:01
    Checking link statuses...
    Port 0 Link Up - speed 10000 Mbps - full-duplex
    Done
    testpmd>
    
Multiple Queue Set per DPDK port configuration

There are two types of VFs:

  • Primary VF
  • Secondary VF

Each port consists of a primary VF and n secondary VF(s). Each VF provides 8 Tx/Rx queues to a port. When a given port is configured to use more than 8 queues, it requires one (or more) secondary VF. Each secondary VF adds 8 additional queues to the queue set.

During PMD driver initialization, the primary VF’s are enumerated by checking the specific flag (see sqs message in DPDK boot log - sqs indicates secondary queue set). They are at the beginning of VF list (the remain ones are secondary VF’s).

The primary VFs are used as master queue sets. Secondary VFs provide additional queue sets for primary ones. If a port is configured for more then 8 queues than it will request for additional queues from secondary VFs.

Secondary VFs cannot be shared between primary VFs.

Primary VFs are present on the beginning of the ‘Network devices using kernel driver’ list, secondary VFs are on the remaining on the remaining part of the list.

Note

The VNIC driver in the multiqueue setup works differently than other drivers like ixgbe. We need to bind separately each specific queue set device with the usertools/dpdk-devbind.py utility.

Note

Depending on the hardware used, the kernel driver sets a threshold vf_id. VFs that try to attached with an id below or equal to this boundary are considered primary VFs. VFs that try to attach with an id above this boundary are considered secondary VFs.

Example device binding

If a system has three interfaces, a total of 18 VF devices will be created on a non-NUMA machine.

Note

NUMA systems have 12 VFs per port and non-NUMA 6 VFs per port.

# usertools/dpdk-devbind.py --status

Network devices using DPDK-compatible driver
============================================
<none>

Network devices using kernel driver
===================================
0000:01:10.0 'Device a026' if= drv=thunder-BGX unused=vfio-pci,uio_pci_generic
0000:01:10.1 'Device a026' if= drv=thunder-BGX unused=vfio-pci,uio_pci_generic
0002:01:00.0 'Device a01e' if= drv=thunder-nic unused=vfio-pci,uio_pci_generic
0002:01:00.1 'Device 0011' if=eth0 drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:00.2 'Device 0011' if=eth1 drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:00.3 'Device 0011' if=eth2 drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:00.4 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:00.5 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:00.6 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:00.7 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:01.0 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:01.1 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:01.2 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:01.3 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:01.4 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:01.5 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:01.6 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:01.7 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:02.0 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:02.1 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic
0002:01:02.2 'Device 0011' if= drv=thunder-nicvf unused=vfio-pci,uio_pci_generic

Other network devices
=====================
0002:00:03.0 'Device a01f' unused=vfio-pci,uio_pci_generic

We want to bind two physical interfaces with 24 queues each device, we attach two primary VFs and four secondary queues. In our example we choose two 10G interfaces eth1 (0002:01:00.2) and eth2 (0002:01:00.3). We will choose four secondary queue sets from the ending of the list (0002:01:01.7-0002:01:02.2).

  1. Bind two primary VFs to the vfio-pci driver:

    usertools/dpdk-devbind.py -b vfio-pci 0002:01:00.2
    usertools/dpdk-devbind.py -b vfio-pci 0002:01:00.3
    
  2. Bind four primary VFs to the vfio-pci driver:

    usertools/dpdk-devbind.py -b vfio-pci 0002:01:01.7
    usertools/dpdk-devbind.py -b vfio-pci 0002:01:02.0
    usertools/dpdk-devbind.py -b vfio-pci 0002:01:02.1
    usertools/dpdk-devbind.py -b vfio-pci 0002:01:02.2
    

The nicvf thunderx driver will make use of attached secondary VFs automatically during the interface configuration stage.

Limitations

CRC striping

The ThunderX SoC family NICs strip the CRC for every packets coming into the host interface. So, CRC will be stripped even when the rxmode.hw_strip_crc member is set to 0 in struct rte_eth_conf.

Maximum packet length

The ThunderX SoC family NICs support a maximum of a 9K jumbo frame. The value is fixed and cannot be changed. So, even when the rxmode.max_rx_pkt_len member of struct rte_eth_conf is set to a value lower than 9200, frames up to 9200 bytes can still reach the host interface.

Maximum packet segments

The ThunderX SoC family NICs support up to 12 segments per packet when working in scatter/gather mode. So, setting MTU will result with EINVAL when the frame size does not fit in the maximum number of segments.

Poll Mode Driver for Emulated Virtio NIC

Virtio is a para-virtualization framework initiated by IBM, and supported by KVM hypervisor. In the Data Plane Development Kit (DPDK), we provide a virtio Poll Mode Driver (PMD) as a software solution, comparing to SRIOV hardware solution,

for fast guest VM to guest VM communication and guest VM to host communication.

Vhost is a kernel acceleration module for virtio qemu backend. The DPDK extends kni to support vhost raw socket interface, which enables vhost to directly read/ write packets from/to a physical port. With this enhancement, virtio could achieve quite promising performance.

For basic qemu-KVM installation and other Intel EM poll mode driver in guest VM, please refer to Chapter “Driver for VM Emulated Devices”.

In this chapter, we will demonstrate usage of virtio PMD driver with two backends, standard qemu vhost back end and vhost kni back end.

Virtio Implementation in DPDK

For details about the virtio spec, refer to Virtio PCI Card Specification written by Rusty Russell.

As a PMD, virtio provides packet reception and transmission callbacks virtio_recv_pkts and virtio_xmit_pkts.

In virtio_recv_pkts, index in range [vq->vq_used_cons_idx , vq->vq_ring.used->idx) in vring is available for virtio to burst out.

In virtio_xmit_pkts, same index range in vring is available for virtio to clean. Virtio will enqueue to be transmitted packets into vring, advance the vq->vq_ring.avail->idx, and then notify the host back end if necessary.

Features and Limitations of virtio PMD

In this release, the virtio PMD driver provides the basic functionality of packet reception and transmission.

  • It supports merge-able buffers per packet when receiving packets and scattered buffer per packet when transmitting packets. The packet size supported is from 64 to 1518.
  • It supports multicast packets and promiscuous mode.
  • The descriptor number for the Rx/Tx queue is hard-coded to be 256 by qemu 2.7 and below. If given a different descriptor number by the upper application, the virtio PMD generates a warning and fall back to the hard-coded value. Rx queue size can be configureable and up to 1024 since qemu 2.8 and above. Rx queue size is 256 by default. Tx queue size is still hard-coded to be 256.
  • Features of mac/vlan filter are supported, negotiation with vhost/backend are needed to support them. When backend can’t support vlan filter, virtio app on guest should disable vlan filter to make sure the virtio port is configured correctly. E.g. specify ‘–disable-hw-vlan’ in testpmd command line.
  • “RTE_PKTMBUF_HEADROOM” should be defined no less than “sizeof(struct virtio_net_hdr_mrg_rxbuf)”, which is 12 bytes when mergeable or “VIRTIO_F_VERSION_1” is set. no less than “sizeof(struct virtio_net_hdr)”, which is 10 bytes, when using non-mergeable.
  • Virtio does not support runtime configuration.
  • Virtio supports Link State interrupt.
  • Virtio supports Rx interrupt (so far, only support 1:1 mapping for queue/interrupt).
  • Virtio supports software vlan stripping and inserting.
  • Virtio supports using port IO to get PCI resource when uio/igb_uio module is not available.

Prerequisites

The following prerequisites apply:

  • In the BIOS, turn VT-x and VT-d on
  • Linux kernel with KVM module; vhost module loaded and ioeventfd supported. Qemu standard backend without vhost support isn’t tested, and probably isn’t supported.

Virtio with kni vhost Back End

This section demonstrates kni vhost back end example setup for Phy-VM Communication.

_images/host_vm_comms.png

Host2VM Communication Example Using kni vhost Back End

Host2VM communication example

  1. Load the kni kernel module:

    insmod rte_kni.ko
    

    Other basic DPDK preparations like hugepage enabling, uio port binding are not listed here. Please refer to the DPDK Getting Started Guide for detailed instructions.

  2. Launch the kni user application:

    examples/kni/build/app/kni -l 0-3 -n 4 -- -p 0x1 -P --config="(0,1,3)"
    

    This command generates one network device vEth0 for physical port. If specify more physical ports, the generated network device will be vEth1, vEth2, and so on.

    For each physical port, kni creates two user threads. One thread loops to fetch packets from the physical NIC port into the kni receive queue. The other user thread loops to send packets in the kni transmit queue.

    For each physical port, kni also creates a kernel thread that retrieves packets from the kni receive queue, place them onto kni’s raw socket’s queue and wake up the vhost kernel thread to exchange packets with the virtio virt queue.

    For more details about kni, please refer to Kernel NIC Interface.

  3. Enable the kni raw socket functionality for the specified physical NIC port, get the generated file descriptor and set it in the qemu command line parameter. Always remember to set ioeventfd_on and vhost_on.

    Example:

    echo 1 > /sys/class/net/vEth0/sock_en
    fd=`cat /sys/class/net/vEth0/sock_fd`
    exec qemu-system-x86_64 -enable-kvm -cpu host \
    -m 2048 -smp 4 -name dpdk-test1-vm1 \
    -drive file=/data/DPDKVMS/dpdk-vm.img \
    -netdev tap, fd=$fd,id=mynet_kni, script=no,vhost=on \
    -device virtio-net-pci,netdev=mynet_kni,bus=pci.0,addr=0x3,ioeventfd=on \
    -vnc:1 -daemonize
    

    In the above example, virtio port 0 in the guest VM will be associated with vEth0, which in turns corresponds to a physical port, which means received packets come from vEth0, and transmitted packets is sent to vEth0.

  4. In the guest, bind the virtio device to the uio_pci_generic kernel module and start the forwarding application. When the virtio port in guest bursts Rx, it is getting packets from the raw socket’s receive queue. When the virtio port bursts Tx, it is sending packet to the tx_q.

    modprobe uio
    echo 512 > /sys/devices/system/node/node0/hugepages/hugepages-2048kB/nr_hugepages
    modprobe uio_pci_generic
    python usertools/dpdk-devbind.py -b uio_pci_generic 00:03.0
    

    We use testpmd as the forwarding application in this example.

    _images/console.png

    Running testpmd

  5. Use IXIA packet generator to inject a packet stream into the KNI physical port.

    The packet reception and transmission flow path is:

    IXIA packet generator->82599 PF->KNI Rx queue->KNI raw socket queue->Guest VM virtio port 0 Rx burst->Guest VM virtio port 0 Tx burst-> KNI Tx queue ->82599 PF-> IXIA packet generator

Virtio with qemu virtio Back End

_images/host_vm_comms_qemu.png

Host2VM Communication Example Using qemu vhost Back End

qemu-system-x86_64 -enable-kvm -cpu host -m 2048 -smp 2 -mem-path /dev/
hugepages -mem-prealloc
-drive file=/data/DPDKVMS/dpdk-vm1
-netdev tap,id=vm1_p1,ifname=tap0,script=no,vhost=on
-device virtio-net-pci,netdev=vm1_p1,bus=pci.0,addr=0x3,ioeventfd=on
-device pci-assign,host=04:10.1 \

In this example, the packet reception flow path is:

IXIA packet generator->82599 PF->Linux Bridge->TAP0’s socket queue-> Guest VM virtio port 0 Rx burst-> Guest VM 82599 VF port1 Tx burst-> IXIA packet generator

The packet transmission flow is:

IXIA packet generator-> Guest VM 82599 VF port1 Rx burst-> Guest VM virtio port 0 Tx burst-> tap -> Linux Bridge->82599 PF-> IXIA packet generator

Virtio PMD Rx/Tx Callbacks

Virtio driver has 3 Rx callbacks and 2 Tx callbacks.

Rx callbacks:

  1. virtio_recv_pkts: Regular version without mergeable Rx buffer support.
  2. virtio_recv_mergeable_pkts: Regular version with mergeable Rx buffer support.
  3. virtio_recv_pkts_vec: Vector version without mergeable Rx buffer support, also fixes the available ring indexes and uses vector instructions to optimize performance.

Tx callbacks:

  1. virtio_xmit_pkts: Regular version.
  2. virtio_xmit_pkts_simple: Vector version fixes the available ring indexes to optimize performance.

By default, the non-vector callbacks are used:

  • For Rx: If mergeable Rx buffers is disabled then virtio_recv_pkts is used; otherwise virtio_recv_mergeable_pkts.
  • For Tx: virtio_xmit_pkts.

Vector callbacks will be used when:

  • txq_flags is set to VIRTIO_SIMPLE_FLAGS (0xF01), which implies:
    • Single segment is specified.
    • No offload support is needed.
  • Mergeable Rx buffers is disabled.

The corresponding callbacks are:

  • For Rx: virtio_recv_pkts_vec.
  • For Tx: virtio_xmit_pkts_simple.

Example of using the vector version of the virtio poll mode driver in testpmd:

testpmd -l 0-2 -n 4 -- -i --txqflags=0xF01 --rxq=1 --txq=1 --nb-cores=1

Interrupt mode

There are three kinds of interrupts from a virtio device over PCI bus: config interrupt, Rx interrupts, and Tx interrupts. Config interrupt is used for notification of device configuration changes, especially link status (lsc). Interrupt mode is translated into Rx interrupts in the context of DPDK.

Note

Virtio PMD already has support for receiving lsc from qemu when the link status changes, especially when vhost user disconnects. However, it fails to do that if the VM is created by qemu 2.6.2 or below, since the capability to detect vhost user disconnection is introduced in qemu 2.7.0.

Prerequisites for Rx interrupts

To support Rx interrupts, #. Check if guest kernel supports VFIO-NOIOMMU:

Linux started to support VFIO-NOIOMMU since 4.8.0. Make sure the guest kernel is compiled with:

CONFIG_VFIO_NOIOMMU=y
  1. Properly set msix vectors when starting VM:

    Enable multi-queue when starting VM, and specify msix vectors in qemu cmdline. (N+1) is the minimum, and (2N+2) is mostly recommended.

    $(QEMU) ... -device virtio-net-pci,mq=on,vectors=2N+2 ...
    
  2. In VM, insert vfio module in NOIOMMU mode:

    modprobe vfio enable_unsafe_noiommu_mode=1
    modprobe vfio-pci
    
  3. In VM, bind the virtio device with vfio-pci:

    python usertools/dpdk-devbind.py -b vfio-pci 00:03.0
    
Example

Here we use l3fwd-power as an example to show how to get started.

Example:

$ l3fwd-power -l 0-1 -- -p 1 -P --config="(0,0,1)" \
                                       --no-numa --parse-ptype

Poll Mode Driver that wraps vhost library

This PMD is a thin wrapper of the DPDK vhost library. The user can handle virtqueues as one of normal DPDK port.

Vhost Implementation in DPDK

Please refer to Chapter “Vhost Library” of DPDK Programmer’s Guide to know detail of vhost.

Features and Limitations of vhost PMD

Currently, the vhost PMD provides the basic functionality of packet reception, transmission and event handling.

  • It has multiple queues support.
  • It supports RTE_ETH_EVENT_INTR_LSC and RTE_ETH_EVENT_QUEUE_STATE events.
  • It supports Port Hotplug functionality.
  • Don’t need to stop RX/TX, when the user wants to stop a guest or a virtio-net driver on guest.

Vhost PMD arguments

The user can specify below arguments in –vdev option.

  1. iface:

    It is used to specify a path to connect to a QEMU virtio-net device.

  2. queues:

    It is used to specify the number of queues virtio-net device has. (Default: 1)

  3. iommu-support:

    It is used to enable iommu support in vhost library. (Default: 0 (disabled))

Vhost PMD event handling

This section describes how to handle vhost PMD events.

The user can register an event callback handler with rte_eth_dev_callback_register(). The registered callback handler will be invoked with one of below event types.

  1. RTE_ETH_EVENT_INTR_LSC:

    It means link status of the port was changed.

  2. RTE_ETH_EVENT_QUEUE_STATE:

    It means some of queue statuses were changed. Call rte_eth_vhost_get_queue_event() in the callback handler. Because changing multiple statuses may occur only one event, call the function repeatedly as long as it doesn’t return negative value.

Vhost PMD with testpmd application

This section demonstrates vhost PMD with testpmd DPDK sample application.

  1. Launch the testpmd with vhost PMD:

    ./testpmd -l 0-3 -n 4 --vdev 'net_vhost0,iface=/tmp/sock0,queues=1' -- -i
    

    Other basic DPDK preparations like hugepage enabling here. Please refer to the DPDK Getting Started Guide for detailed instructions.

  2. Launch the QEMU:

    qemu-system-x86_64 <snip>
                -chardev socket,id=chr0,path=/tmp/sock0 \
                -netdev vhost-user,id=net0,chardev=chr0,vhostforce,queues=1 \
                -device virtio-net-pci,netdev=net0
    

    This command attaches one virtio-net device to QEMU guest. After initialization processes between QEMU and DPDK vhost library are done, status of the port will be linked up.

Poll Mode Driver for Paravirtual VMXNET3 NIC

The VMXNET3 adapter is the next generation of a paravirtualized NIC, introduced by VMware* ESXi. It is designed for performance, offers all the features available in VMXNET2, and adds several new features such as, multi-queue support (also known as Receive Side Scaling, RSS), IPv6 offloads, and MSI/MSI-X interrupt delivery. One can use the same device in a DPDK application with VMXNET3 PMD introduced in DPDK API.

In this chapter, two setups with the use of the VMXNET3 PMD are demonstrated:

  1. Vmxnet3 with a native NIC connected to a vSwitch
  2. Vmxnet3 chaining VMs connected to a vSwitch

VMXNET3 Implementation in the DPDK

For details on the VMXNET3 device, refer to the VMXNET3 driver’s vmxnet3 directory and support manual from VMware*.

For performance details, refer to the following link from VMware:

http://www.vmware.com/pdf/vsp_4_vmxnet3_perf.pdf

As a PMD, the VMXNET3 driver provides the packet reception and transmission callbacks, vmxnet3_recv_pkts and vmxnet3_xmit_pkts.

The VMXNET3 PMD handles all the packet buffer memory allocation and resides in guest address space and it is solely responsible to free that memory when not needed. The packet buffers and features to be supported are made available to hypervisor via VMXNET3 PCI configuration space BARs. During RX/TX, the packet buffers are exchanged by their GPAs, and the hypervisor loads the buffers with packets in the RX case and sends packets to vSwitch in the TX case.

The VMXNET3 PMD is compiled with vmxnet3 device headers. The interface is similar to that of the other PMDs available in the DPDK API. The driver pre-allocates the packet buffers and loads the command ring descriptors in advance. The hypervisor fills those packet buffers on packet arrival and write completion ring descriptors, which are eventually pulled by the PMD. After reception, the DPDK application frees the descriptors and loads new packet buffers for the coming packets. The interrupts are disabled and there is no notification required. This keeps performance up on the RX side, even though the device provides a notification feature.

In the transmit routine, the DPDK application fills packet buffer pointers in the descriptors of the command ring and notifies the hypervisor. In response the hypervisor takes packets and passes them to the vSwitch, It writes into the completion descriptors ring. The rings are read by the PMD in the next transmit routine call and the buffers and descriptors are freed from memory.

Features and Limitations of VMXNET3 PMD

In release 1.6.0, the VMXNET3 PMD provides the basic functionality of packet reception and transmission. There are several options available for filtering packets at VMXNET3 device level including:

  1. MAC Address based filtering:
    • Unicast, Broadcast, All Multicast modes - SUPPORTED BY DEFAULT
    • Multicast with Multicast Filter table - NOT SUPPORTED
    • Promiscuous mode - SUPPORTED
    • RSS based load balancing between queues - SUPPORTED
  2. VLAN filtering:
    • VLAN tag based filtering without load balancing - SUPPORTED

Note

  • Release 1.6.0 does not support separate headers and body receive cmd_ring and hence, multiple segment buffers are not supported. Only cmd_ring_0 is used for packet buffers, one for each descriptor.
  • Receive and transmit of scattered packets is not supported.
  • Multicast with Multicast Filter table is not supported.

Prerequisites

The following prerequisites apply:

  • Before starting a VM, a VMXNET3 interface to a VM through VMware vSphere Client must be assigned. This is shown in the figure below.
_images/vmxnet3_int.png

Assigning a VMXNET3 interface to a VM using VMware vSphere Client

Note

Depending on the Virtual Machine type, the VMware vSphere Client shows Ethernet adaptors while adding an Ethernet device. Ensure that the VM type used offers a VMXNET3 device. Refer to the VMware documentation for a listed of VMs.

Note

Follow the DPDK Getting Started Guide to setup the basic DPDK environment.

Note

Follow the DPDK Sample Application’s User Guide, L2 Forwarding/L3 Forwarding and TestPMD for instructions on how to run a DPDK application using an assigned VMXNET3 device.

VMXNET3 with a Native NIC Connected to a vSwitch

This section describes an example setup for Phy-vSwitch-VM-Phy communication.

_images/vswitch_vm.png

VMXNET3 with a Native NIC Connected to a vSwitch

Note

Other instructions on preparing to use DPDK such as, hugepage enabling, uio port binding are not listed here. Please refer to DPDK Getting Started Guide and DPDK Sample Application’s User Guide for detailed instructions.

The packet reception and transmission flow path is:

Packet generator -> 82576
                 -> VMware ESXi vSwitch
                 -> VMXNET3 device
                 -> Guest VM VMXNET3 port 0 rx burst
                 -> Guest VM 82599 VF port 0 tx burst
                 -> 82599 VF
                 -> Packet generator

VMXNET3 Chaining VMs Connected to a vSwitch

The following figure shows an example VM-to-VM communication over a Phy-VM-vSwitch-VM-Phy communication channel.

_images/vm_vm_comms.png

VMXNET3 Chaining VMs Connected to a vSwitch

Note

When using the L2 Forwarding or L3 Forwarding applications, a destination MAC address needs to be written in packets to hit the other VM’s VMXNET3 interface.

In this example, the packet flow path is:

Packet generator -> 82599 VF
                 -> Guest VM 82599 port 0 rx burst
                 -> Guest VM VMXNET3 port 1 tx burst
                 -> VMXNET3 device
                 -> VMware ESXi vSwitch
                 -> VMXNET3 device
                 -> Guest VM VMXNET3 port 0 rx burst
                 -> Guest VM 82599 VF port 1 tx burst
                 -> 82599 VF
                 -> Packet generator

Libpcap and Ring Based Poll Mode Drivers

In addition to Poll Mode Drivers (PMDs) for physical and virtual hardware, the DPDK also includes two pure-software PMDs. These two drivers are:

  • A libpcap -based PMD (librte_pmd_pcap) that reads and writes packets using libpcap, - both from files on disk, as well as from physical NIC devices using standard Linux kernel drivers.
  • A ring-based PMD (librte_pmd_ring) that allows a set of software FIFOs (that is, rte_ring) to be accessed using the PMD APIs, as though they were physical NICs.

Note

The libpcap -based PMD is disabled by default in the build configuration files, owing to an external dependency on the libpcap development files which must be installed on the board. Once the libpcap development files are installed, the library can be enabled by setting CONFIG_RTE_LIBRTE_PMD_PCAP=y and recompiling the DPDK.

Using the Drivers from the EAL Command Line

For ease of use, the DPDK EAL also has been extended to allow pseudo-Ethernet devices, using one or more of these drivers, to be created at application startup time during EAL initialization.

To do so, the –vdev= parameter must be passed to the EAL. This takes take options to allow ring and pcap-based Ethernet to be allocated and used transparently by the application. This can be used, for example, for testing on a virtual machine where there are no Ethernet ports.

Libpcap-based PMD

Pcap-based devices can be created using the virtual device –vdev option. The device name must start with the net_pcap prefix followed by numbers or letters. The name is unique for each device. Each device can have multiple stream options and multiple devices can be used. Multiple device definitions can be arranged using multiple –vdev. Device name and stream options must be separated by commas as shown below:

$RTE_TARGET/app/testpmd -l 0-3 -n 4 \
    --vdev 'net_pcap0,stream_opt0=..,stream_opt1=..' \
    --vdev='net_pcap1,stream_opt0=..'
Device Streams

Multiple ways of stream definitions can be assessed and combined as long as the following two rules are respected:

  • A device is provided with two different streams - reception and transmission.
  • A device is provided with one network interface name used for reading and writing packets.

The different stream types are:

  • rx_pcap: Defines a reception stream based on a pcap file. The driver reads each packet within the given pcap file as if it was receiving it from the wire. The value is a path to a valid pcap file.

    rx_pcap=/path/to/file.pcap

  • tx_pcap: Defines a transmission stream based on a pcap file. The driver writes each received packet to the given pcap file. The value is a path to a pcap file. The file is overwritten if it already exists and it is created if it does not.

    tx_pcap=/path/to/file.pcap

  • rx_iface: Defines a reception stream based on a network interface name. The driver reads packets coming from the given interface using the Linux kernel driver for that interface. The value is an interface name.

    rx_iface=eth0

  • tx_iface: Defines a transmission stream based on a network interface name. The driver sends packets to the given interface using the Linux kernel driver for that interface. The value is an interface name.

    tx_iface=eth0

  • iface: Defines a device mapping a network interface. The driver both reads and writes packets from and to the given interface. The value is an interface name.

    iface=eth0

Examples of Usage

Read packets from one pcap file and write them to another:

$RTE_TARGET/app/testpmd -l 0-3 -n 4 \
    --vdev 'net_pcap0,rx_pcap=file_rx.pcap,tx_pcap=file_tx.pcap' \
    -- --port-topology=chained

Read packets from a network interface and write them to a pcap file:

$RTE_TARGET/app/testpmd -l 0-3 -n 4 \
    --vdev 'net_pcap0,rx_iface=eth0,tx_pcap=file_tx.pcap' \
    -- --port-topology=chained

Read packets from a pcap file and write them to a network interface:

$RTE_TARGET/app/testpmd -l 0-3 -n 4 \
    --vdev 'net_pcap0,rx_pcap=file_rx.pcap,tx_iface=eth1' \
    -- --port-topology=chained

Forward packets through two network interfaces:

$RTE_TARGET/app/testpmd -l 0-3 -n 4 \
    --vdev 'net_pcap0,iface=eth0' --vdev='net_pcap1;iface=eth1'
Using libpcap-based PMD with the testpmd Application

One of the first things that testpmd does before starting to forward packets is to flush the RX streams by reading the first 512 packets on every RX stream and discarding them. When using a libpcap-based PMD this behavior can be turned off using the following command line option:

--no-flush-rx

It is also available in the runtime command line:

set flush_rx on/off

It is useful for the case where the rx_pcap is being used and no packets are meant to be discarded. Otherwise, the first 512 packets from the input pcap file will be discarded by the RX flushing operation.

$RTE_TARGET/app/testpmd -l 0-3 -n 4 \
    --vdev 'net_pcap0,rx_pcap=file_rx.pcap,tx_pcap=file_tx.pcap' \
    -- --port-topology=chained --no-flush-rx
Rings-based PMD

To run a DPDK application on a machine without any Ethernet devices, a pair of ring-based rte_ethdevs can be used as below. The device names passed to the –vdev option must start with net_ring and take no additional parameters. Multiple devices may be specified, separated by commas.

./testpmd -l 1-3 -n 4 --vdev=net_ring0 --vdev=net_ring1 -- -i
EAL: Detected lcore 1 as core 1 on socket 0
...

Interactive-mode selected
Configuring Port 0 (socket 0)
Configuring Port 1 (socket 0)
Checking link statuses...
Port 0 Link Up - speed 10000 Mbps - full-duplex
Port 1 Link Up - speed 10000 Mbps - full-duplex
Done

testpmd> start tx_first
io packet forwarding - CRC stripping disabled - packets/burst=16
nb forwarding cores=1 - nb forwarding ports=2
RX queues=1 - RX desc=128 - RX free threshold=0
RX threshold registers: pthresh=8 hthresh=8 wthresh=4
TX queues=1 - TX desc=512 - TX free threshold=0
TX threshold registers: pthresh=36 hthresh=0 wthresh=0
TX RS bit threshold=0 - TXQ flags=0x0

testpmd> stop
Telling cores to stop...
Waiting for lcores to finish...
_images/forward_stats.png
+++++++++++++++ Accumulated forward statistics for allports++++++++++
RX-packets: 462384736  RX-dropped: 0 RX-total: 462384736
TX-packets: 462384768  TX-dropped: 0 TX-total: 462384768
+++++++++++++++++++++++++++++++++++++++++++++++++++++

Done.
Using the Poll Mode Driver from an Application

Both drivers can provide similar APIs to allow the user to create a PMD, that is, rte_ethdev structure, instances at run-time in the end-application, for example, using rte_eth_from_rings() or rte_eth_from_pcaps() APIs. For the rings-based PMD, this functionality could be used, for example, to allow data exchange between cores using rings to be done in exactly the same way as sending or receiving packets from an Ethernet device. For the libpcap-based PMD, it allows an application to open one or more pcap files and use these as a source of packet input to the application.

Usage Examples

To create two pseudo-Ethernet ports where all traffic sent to a port is looped back for reception on the same port (error handling omitted for clarity):

#define RING_SIZE 256
#define NUM_RINGS 2
#define SOCKET0 0

struct rte_ring *ring[NUM_RINGS];
int port0, port1;

ring[0] = rte_ring_create("R0", RING_SIZE, SOCKET0, RING_F_SP_ENQ|RING_F_SC_DEQ);
ring[1] = rte_ring_create("R1", RING_SIZE, SOCKET0, RING_F_SP_ENQ|RING_F_SC_DEQ);

/* create two ethdev's */

port0 = rte_eth_from_rings("net_ring0", ring, NUM_RINGS, ring, NUM_RINGS, SOCKET0);
port1 = rte_eth_from_rings("net_ring1", ring, NUM_RINGS, ring, NUM_RINGS, SOCKET0);

To create two pseudo-Ethernet ports where the traffic is switched between them, that is, traffic sent to port 0 is read back from port 1 and vice-versa, the final two lines could be changed as below:

port0 = rte_eth_from_rings("net_ring0", &ring[0], 1, &ring[1], 1, SOCKET0);
port1 = rte_eth_from_rings("net_ring1", &ring[1], 1, &ring[0], 1, SOCKET0);

This type of configuration could be useful in a pipeline model, for example, where one may want to have inter-core communication using pseudo Ethernet devices rather than raw rings, for reasons of API consistency.

Enqueuing and dequeuing items from an rte_ring using the rings-based PMD may be slower than using the native rings API. This is because DPDK Ethernet drivers make use of function pointers to call the appropriate enqueue or dequeue functions, while the rte_ring specific functions are direct function calls in the code and are often inlined by the compiler.

Once an ethdev has been created, for either a ring or a pcap-based PMD, it should be configured and started in the same way as a regular Ethernet device, that is, by calling rte_eth_dev_configure() to set the number of receive and transmit queues, then calling rte_eth_rx_queue_setup() / tx_queue_setup() for each of those queues and finally calling rte_eth_dev_start() to allow transmission and reception of packets to begin.

Fail-safe poll mode driver library

The Fail-safe poll mode driver library (librte_pmd_failsafe) is a virtual device that allows using any device supporting hotplug (sudden device removal and plugging on its bus), without modifying other components relying on such device (application, other PMDs).

Additionally to the Seamless Hotplug feature, the Fail-safe PMD offers the ability to redirect operations to secondary devices when the primary has been removed from the system.

Note

The library is enabled by default. You can enable it or disable it manually by setting the CONFIG_RTE_LIBRTE_PMD_FAILSAFE configuration option.

Features

The Fail-safe PMD only supports a limited set of features. If you plan to use a device underneath the Fail-safe PMD with a specific feature, this feature must be supported by the Fail-safe PMD to avoid throwing any error.

A notable exception is the device removal feature. The fail-safe PMD being a virtual device, it cannot currently be removed in the sense of a specific bus hotplug, like for PCI for example. It will however enable this feature for its sub-device automatically, detecting those that are capable and register the relevant callback for such event.

Check the feature matrix for the complete set of supported features.

Compilation option

This option can be modified in the $RTE_TARGET/build/.config file.

  • CONFIG_RTE_LIBRTE_PMD_FAILSAFE (default y)

    Toggle compiling librte_pmd_failsafe.

Using the Fail-safe PMD from the EAL command line

The Fail-safe PMD can be used like most other DPDK virtual devices, by passing a --vdev parameter to the EAL when starting the application. The device name must start with the net_failsafe prefix, followed by numbers or letters. This name must be unique for each device. Each fail-safe instance must have at least one sub-device, up to RTE_MAX_ETHPORTS-1.

A sub-device can be any legal DPDK device, including possibly another fail-safe instance.

Fail-safe command line parameters
  • dev(<iface>) parameter

    This parameter allows the user to define a sub-device. The <iface> part of this parameter must be a valid device definition. It could be the argument provided to any -w device specification or the argument that would be given to a --vdev parameter (including a fail-safe). Enclosing the device definition within parenthesis here allows using additional sub-device parameters if need be. They will be passed on to the sub-device.

  • exec(<shell command>) parameter

    This parameter allows the user to provide a command to the fail-safe PMD to execute and define a sub-device. It is done within a regular shell context. The first line of its output is read by the fail-safe PMD and otherwise interpreted as if passed by the regular dev parameter. Any other line is discarded. If the command fail or output an incorrect string, the sub-device is not initialized. All commas within the shell command are replaced by spaces before executing the command. This helps using scripts to specify devices.

  • mac parameter [MAC address]

    This parameter allows the user to set a default MAC address to the fail-safe and all of its sub-devices. If no default mac address is provided, the fail-safe PMD will read the MAC address of the first of its sub-device to be successfully probed and use it as its default MAC address, trying to set it to all of its other sub-devices. If no sub-device was successfully probed at initialization, then a random MAC address is generated, that will be subsequently applied to all sub-device once they are probed.

  • hotplug_poll parameter [UINT64] (default 2000)

    This parameter allows the user to configure the amount of time in milliseconds between two slave upkeep round.

Usage example

This section shows some example of using testpmd with a fail-safe PMD.

  1. To build a PMD and configure DPDK, refer to the document compiling and testing a PMD for a NIC.

  2. Start testpmd. The slave device should be blacklisted from normal EAL operations to avoid probing it twice when in PCI blacklist mode.

    $RTE_TARGET/build/app/testpmd -c 0xff -n 4 \
       --vdev 'net_failsafe0,mac=de:ad:be:ef:01:02,dev(84:00.0),dev(net_ring0)' \
       -b 84:00.0 -b 00:04.0 -- -i
    

    If the slave device being used is not blacklisted, it will be probed by the EAL first. When the fail-safe then tries to initialize it the probe operation fails.

    Note that PCI blacklist mode is the default PCI operating mode.

  3. Alternatively, it can be used alongside any other device in whitelist mode.

    $RTE_TARGET/build/app/testpmd -c 0xff -n 4 \
       --vdev 'net_failsafe0,mac=de:ad:be:ef:01:02,dev(84:00.0),dev(net_ring0)' \
       -w 81:00.0 -- -i
    
  4. Start testpmd using a flexible device definition

    $RTE_TARGET/build/app/testpmd -c 0xff -n 4 --no-pci \
       --vdev='net_failsafe0,exec(echo 84:00.0)' -- -i
    

Using the Fail-safe PMD from an application

This driver strives to be as seamless as possible to existing applications, in order to propose the hotplug functionality in the easiest way possible.

Care must be taken, however, to respect the ether API concerning device access, and in particular, using the RTE_ETH_FOREACH_DEV macro to iterate over ethernet devices, instead of directly accessing them or by writing one’s own device iterator.

Plug-in feature

A sub-device can be defined without existing on the system when the fail-safe PMD is initialized. Upon probing this device, the fail-safe PMD will detect its absence and postpone its use. It will then register for a periodic check on any missing sub-device.

During this time, the fail-safe PMD can be used normally, configured and told to emit and receive packets. It will store any applied configuration, and try to apply it upon the probing of its missing sub-device. After this configuration pass, the new sub-device will be synchronized with other sub-devices, i.e. be started if the fail-safe PMD has been started by the user before.

Plug-out feature

A sub-device supporting the device removal event can be removed from its bus at any time. The fail-safe PMD will register a callback for such event and react accordingly. It will try to safely stop, close and uninit the sub-device having emitted this event, allowing it to free its eventual resources.

Fail-safe glossary

Fallback device : Secondary device
The fail-safe will fail-over onto this device when the preferred device is absent.
Preferred device : Primary device
The first declared sub-device in the fail-safe parameters. When this device is plugged, it is always used as emitting device. It is the main sub-device and is used as target for configuration operations if there is any ambiguity.
Upkeep round
Periodical process when slaves are serviced. Each devices having a state different to that of the fail-safe device itself, is synchronized with it. Additionally, each slave having the remove flag set are cleaned-up.
Slave
In the context of the fail-safe PMD, synonymous to sub-device.
Sub-device
A device being utilized by the fail-safe PMD. This is another PMD running underneath the fail-safe PMD. Any sub-device can disappear at any time. The fail-safe will ensure that the device removal happens gracefully.

Figures

Fig. 17.15 Virtualization for a Single Port NIC in SR-IOV Mode

Fig. 17.16 Performance Benchmark Setup

Fig. 17.17 Fast Host-based Packet Processing

Fig. 17.18 Inter-VM Communication

Fig. 30.2 Host2VM Communication Example Using kni vhost Back End

Fig. 30.4 Host2VM Communication Example Using qemu vhost Back End

Fig. 32.1 Assigning a VMXNET3 interface to a VM using VMware vSphere Client

Fig. 32.2 VMXNET3 with a Native NIC Connected to a vSwitch

Fig. 32.3 VMXNET3 Chaining VMs Connected to a vSwitch

Crypto Device Drivers

Crypto Device Supported Functionality Matrices

Supported Feature Flags

Features availability in crypto drivers
Feature a e s n i _ g c m a e s n i _ m b a r m v 8 d p a a 2 _ s e c d p a a _ s e c k a s u m i m r v l n u l l o p e n s s l q a t s n o w 3 g z u c
Symmetric crypto Y Y Y Y Y Y Y Y Y Y Y Y
Asymmetric crypto                        
Sym operation chaining Y Y Y Y Y Y Y Y Y Y Y Y
HW Accelerated       Y Y         Y    
Protocol offload       Y                
CPU SSE Y Y                    
CPU AVX Y Y                    
CPU AVX2 Y Y                    
CPU AVX512   Y                    
CPU AESNI Y Y                    
CPU NEON     Y                  
CPU ARM CE     Y                  

Supported Cipher Algorithms

Cipher algorithms in crypto drivers
Cipher algorithm a e s n i _ g c m a e s n i _ m b a r m v 8 d p a a 2 _ s e c d p a a _ s e c k a s u m i m r v l n u l l o p e n s s l q a t s n o w 3 g z u c
NULL               Y   Y    
AES CBC (128)   Y Y Y Y   Y   Y Y    
AES CBC (192)   Y   Y Y   Y   Y Y    
AES CBC (256)   Y   Y Y   Y   Y Y    
AES CTR (128)   Y   Y Y   Y   Y Y    
AES CTR (192)   Y   Y Y   Y   Y Y    
AES CTR (256)   Y   Y Y   Y   Y Y    
AES DOCSIS BPI   Y               Y    
3DES CBC       Y Y   Y   Y Y    
3DES CTR             Y   Y Y    
DES CBC   Y               Y    
DES DOCSIS BPI   Y             Y Y    
SNOW3G UEA2                   Y Y  
KASUMI F8           Y       Y    
ZUC EEA3                   Y   Y

Supported Authentication Algorithms

Authentication algorithms in crypto drivers
Authentication algorithm a e s n i _ g c m a e s n i _ m b a r m v 8 d p a a 2 _ s e c d p a a _ s e c k a s u m i m r v l n u l l o p e n s s l q a t s n o w 3 g z u c
NULL               Y   Y    
MD5             Y   Y      
MD5 HMAC   Y   Y Y   Y   Y Y    
SHA1             Y   Y      
SHA1 HMAC   Y Y Y Y   Y   Y Y    
SHA224                 Y      
SHA224 HMAC   Y   Y Y       Y Y    
SHA256             Y   Y      
SHA256 HMAC   Y Y Y Y   Y   Y Y    
SHA384             Y   Y      
SHA384 HMAC   Y   Y Y   Y   Y Y    
SHA512             Y   Y      
SHA512 HMAC   Y   Y Y   Y   Y Y    
AES XCBC MAC   Y               Y    
AES GMAC Y           Y   Y Y    
SNOW3G UIA2                   Y Y  
KASUMI F9           Y       Y    
ZUC EIA3                   Y   Y

Supported AEAD Algorithms

AEAD algorithms in crypto drivers
AEAD algorithm a e s n i _ g c m a e s n i _ m b a r m v 8 d p a a 2 _ s e c d p a a _ s e c k a s u m i m r v l n u l l o p e n s s l q a t s n o w 3 g z u c
AES GCM (128) Y     Y Y   Y   Y Y    
AES GCM (192) Y     Y Y       Y Y    
AES GCM (256) Y     Y Y       Y Y    
AES CCM (128)                 Y      
AES CCM (192)                 Y      
AES CCM (256)                 Y      

AESN-NI Multi Buffer Crypto Poll Mode Driver

The AESNI MB PMD (librte_pmd_aesni_mb) provides poll mode crypto driver support for utilizing Intel multi buffer library, see the white paper Fast Multi-buffer IPsec Implementations on Intel® Architecture Processors.

The AES-NI MB PMD has current only been tested on Fedora 21 64-bit with gcc.

Features

AESNI MB PMD has support for:

Cipher algorithms:

  • RTE_CRYPTO_CIPHER_AES128_CBC
  • RTE_CRYPTO_CIPHER_AES192_CBC
  • RTE_CRYPTO_CIPHER_AES256_CBC
  • RTE_CRYPTO_CIPHER_AES128_CTR
  • RTE_CRYPTO_CIPHER_AES192_CTR
  • RTE_CRYPTO_CIPHER_AES256_CTR
  • RTE_CRYPTO_CIPHER_AES_DOCSISBPI
  • RTE_CRYPTO_CIPHER_DES_CBC
  • RTE_CRYPTO_CIPHER_DES_DOCSISBPI

Hash algorithms:

  • RTE_CRYPTO_HASH_MD5_HMAC
  • RTE_CRYPTO_HASH_SHA1_HMAC
  • RTE_CRYPTO_HASH_SHA224_HMAC
  • RTE_CRYPTO_HASH_SHA256_HMAC
  • RTE_CRYPTO_HASH_SHA384_HMAC
  • RTE_CRYPTO_HASH_SHA512_HMAC
  • RTE_CRYPTO_HASH_AES_XCBC_HMAC

Limitations

  • Chained mbufs are not supported.
  • Only in-place is currently supported (destination address is the same as source address).

Installation

To build DPDK with the AESNI_MB_PMD the user is required to download the multi-buffer library from here and compile it on their user system before building DPDK. The latest version of the library supported by this PMD is v0.47, which can be downloaded from https://github.com/01org/intel-ipsec-mb/archive/v0.47.zip.

make

As a reference, the following table shows a mapping between the past DPDK versions and the Multi-Buffer library version supported by them:

DPDK and Multi-Buffer library version compatibility
DPDK version Multi-buffer library version
2.2 - 16.11 0.43 - 0.44
17.02 0.44
17.05 - 17.08 0.45 - 0.47
17.11+ 0.47

Initialization

In order to enable this virtual crypto PMD, user must:

  • Export the environmental variable AESNI_MULTI_BUFFER_LIB_PATH with the path where the library was extracted.
  • Build the multi buffer library (explained in Installation section).
  • Set CONFIG_RTE_LIBRTE_PMD_AESNI_MB=y in config/common_base.

To use the PMD in an application, user must:

  • Call rte_vdev_init(“crypto_aesni_mb”) within the application.
  • Use –vdev=”crypto_aesni_mb” in the EAL options, which will call rte_vdev_init() internally.

The following parameters (all optional) can be provided in the previous two calls:

  • socket_id: Specify the socket where the memory for the device is going to be allocated (by default, socket_id will be the socket where the core that is creating the PMD is running on).
  • max_nb_queue_pairs: Specify the maximum number of queue pairs in the device (8 by default).
  • max_nb_sessions: Specify the maximum number of sessions that can be created (2048 by default).

Example:

./l2fwd-crypto -l 1 -n 4 --vdev="crypto_aesni_mb,socket_id=0,max_nb_sessions=128" \
-- -p 1 --cdev SW --chain CIPHER_HASH --cipher_algo "aes-cbc" --auth_algo "sha1-hmac"

Extra notes

For AES Counter mode (AES-CTR), the library supports two different sizes for Initialization Vector (IV):

  • 12 bytes: used mainly for IPSec, as it requires 12 bytes from the user, which internally are appended the counter block (4 bytes), which is set to 1 for the first block (no padding required from the user)
  • 16 bytes: when passing 16 bytes, the library will take them and use the last 4 bytes as the initial counter block for the first block.

AES-NI GCM Crypto Poll Mode Driver

The AES-NI GCM PMD (librte_pmd_aesni_gcm) provides poll mode crypto driver support for utilizing Intel multi buffer library (see AES-NI Multi-buffer PMD documentation to learn more about it, including installation).

Features

AESNI GCM PMD has support for:

Authentication algorithms:

  • RTE_CRYPTO_AUTH_AES_GMAC

AEAD algorithms:

  • RTE_CRYPTO_AEAD_AES_GCM

Limitations

  • Chained mbufs are supported but only out-of-place (destination mbuf must be contiguous).
  • Cipher only is not supported.

Installation

To build DPDK with the AESNI_GCM_PMD the user is required to download the multi-buffer library from here and compile it on their user system before building DPDK. The latest version of the library supported by this PMD is v0.47, which can be downloaded in https://github.com/01org/intel-ipsec-mb/archive/v0.47.zip.

make

As a reference, the following table shows a mapping between the past DPDK versions and the external crypto libraries supported by them:

DPDK and external crypto library version compatibility
DPDK version Crypto library version
16.04 - 16.11 Multi-buffer library 0.43 - 0.44
17.02 - 17.05 ISA-L Crypto v2.18
17.08+ Multi-buffer library 0.46+

Initialization

In order to enable this virtual crypto PMD, user must:

  • Export the environmental variable AESNI_MULTI_BUFFER_LIB_PATH with the path where the library was extracted.
  • Build the multi buffer library (explained in Installation section).
  • Set CONFIG_RTE_LIBRTE_PMD_AESNI_GCM=y in config/common_base.

To use the PMD in an application, user must:

  • Call rte_vdev_init(“crypto_aesni_gcm”) within the application.
  • Use –vdev=”crypto_aesni_gcm” in the EAL options, which will call rte_vdev_init() internally.

The following parameters (all optional) can be provided in the previous two calls:

  • socket_id: Specify the socket where the memory for the device is going to be allocated (by default, socket_id will be the socket where the core that is creating the PMD is running on).
  • max_nb_queue_pairs: Specify the maximum number of queue pairs in the device (8 by default).
  • max_nb_sessions: Specify the maximum number of sessions that can be created (2048 by default).

Example:

./l2fwd-crypto -l 1 -n 4 --vdev="crypto_aesni_gcm,socket_id=0,max_nb_sessions=128" \
-- -p 1 --cdev SW --chain AEAD --aead_algo "aes-gcm"

ARMv8 Crypto Poll Mode Driver

This code provides the initial implementation of the ARMv8 crypto PMD. The driver uses ARMv8 cryptographic extensions to process chained crypto operations in an optimized way. The core functionality is provided by a low-level library, written in the assembly code.

Features

ARMv8 Crypto PMD has support for the following algorithm pairs:

Supported cipher algorithms:

  • RTE_CRYPTO_CIPHER_AES_CBC

Supported authentication algorithms:

  • RTE_CRYPTO_AUTH_SHA1_HMAC
  • RTE_CRYPTO_AUTH_SHA256_HMAC

Installation

In order to enable this virtual crypto PMD, user must:

  • Download ARMv8 crypto library source code from here
  • Export the environmental variable ARMV8_CRYPTO_LIB_PATH with the path where the armv8_crypto library was downloaded or cloned.
  • Build the library by invoking:
make -C $ARMV8_CRYPTO_LIB_PATH/
  • Set CONFIG_RTE_LIBRTE_PMD_ARMV8_CRYPTO=y in config/defconfig_arm64-armv8a-linuxapp-gcc

The corresponding device can be created only if the following features are supported by the CPU:

  • RTE_CPUFLAG_AES
  • RTE_CPUFLAG_SHA1
  • RTE_CPUFLAG_SHA2
  • RTE_CPUFLAG_NEON

Initialization

User can use app/test application to check how to use this PMD and to verify crypto processing.

Test name is cryptodev_sw_armv8_autotest. For performance test cryptodev_sw_armv8_perftest can be used.

Limitations

  • Maximum number of sessions is 2048.
  • Only chained operations are supported.
  • AES-128-CBC is the only supported cipher variant.
  • Cipher input data has to be a multiple of 16 bytes.
  • Digest input data has to be a multiple of 8 bytes.

NXP DPAA2 CAAM (DPAA2_SEC)

The DPAA2_SEC PMD provides poll mode crypto driver support for NXP DPAA2 CAAM hardware accelerator.

Architecture

SEC is the SOC’s security engine, which serves as NXP’s latest cryptographic acceleration and offloading hardware. It combines functions previously implemented in separate modules to create a modular and scalable acceleration and assurance engine. It also implements block encryption algorithms, stream cipher algorithms, hashing algorithms, public key algorithms, run-time integrity checking, and a hardware random number generator. SEC performs higher-level cryptographic operations than previous NXP cryptographic accelerators. This provides significant improvement to system level performance.

DPAA2_SEC is one of the hardware resource in DPAA2 Architecture. More information on DPAA2 Architecture is described in DPAA2 Overview.

DPAA2_SEC PMD is one of DPAA2 drivers which interacts with Management Complex (MC) portal to access the hardware object - DPSECI. The MC provides access to create, discover, connect, configure and destroy dpseci objects in DPAA2_SEC PMD.

DPAA2_SEC PMD also uses some of the other hardware resources like buffer pools, queues, queue portals to store and to enqueue/dequeue data to the hardware SEC.

DPSECI objects are detected by PMD using a resource container called DPRC (like in DPAA2 Overview).

For example:

DPRC.1 (bus)
  |
  +--+--------+-------+-------+-------+---------+
     |        |       |       |       |         |
   DPMCP.1  DPIO.1  DPBP.1  DPNI.1  DPMAC.1  DPSECI.1
   DPMCP.2  DPIO.2          DPNI.2  DPMAC.2  DPSECI.2
   DPMCP.3

Implementation

SEC provides platform assurance by working with SecMon, which is a companion logic block that tracks the security state of the SOC. SEC is programmed by means of descriptors (not to be confused with frame descriptors (FDs)) that indicate the operations to be performed and link to the message and associated data. SEC incorporates two DMA engines to fetch the descriptors, read the message data, and write the results of the operations. The DMA engine provides a scatter/gather capability so that SEC can read and write data scattered in memory. SEC may be configured by means of software for dynamic changes in byte ordering. The default configuration for this version of SEC is little-endian mode.

A block diagram similar to dpaa2 NIC is shown below to show where DPAA2_SEC fits in the DPAA2 Bus model

                                   +----------------+
                                   | DPDK DPAA2_SEC |
                                   |     PMD        |
                                   +----------------+       +------------+
                                   |  MC SEC object |.......|  Mempool   |
                . . . . . . . . .  |   (DPSECI)     |       |  (DPBP)    |
               .                   +---+---+--------+       +-----+------+
              .                        ^   |                      .
             .                         |   |<enqueue,             .
            .                          |   | dequeue>             .
           .                           |   |                      .
          .                        +---+---V----+                 .
         .      . . . . . . . . . .| DPIO driver|                 .
        .      .                   |  (DPIO)    |                 .
       .      .                    +-----+------+                 .
      .      .                     |  QBMAN     |                 .
     .      .                      |  Driver    |                 .
+----+------+-------+              +-----+----- |                 .
|   dpaa2 bus       |                    |                        .
|   VFIO fslmc-bus  |....................|.........................
|                   |                    |
|     /bus/fslmc    |                    |
+-------------------+                    |
                                         |
========================== HARDWARE =====|=======================
                                       DPIO
                                         |
                                       DPSECI---DPBP
=========================================|========================

Features

The DPAA2_SEC PMD has support for:

Cipher algorithms:

  • RTE_CRYPTO_CIPHER_3DES_CBC
  • RTE_CRYPTO_CIPHER_AES128_CBC
  • RTE_CRYPTO_CIPHER_AES192_CBC
  • RTE_CRYPTO_CIPHER_AES256_CBC
  • RTE_CRYPTO_CIPHER_AES128_CTR
  • RTE_CRYPTO_CIPHER_AES192_CTR
  • RTE_CRYPTO_CIPHER_AES256_CTR

Hash algorithms:

  • RTE_CRYPTO_AUTH_SHA1_HMAC
  • RTE_CRYPTO_AUTH_SHA224_HMAC
  • RTE_CRYPTO_AUTH_SHA256_HMAC
  • RTE_CRYPTO_AUTH_SHA384_HMAC
  • RTE_CRYPTO_AUTH_SHA512_HMAC
  • RTE_CRYPTO_AUTH_MD5_HMAC

AEAD algorithms:

  • RTE_CRYPTO_AEAD_AES_GCM

Supported DPAA2 SoCs

  • LS2080A/LS2040A
  • LS2084A/LS2044A
  • LS2088A/LS2048A
  • LS1088A/LS1048A

Limitations

  • Chained mbufs are not supported.
  • Hash followed by Cipher mode is not supported
  • Only supports the session-oriented API implementation (session-less APIs are not supported).

Prerequisites

DPAA2_SEC driver has similar pre-requisites as described in DPAA2 Overview. The following dependencies are not part of DPDK and must be installed separately:

  • NXP Linux SDK

    NXP Linux software development kit (SDK) includes support for the family of QorIQ® ARM-Architecture-based system on chip (SoC) processors and corresponding boards.

    It includes the Linux board support packages (BSPs) for NXP SoCs, a fully operational tool chain, kernel and board specific modules.

    SDK and related information can be obtained from: NXP QorIQ SDK.

  • DPDK Extra Scripts

    DPAA2 based resources can be configured easily with the help of ready scripts as provided in the DPDK helper repository.

    DPDK Extra Scripts.

Currently supported by DPDK:

  • NXP SDK 17.08+.
  • MC Firmware version 10.3.1 and higher.
  • Supported architectures: arm64 LE.
  • Follow the DPDK Getting Started Guide for Linux to setup the basic DPDK environment.

Pre-Installation Configuration

Config File Options

Basic DPAA2 config file options are described in DPAA2 Overview. In addition to those, the following options can be modified in the config file to enable DPAA2_SEC PMD.

Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_LIBRTE_PMD_DPAA2_SEC (default n) By default it is only enabled in defconfig_arm64-dpaa2-* config. Toggle compilation of the librte_pmd_dpaa2_sec driver.
  • CONFIG_RTE_LIBRTE_DPAA2_SEC_DEBUG_INIT (default n) Toggle display of initialization related driver messages
  • CONFIG_RTE_LIBRTE_DPAA2_SEC_DEBUG_DRIVER (default n) Toggle display of driver runtime messages
  • CONFIG_RTE_LIBRTE_DPAA2_SEC_DEBUG_RX (default n) Toggle display of receive fast path run-time message
  • CONFIG_RTE_DPAA2_SEC_PMD_MAX_NB_SESSIONS By default it is set as 2048 in defconfig_arm64-dpaa2-* config. It indicates Number of sessions to create in the session memory pool on a single DPAA2 SEC device.

Installations

To compile the DPAA2_SEC PMD for Linux arm64 gcc target, run the following make command:

cd <DPDK-source-directory>
make config T=arm64-dpaa2-linuxapp-gcc install

NXP DPAA CAAM (DPAA_SEC)

The DPAA_SEC PMD provides poll mode crypto driver support for NXP DPAA CAAM hardware accelerator.

Architecture

SEC is the SOC’s security engine, which serves as NXP’s latest cryptographic acceleration and offloading hardware. It combines functions previously implemented in separate modules to create a modular and scalable acceleration and assurance engine. It also implements block encryption algorithms, stream cipher algorithms, hashing algorithms, public key algorithms, run-time integrity checking, and a hardware random number generator. SEC performs higher-level cryptographic operations than previous NXP cryptographic accelerators. This provides significant improvement to system level performance.

DPAA_SEC is one of the hardware resource in DPAA Architecture. More information on DPAA Architecture is described in DPAA Overview.

DPAA_SEC PMD is one of DPAA drivers which interacts with QBMAN to create, configure and destroy the device instance using queue pair with CAAM portal.

DPAA_SEC PMD also uses some of the other hardware resources like buffer pools, queues, queue portals to store and to enqueue/dequeue data to the hardware SEC.

Implementation

SEC provides platform assurance by working with SecMon, which is a companion logic block that tracks the security state of the SOC. SEC is programmed by means of descriptors (not to be confused with frame descriptors (FDs)) that indicate the operations to be performed and link to the message and associated data. SEC incorporates two DMA engines to fetch the descriptors, read the message data, and write the results of the operations. The DMA engine provides a scatter/gather capability so that SEC can read and write data scattered in memory. SEC may be configured by means of software for dynamic changes in byte ordering. The default configuration for this version of SEC is little-endian mode.

Features

The DPAA PMD has support for:

Cipher algorithms:

  • RTE_CRYPTO_CIPHER_3DES_CBC
  • RTE_CRYPTO_CIPHER_AES128_CBC
  • RTE_CRYPTO_CIPHER_AES192_CBC
  • RTE_CRYPTO_CIPHER_AES256_CBC
  • RTE_CRYPTO_CIPHER_AES128_CTR
  • RTE_CRYPTO_CIPHER_AES192_CTR
  • RTE_CRYPTO_CIPHER_AES256_CTR

Hash algorithms:

  • RTE_CRYPTO_AUTH_SHA1_HMAC
  • RTE_CRYPTO_AUTH_SHA224_HMAC
  • RTE_CRYPTO_AUTH_SHA256_HMAC
  • RTE_CRYPTO_AUTH_SHA384_HMAC
  • RTE_CRYPTO_AUTH_SHA512_HMAC
  • RTE_CRYPTO_AUTH_MD5_HMAC

AEAD algorithms:

  • RTE_CRYPTO_AEAD_AES_GCM

Supported DPAA SoCs

  • LS1046A/LS1026A
  • LS1043A/LS1023A

Limitations

  • Chained mbufs are not supported.
  • Hash followed by Cipher mode is not supported
  • Only supports the session-oriented API implementation (session-less APIs are not supported).

Prerequisites

DPAA_SEC driver has similar pre-requisites as described in DPAA Overview. The following dependencies are not part of DPDK and must be installed separately:

  • NXP Linux SDK

    NXP Linux software development kit (SDK) includes support for the family of QorIQ® ARM-Architecture-based system on chip (SoC) processors and corresponding boards.

    It includes the Linux board support packages (BSPs) for NXP SoCs, a fully operational tool chain, kernel and board specific modules.

    SDK and related information can be obtained from: NXP QorIQ SDK.

  • DPDK Extras Scripts

    DPAA based resources can be configured easily with the help of ready scripts as provided in the DPDK Extras repository.

    DPDK Extras Scripts.

Currently supported by DPDK:

Pre-Installation Configuration

Config File Options

Basic DPAA config file options are described in DPAA Overview. In addition to those, the following options can be modified in the config file to enable DPAA_SEC PMD.

Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_LIBRTE_PMD_DPAA_SEC (default n) By default it is only enabled in defconfig_arm64-dpaa-* config. Toggle compilation of the librte_pmd_dpaa_sec driver.
  • CONFIG_RTE_LIBRTE_DPAA_SEC_DEBUG_INIT (default n) Toggle display of initialization related driver messages
  • CONFIG_RTE_LIBRTE_DPAA_SEC_DEBUG_DRIVER (default n) Toggle display of driver runtime messages
  • CONFIG_RTE_LIBRTE_DPAA_SEC_DEBUG_RX (default n) Toggle display of receive fast path run-time message
  • CONFIG_RTE_DPAA_SEC_PMD_MAX_NB_SESSIONS By default it is set as 2048 in defconfig_arm64-dpaa-* config. It indicates Number of sessions to create in the session memory pool on a single DPAA SEC device.

Installations

To compile the DPAA_SEC PMD for Linux arm64 gcc target, run the following make command:

cd <DPDK-source-directory>
make config T=arm64-dpaa-linuxapp-gcc install

KASUMI Crypto Poll Mode Driver

The KASUMI PMD (librte_pmd_kasumi) provides poll mode crypto driver support for utilizing Intel Libsso library, which implements F8 and F9 functions for KASUMI UEA1 cipher and UIA1 hash algorithms.

Features

KASUMI PMD has support for:

Cipher algorithm:

  • RTE_CRYPTO_CIPHER_KASUMI_F8

Authentication algorithm:

  • RTE_CRYPTO_AUTH_KASUMI_F9

Limitations

  • Chained mbufs are not supported.
  • KASUMI(F9) supported only if hash offset and length field is byte-aligned.
  • In-place bit-level operations for KASUMI(F8) are not supported (if length and/or offset of data to be ciphered is not byte-aligned).

Installation

To build DPDK with the KASUMI_PMD the user is required to download the export controlled libsso_kasumi library, by requesting it from https://networkbuilders.intel.com/network-technologies/dpdk. Once approval has been granted, the user needs to log in https://networkbuilders.intel.com/dpdklogin and click on “Kasumi Bit Stream crypto library” link, to download the library. After downloading the library, the user needs to unpack and compile it on their system before building DPDK:

make

Note: When encrypting with KASUMI F8, by default the library encrypts full blocks of 8 bytes, regardless the number of bytes to be encrypted provided (which leads to a possible buffer overflow). To avoid this situation, it is necessary not to pass 3GPP_SAFE_BUFFERS as a compilation flag. Also, this is required when using chained operations (cipher-then-auth/auth-then-cipher). For this, in the Makefile of the library, make sure that this flag is commented out:

#EXTRA_CFLAGS  += -D_3GPP_SAFE_BUFFERS

Note: To build the PMD as a shared library, the libsso_kasumi library must be built as follows:

make KASUMI_CFLAGS=-DKASUMI_C

Initialization

In order to enable this virtual crypto PMD, user must:

  • Export the environmental variable LIBSSO_KASUMI_PATH with the path where the library was extracted (kasumi folder).
  • Build the LIBSSO library (explained in Installation section).
  • Set CONFIG_RTE_LIBRTE_PMD_KASUMI=y in config/common_base.

To use the PMD in an application, user must:

  • Call rte_vdev_init(“crypto_kasumi”) within the application.
  • Use –vdev=”crypto_kasumi” in the EAL options, which will call rte_vdev_init() internally.

The following parameters (all optional) can be provided in the previous two calls:

  • socket_id: Specify the socket where the memory for the device is going to be allocated (by default, socket_id will be the socket where the core that is creating the PMD is running on).
  • max_nb_queue_pairs: Specify the maximum number of queue pairs in the device (8 by default).
  • max_nb_sessions: Specify the maximum number of sessions that can be created (2048 by default).

Example:

./l2fwd-crypto -l 1 -n 4 --vdev="crypto_kasumi,socket_id=0,max_nb_sessions=128" \
-- -p 1 --cdev SW --chain CIPHER_ONLY --cipher_algo "kasumi-f8"

Extra notes on KASUMI F9

When using KASUMI F9 authentication algorithm, the input buffer must be constructed according to the 3GPP KASUMI specifications (section 4.4, page 13): http://cryptome.org/3gpp/35201-900.pdf. Input buffer has to have COUNT (4 bytes), FRESH (4 bytes), MESSAGE and DIRECTION (1 bit) concatenated. After the DIRECTION bit, a single ‘1’ bit is appended, followed by between 0 and 7 ‘0’ bits, so that the total length of the buffer is multiple of 8 bits. Note that the actual message can be any length, specified in bits.

Once this buffer is passed this way, when creating the crypto operation, length of data to authenticate (op.sym.auth.data.length) must be the length of all the items described above, including the padding at the end. Also, offset of data to authenticate (op.sym.auth.data.offset) must be such that points at the start of the COUNT bytes.

OpenSSL Crypto Poll Mode Driver

This code provides the initial implementation of the openssl poll mode driver. All cryptography operations are using Openssl library crypto API. Each algorithm uses EVP interface from openssl API - which is recommended by Openssl maintainers.

For more details about openssl library please visit openssl webpage: https://www.openssl.org/

Features

OpenSSL PMD has support for:

Supported cipher algorithms: * RTE_CRYPTO_CIPHER_3DES_CBC * RTE_CRYPTO_CIPHER_AES_CBC * RTE_CRYPTO_CIPHER_AES_CTR * RTE_CRYPTO_CIPHER_3DES_CTR * RTE_CRYPTO_CIPHER_DES_DOCSISBPI

Supported authentication algorithms: * RTE_CRYPTO_AUTH_AES_GMAC * RTE_CRYPTO_AUTH_MD5 * RTE_CRYPTO_AUTH_SHA1 * RTE_CRYPTO_AUTH_SHA224 * RTE_CRYPTO_AUTH_SHA256 * RTE_CRYPTO_AUTH_SHA384 * RTE_CRYPTO_AUTH_SHA512 * RTE_CRYPTO_AUTH_MD5_HMAC * RTE_CRYPTO_AUTH_SHA1_HMAC * RTE_CRYPTO_AUTH_SHA224_HMAC * RTE_CRYPTO_AUTH_SHA256_HMAC * RTE_CRYPTO_AUTH_SHA384_HMAC * RTE_CRYPTO_AUTH_SHA512_HMAC

Supported AEAD algorithms: * RTE_CRYPTO_AEAD_AES_GCM * RTE_CRYPTO_AEAD_AES_CCM

Installation

To compile openssl PMD, it has to be enabled in the config/common_base file and appropriate openssl packages have to be installed in the build environment.

The newest openssl library version is supported: * 1.0.2h-fips 3 May 2016. Older versions that were also verified: * 1.0.1f 6 Jan 2014 * 1.0.1 14 Mar 2012

For Ubuntu 14.04 LTS these packages have to be installed in the build system: sudo apt-get install openssl sudo apt-get install libc6-dev-i386 (for i686-native-linuxapp-gcc target)

This code was also verified on Fedora 24. This code was NOT yet verified on FreeBSD.

Initialization

User can use app/test application to check how to use this pmd and to verify crypto processing.

Test name is cryptodev_openssl_autotest. For performance test cryptodev_openssl_perftest can be used.

To verify real traffic l2fwd-crypto example can be used with this command:

sudo ./build/l2fwd-crypto -l 0-1 -n 4 --vdev "crypto_openssl"
--vdev "crypto_openssl"-- -p 0x3 --chain CIPHER_HASH
--cipher_op ENCRYPT --cipher_algo AES_CBC
--cipher_key 00:01:02:03:04:05:06:07:08:09:0a:0b:0c:0d:0e:0f
--iv 00:01:02:03:04:05:06:07:08:09:0a:0b:0c:0d:0e:ff
--auth_op GENERATE --auth_algo SHA1_HMAC
--auth_key 11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11
:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11
:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11

Limitations

  • Maximum number of sessions is 2048.
  • Chained mbufs are supported only for source mbuf (destination must be contiguous).
  • Hash only is not supported for GCM and GMAC.
  • Cipher only is not supported for GCM and GMAC.

MRVL Crypto Poll Mode Driver

The MRVL CRYPTO PMD (librte_crypto_mrvl_pmd) provides poll mode crypto driver support by utilizing MUSDK library, which provides cryptographic operations acceleration by using Security Acceleration Engine (EIP197) directly from user-space with minimum overhead and high performance.

Features

MRVL CRYPTO PMD has support for:

  • Symmetric crypto
  • Sym operation chaining
  • AES CBC (128)
  • AES CBC (192)
  • AES CBC (256)
  • AES CTR (128)
  • AES CTR (192)
  • AES CTR (256)
  • 3DES CBC
  • 3DES CTR
  • MD5
  • MD5 HMAC
  • SHA1
  • SHA1 HMAC
  • SHA256
  • SHA256 HMAC
  • SHA384
  • SHA384 HMAC
  • SHA512
  • SHA512 HMAC
  • AES GCM (128)

Limitations

  • Hardware only supports scenarios where ICV (digest buffer) is placed just after the authenticated data. Other placement will result in error.

  • Before running crypto test suite it is advised to increase limit of opened files:

    ulimit -n 20000
    

Installation

MRVL CRYPTO PMD driver compilation is disabled by default due to external dependencies. Currently there are two driver specific compilation options in config/common_base available:

  • CONFIG_RTE_LIBRTE_MRVL_CRYPTO (default n)

    Toggle compilation of the librte_pmd_mrvl driver.

  • CONFIG_RTE_LIBRTE_MRVL_CRYPTO_DEBUG (default n)

    Toggle display of debugging messages.

During compilation external MUSDK library, which provides direct access to Marvell’s EIP197 cryptographic engine, is necessary. Library sources are available here.

Alternatively, prebuilt library can be downloaded from Marvell Extranet. Once approval has been granted, library can be found by typing musdk in the search box.

For MUSDK library build instructions please refer to doc/musdk_get_started.txt in library sources directory.

MUSDK requires out of tree kernel modules to work. Kernel tree needed to build them is available here.

Initialization

After successfully building MRVL CRYPTO PMD, the following modules need to be loaded:

insmod musdk_uio.ko
insmod mvpp2x_sysfs.ko
insmod mv_pp_uio.ko
insmod mv_sam_uio.ko
insmod crypto_safexcel.ko
  • musdk_uio.ko, mv_pp2_uio.ko and mv_sam_uio.ko are distributed together with MUSDK library.
  • crypto_safexcel.ko is an in-kernel module.
  • mvpp2x_sysfs.ko can be build from sources available here.

The following parameters (all optional) are exported by the driver:

  • max_nb_queue_pairs: maximum number of queue pairs in the device (8 by default).
  • max_nb_sessions: maximum number of sessions that can be created (2048 by default).
  • socket_id: socket on which to allocate the device resources on.

l2fwd-crypto example application can be used to verify MRVL CRYPTO PMD operation:

./l2fwd-crypto --vdev=net_mrvl,iface=eth0 --vdev=crypto_mrvl -- \
  --cipher_op ENCRYPT --cipher_algo aes-cbc \
  --cipher_key 00:01:02:03:04:05:06:07:08:09:0a:0b:0c:0d:0e:0f  \
  --auth_op GENERATE --auth_algo sha1-hmac \
  --auth_key 10:11:12:13:14:15:16:17:18:19:1a:1b:1c:1d:1e:1f

Example output:

[...]
AAD: at [0x7f253ceb80], len=
P ID 0 configuration ----
Port mode               : KR
MAC status              : disabled
Link status             : link up
Port speed              : 10G
Port duplex             : full
Port: Egress enable tx_port_num=16 qmap=0x1
PORT: Port0 - link
P ID 0 configuration ----
Port mode               : KR
MAC status              : disabled
Link status             : link down
Port speed              : 10G
Port duplex             : full
Port: Egress enable tx_port_num=16 qmap=0x1
Port 0, MAC address: 00:50:43:02:21:20


Checking link statusdone
Port 0 Link Up - speed 0 Mbps - full-duplex
Lcore 0: RX port 0
Allocated session pool on socket 0
eip197: 0:0 registers: paddr: 0xf2880000, vaddr: 0x0x7f56a80000
DMA buffer (131136 bytes) for CDR #0 allocated: paddr = 0xb0585e00, vaddr = 0x7f09384e00
DMA buffer (131136 bytes) for RDR #0 allocated: paddr = 0xb05a5f00, vaddr = 0x7f093a4f00
DMA buffers allocated for 2049 operations. Tokens - 256 bytes
Lcore 0: cryptodev 0
L2FWD: lcore 1 has nothing to do
L2FWD: lcore 2 has nothing to do
L2FWD: lcore 3 has nothing to do
L2FWD: entering main loop on lcore 0
L2FWD:  -- lcoreid=0 portid=0
L2FWD:  -- lcoreid=0 cryptoid=0
Options:-
nportmask: ffffffff
ports per lcore: 1
refresh period : 10000
single lcore mode: disabled
stats_printing: enabled
sessionless crypto: disabled

Crypto chain: Input --> Encrypt --> Auth generate --> Output

---- Cipher information ---
Algorithm: aes-cbc
Cipher key: at [0x7f56db4e80], len=16
00000000: 00 01 02 03 04 05 06 07 08 09 0A 0B 0C 0D 0E 0F | ................
IV: at [0x7f56db4b80], len=16
00000000: 20 F0 63 0E 45 EB 2D 84 72 D4 13 6E 36 B5 AF FE |  .c.E.-.r..n6...

---- Authentication information ---
Algorithm: sha1-hmac
Auth key: at [0x7f56db4d80], len=16
00000000: 10 11 12 13 14 15 16 17 18 19 1A 1B 1C 1D 1E 1F | ................
IV: at [0x7f56db4a80], len=0
AAD: at [0x7f253ceb80], len=

Null Crypto Poll Mode Driver

The Null Crypto PMD (librte_pmd_null_crypto) provides a crypto poll mode driver which provides a minimal implementation for a software crypto device. As a null device it does not modify the data in the mbuf on which the crypto operation is to operate and it only has support for a single cipher and authentication algorithm.

When a burst of mbufs is submitted to a Null Crypto PMD for processing then each mbuf in the burst will be enqueued in an internal buffer for collection on a dequeue call as long as the mbuf has a valid rte_mbuf_offload operation with a valid rte_cryptodev_session or rte_crypto_xform chain of operations.

Features

Modes:

  • RTE_CRYPTO_XFORM_CIPHER ONLY
  • RTE_CRYPTO_XFORM_AUTH ONLY
  • RTE_CRYPTO_XFORM_CIPHER THEN RTE_CRYPTO_XFORM_AUTH
  • RTE_CRYPTO_XFORM_AUTH THEN RTE_CRYPTO_XFORM_CIPHER

Cipher algorithms:

  • RTE_CRYPTO_CIPHER_NULL

Authentication algorithms:

  • RTE_CRYPTO_AUTH_NULL

Limitations

  • Only in-place is currently supported (destination address is the same as source address).

Installation

The Null Crypto PMD is enabled and built by default in both the Linux and FreeBSD builds.

Initialization

To use the PMD in an application, user must:

  • Call rte_vdev_init(“crypto_null”) within the application.
  • Use –vdev=”crypto_null” in the EAL options, which will call rte_vdev_init() internally.

The following parameters (all optional) can be provided in the previous two calls:

  • socket_id: Specify the socket where the memory for the device is going to be allocated (by default, socket_id will be the socket where the core that is creating the PMD is running on).
  • max_nb_queue_pairs: Specify the maximum number of queue pairs in the device (8 by default).
  • max_nb_sessions: Specify the maximum number of sessions that can be created (2048 by default).

Example:

./l2fwd-crypto -l 1 -n 4 --vdev="crypto_null,socket_id=0,max_nb_sessions=128" \
-- -p 1 --cdev SW --chain CIPHER_ONLY --cipher_algo "null"

Cryptodev Scheduler Poll Mode Driver Library

Scheduler PMD is a software crypto PMD, which has the capabilities of attaching hardware and/or software cryptodevs, and distributes ingress crypto ops among them in a certain manner.

_images/scheduler-overview.svg

Cryptodev Scheduler Overview

The Cryptodev Scheduler PMD library (librte_pmd_crypto_scheduler) acts as a software crypto PMD and shares the same API provided by librte_cryptodev. The PMD supports attaching multiple crypto PMDs, software or hardware, as slaves, and distributes the crypto workload to them with certain behavior. The behaviors are categorizes as different “modes”. Basically, a scheduling mode defines certain actions for scheduling crypto ops to its slaves.

The librte_pmd_crypto_scheduler library exports a C API which provides an API for attaching/detaching slaves, set/get scheduling modes, and enable/disable crypto ops reordering.

Limitations

  • Sessionless crypto operation is not supported
  • OOP crypto operation is not supported when the crypto op reordering feature is enabled.

Installation

To build DPDK with CRYTPO_SCHEDULER_PMD the user is required to set CONFIG_RTE_LIBRTE_PMD_CRYPTO_SCHEDULER=y in config/common_base, and recompile DPDK

Initialization

To use the PMD in an application, user must:

  • Call rte_vdev_init(“crypto_scheduler”) within the application.
  • Use –vdev=”crypto_scheduler” in the EAL options, which will call rte_vdev_init() internally.

The following parameters (all optional) can be provided in the previous two calls:

  • socket_id: Specify the socket where the memory for the device is going to be allocated (by default, socket_id will be the socket where the core that is creating the PMD is running on).
  • max_nb_sessions: Specify the maximum number of sessions that can be created. This value may be overwritten internally if there are too many devices are attached.
  • slave: If a cryptodev has been initialized with specific name, it can be attached to the scheduler using this parameter, simply filling the name here. Multiple cryptodevs can be attached initially by presenting this parameter multiple times.
  • mode: Specify the scheduling mode of the PMD. The supported scheduling mode parameter values are specified in the “Cryptodev Scheduler Modes Overview” section.
  • ordering: Specify the status of the crypto operations ordering feature. The value of this parameter can be “enable” or “disable”. This feature is disabled by default.

Example:

... --vdev "crypto_aesni_mb0,name=aesni_mb_1" --vdev "crypto_aesni_mb1,name=aesni_mb_2" --vdev "crypto_scheduler,slave=aesni_mb_1,slave=aesni_mb_2" ...

Note

  • The scheduler cryptodev cannot be started unless the scheduling mode is set and at least one slave is attached. Also, to configure the scheduler in the run-time, like attach/detach slave(s), change scheduling mode, or enable/disable crypto op ordering, one should stop the scheduler first, otherwise an error will be returned.
  • The crypto op reordering feature requires using the userdata field of every mbuf to be processed to store temporary data. By the end of processing, the field is set to pointing to NULL, any previously stored value of this field will be lost.

Cryptodev Scheduler Modes Overview

Currently the Crypto Scheduler PMD library supports following modes of operation:

  • CDEV_SCHED_MODE_ROUNDROBIN:

Initialization mode parameter: round-robin

Round-robin mode, which distributes the enqueued burst of crypto ops among its slaves in a round-robin manner. This mode may help to fill the throughput gap between the physical core and the existing cryptodevs to increase the overall performance.

  • CDEV_SCHED_MODE_PKT_SIZE_DISTR:

Initialization mode parameter: packet-size-distr

Packet-size based distribution mode, which works with 2 slaves, the primary slave and the secondary slave, and distributes the enqueued crypto operations to them based on their data lengths. A crypto operation will be distributed to the primary slave if its data length is equal to or bigger than the designated threshold, otherwise it will be handled by the secondary slave.

A typical usecase in this mode is with the QAT cryptodev as the primary and a software cryptodev as the secondary slave. This may help applications to process additional crypto workload than what the QAT cryptodev can handle on its own, by making use of the available CPU cycles to deal with smaller crypto workloads.

The threshold is set to 128 bytes by default. It can be updated by calling function rte_cryptodev_scheduler_option_set. The parameter of option_type must be CDEV_SCHED_OPTION_THRESHOLD and option should point to a rte_cryptodev_scheduler_threshold_option structure filled with appropriate threshold value. Please NOTE this threshold has be a power-of-2 unsigned integer.

  • CDEV_SCHED_MODE_FAILOVER:

Initialization mode parameter: fail-over

Fail-over mode, which works with 2 slaves, the primary slave and the secondary slave. In this mode, the scheduler will enqueue the incoming crypto operation burst to the primary slave. When one or more crypto operations fail to be enqueued, then they will be enqueued to the secondary slave.

  • CDEV_SCHED_MODE_MULTICORE:

Initialization mode parameter: multi-core

Multi-core mode, which distributes the workload with several (up to eight) worker cores. The enqueued bursts are distributed among the worker cores in a round-robin manner. If scheduler cannot enqueue entire burst to the same worker, it will enqueue the remaining operations to the next available worker. For pure small packet size (64 bytes) traffic however the multi-core mode is not an optimal solution, as it doesn’t give significant per-core performance improvement. For mixed traffic (IMIX) the optimal number of worker cores is around 2-3. For large packets (1.5 Kbytes) scheduler shows linear scaling in performance up to eight cores. Each worker uses its own slave cryptodev. Only software cryptodevs are supported. Only the same type of cryptodevs should be used concurrently.

The multi-core mode uses one extra parameter:

  • corelist: Semicolon-separated list of logical cores to be used as workers. The number of worker cores should be equal to the number of slave cryptodevs. These cores should be present in EAL core list parameter and should not be used by the application or any other process.
Example:
... –vdev “crypto_aesni_mb1,name=aesni_mb_1” –vdev “crypto_aesni_mb_pmd2,name=aesni_mb_2” –vdev “crypto_scheduler,slave=aesni_mb_1,slave=aesni_mb_2,mode=multi-core,corelist=23;24” ...

SNOW 3G Crypto Poll Mode Driver

The SNOW 3G PMD (librte_pmd_snow3g) provides poll mode crypto driver support for utilizing Intel Libsso library, which implements F8 and F9 functions for SNOW 3G UEA2 cipher and UIA2 hash algorithms.

Features

SNOW 3G PMD has support for:

Cipher algorithm:

  • RTE_CRYPTO_CIPHER_SNOW3G_UEA2

Authentication algorithm:

  • RTE_CRYPTO_AUTH_SNOW3G_UIA2

Limitations

  • Chained mbufs are not supported.
  • SNOW 3G (UIA2) supported only if hash offset field is byte-aligned.
  • In-place bit-level operations for SNOW 3G (UEA2) are not supported (if length and/or offset of data to be ciphered is not byte-aligned).

Installation

To build DPDK with the SNOW3G_PMD the user is required to download the export controlled libsso_snow3g library, by requesting it from https://networkbuilders.intel.com/network-technologies/dpdk. Once approval has been granted, the user needs to log in https://networkbuilders.intel.com/dpdklogin and click on “Snow3G Bit Stream crypto library” link, to download the library. After downloading the library, the user needs to unpack and compile it on their system before building DPDK:

make snow3G

Note: When encrypting with SNOW3G UEA2, by default the library encrypts blocks of 4 bytes, regardless the number of bytes to be encrypted provided (which leads to a possible buffer overflow). To avoid this situation, it is necessary not to pass 3GPP_SAFE_BUFFERS as a compilation flag. For this, in the Makefile of the library, make sure that this flag is commented out.:

#EXTRA_CFLAGS  += -D_3GPP_SAFE_BUFFERS

Initialization

In order to enable this virtual crypto PMD, user must:

  • Export the environmental variable LIBSSO_SNOW3G_PATH with the path where the library was extracted (snow3g folder).
  • Build the LIBSSO_SNOW3G library (explained in Installation section).
  • Set CONFIG_RTE_LIBRTE_PMD_SNOW3G=y in config/common_base.

To use the PMD in an application, user must:

  • Call rte_vdev_init(“crypto_snow3g”) within the application.
  • Use –vdev=”crypto_snow3g” in the EAL options, which will call rte_vdev_init() internally.

The following parameters (all optional) can be provided in the previous two calls:

  • socket_id: Specify the socket where the memory for the device is going to be allocated (by default, socket_id will be the socket where the core that is creating the PMD is running on).
  • max_nb_queue_pairs: Specify the maximum number of queue pairs in the device (8 by default).
  • max_nb_sessions: Specify the maximum number of sessions that can be created (2048 by default).

Example:

./l2fwd-crypto -l 1 -n 4 --vdev="crypto_snow3g,socket_id=0,max_nb_sessions=128" \
-- -p 1 --cdev SW --chain CIPHER_ONLY --cipher_algo "snow3g-uea2"

Intel(R) QuickAssist (QAT) Crypto Poll Mode Driver

The QAT PMD provides poll mode crypto driver support for the following hardware accelerator devices:

  • Intel QuickAssist Technology DH895xCC
  • Intel QuickAssist Technology C62x
  • Intel QuickAssist Technology C3xxx
  • Intel QuickAssist Technology D15xx

Features

The QAT PMD has support for:

Cipher algorithms:

  • RTE_CRYPTO_CIPHER_3DES_CBC
  • RTE_CRYPTO_CIPHER_3DES_CTR
  • RTE_CRYPTO_CIPHER_AES128_CBC
  • RTE_CRYPTO_CIPHER_AES192_CBC
  • RTE_CRYPTO_CIPHER_AES256_CBC
  • RTE_CRYPTO_CIPHER_AES128_CTR
  • RTE_CRYPTO_CIPHER_AES192_CTR
  • RTE_CRYPTO_CIPHER_AES256_CTR
  • RTE_CRYPTO_CIPHER_SNOW3G_UEA2
  • RTE_CRYPTO_CIPHER_NULL
  • RTE_CRYPTO_CIPHER_KASUMI_F8
  • RTE_CRYPTO_CIPHER_DES_CBC
  • RTE_CRYPTO_CIPHER_AES_DOCSISBPI
  • RTE_CRYPTO_CIPHER_DES_DOCSISBPI
  • RTE_CRYPTO_CIPHER_ZUC_EEA3

Hash algorithms:

  • RTE_CRYPTO_AUTH_SHA1_HMAC
  • RTE_CRYPTO_AUTH_SHA224_HMAC
  • RTE_CRYPTO_AUTH_SHA256_HMAC
  • RTE_CRYPTO_AUTH_SHA384_HMAC
  • RTE_CRYPTO_AUTH_SHA512_HMAC
  • RTE_CRYPTO_AUTH_AES_XCBC_MAC
  • RTE_CRYPTO_AUTH_SNOW3G_UIA2
  • RTE_CRYPTO_AUTH_MD5_HMAC
  • RTE_CRYPTO_AUTH_NULL
  • RTE_CRYPTO_AUTH_KASUMI_F9
  • RTE_CRYPTO_AUTH_AES_GMAC
  • RTE_CRYPTO_AUTH_ZUC_EIA3

Supported AEAD algorithms: * RTE_CRYPTO_AEAD_AES_GCM

Limitations

  • Only supports the session-oriented API implementation (session-less APIs are not supported).
  • SNOW 3G (UEA2), KASUMI (F8) and ZUC (EEA3) supported only if cipher length and offset fields are byte-multiple.
  • SNOW 3G (UIA2) and ZUC (EIA3) supported only if hash length and offset fields are byte-multiple.
  • No BSD support as BSD QAT kernel driver not available.
  • ZUC EEA3/EIA3 is not supported by dh895xcc devices
  • Maximum additional authenticated data (AAD) for GCM is 240 bytes long.
  • Queue pairs are not thread-safe (that is, within a single queue pair, RX and TX from different lcores is not supported).

Installation

To enable QAT in DPDK, follow the instructions for modifying the compile-time configuration file as described here.

Quick instructions are as follows:

cd to the top-level DPDK directory
make config T=x86_64-native-linuxapp-gcc
sed -i 's,\(CONFIG_RTE_LIBRTE_PMD_QAT\)=n,\1=y,' build/.config
make

To use the DPDK QAT PMD an SRIOV-enabled QAT kernel driver is required. The VF devices exposed by this driver will be used by the QAT PMD. The devices and available kernel drivers and device ids are :

QAT device generations, devices and drivers
Gen Device Driver Kernel Module Pci Driver PF Did #PFs Vf Did VFs/PF
1 DH895xCC 01.org icp_qa_al n/a 435 1 443 32
1 DH895xCC 4.4+ qat_dh895xcc dh895xcc 435 1 443 32
2 C62x 4.5+ qat_c62x c6xx 37c8 3 37c9 16
2 C3xxx 4.5+ qat_c3xxx c3xxx 19e2 1 19e3 16
2 D15xx p qat_d15xx d15xx 6f54 1 6f55 16

The Driver column indicates either the Linux kernel version in which support for this device was introduced or a driver available on Intel’s 01.org website. There are both linux and 01.org kernel drivers available for some devices. p = release pending.

If you are running on a kernel which includes a driver for your device, see Installation using kernel.org driver below. Otherwise see Installation using 01.org QAT driver.

Installation using kernel.org driver

The examples below are based on the C62x device, if you have a different device use the corresponding values in the above table.

In BIOS ensure that SRIOV is enabled and either:

  • Disable VT-d or
  • Enable VT-d and set "intel_iommu=on iommu=pt" in the grub file.

Check that the QAT driver is loaded on your system, by executing:

lsmod | grep qa

You should see the kernel module for your device listed, e.g.:

qat_c62x               5626  0
intel_qat              82336  1 qat_c62x

Next, you need to expose the Virtual Functions (VFs) using the sysfs file system.

First find the BDFs (Bus-Device-Function) of the physical functions (PFs) of your device, e.g.:

lspci -d : 37c8

You should see output similar to:

1a:00.0 Co-processor: Intel Corporation Device 37c8
3d:00.0 Co-processor: Intel Corporation Device 37c8
3f:00.0 Co-processor: Intel Corporation Device 37c8

Enable the VFs for each PF by echoing the number of VFs per PF to the pci driver:

echo 16 > /sys/bus/pci/drivers/c6xx/0000:1a:00.0/sriov_numvfs
echo 16 > /sys/bus/pci/drivers/c6xx/0000:3d:00.0/sriov_numvfs
echo 16 > /sys/bus/pci/drivers/c6xx/0000:3f:00.0/sriov_numvfs

Check that the VFs are available for use. For example lspci -d:37c9 should list 48 VF devices available for a C62x device.

To complete the installation follow the instructions in Binding the available VFs to the DPDK UIO driver.

Note

If the QAT kernel modules are not loaded and you see an error like Failed to load MMP firmware qat_895xcc_mmp.bin in kernel logs, this may be as a result of not using a distribution, but just updating the kernel directly.

Download firmware from the kernel firmware repo.

Copy qat binaries to /lib/firmware:

cp qat_895xcc.bin /lib/firmware
cp qat_895xcc_mmp.bin /lib/firmware

Change to your linux source root directory and start the qat kernel modules:

insmod ./drivers/crypto/qat/qat_common/intel_qat.ko
insmod ./drivers/crypto/qat/qat_dh895xcc/qat_dh895xcc.ko

Note

If you see the following warning in /var/log/messages it can be ignored: IOMMU should be enabled for SR-IOV to work correctly.

Installation using 01.org QAT driver

Download the latest QuickAssist Technology Driver from 01.org. Consult the Getting Started Guide at the same URL for further information.

The steps below assume you are:

  • Building on a platform with one DH895xCC device.
  • Using package qatmux.l.2.3.0-34.tgz.
  • On Fedora21 kernel 3.17.4-301.fc21.x86_64.

In the BIOS ensure that SRIOV is enabled and VT-d is disabled.

Uninstall any existing QAT driver, for example by running:

  • ./installer.sh uninstall in the directory where originally installed.
  • or rmmod qat_dh895xcc; rmmod intel_qat.

Build and install the SRIOV-enabled QAT driver:

mkdir /QAT
cd /QAT

# Copy qatmux.l.2.3.0-34.tgz to this location
tar zxof qatmux.l.2.3.0-34.tgz

export ICP_WITHOUT_IOMMU=1
./installer.sh install QAT1.6 host

You can use cat /proc/icp_dh895xcc_dev0/version to confirm the driver is correctly installed. You can use lspci -d:443 to confirm the of the 32 VF devices available per DH895xCC device.

To complete the installation - follow instructions in Binding the available VFs to the DPDK UIO driver.

Note

If using a later kernel and the build fails with an error relating to strict_stroul not being available apply the following patch:

/QAT/QAT1.6/quickassist/utilities/downloader/Target_CoreLibs/uclo/include/linux/uclo_platform.h
+ #if LINUX_VERSION_CODE >= KERNEL_VERSION(3,18,5)
+ #define STR_TO_64(str, base, num, endPtr) {endPtr=NULL; if (kstrtoul((str), (base), (num))) printk("Error strtoull convert %s\n", str); }
+ #else
#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,38)
#define STR_TO_64(str, base, num, endPtr) {endPtr=NULL; if (strict_strtoull((str), (base), (num))) printk("Error strtoull convert %s\n", str); }
#else
#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,25)
#define STR_TO_64(str, base, num, endPtr) {endPtr=NULL; strict_strtoll((str), (base), (num));}
#else
#define STR_TO_64(str, base, num, endPtr)                                 \
     do {                                                               \
           if (str[0] == '-')                                           \
           {                                                            \
                *(num) = -(simple_strtoull((str+1), &(endPtr), (base))); \
           }else {                                                      \
                *(num) = simple_strtoull((str), &(endPtr), (base));      \
           }                                                            \
     } while(0)
+ #endif
#endif
#endif

Note

If the build fails due to missing header files you may need to do following:

sudo yum install zlib-devel
sudo yum install openssl-devel

Note

If the build or install fails due to mismatching kernel sources you may need to do the following:

sudo yum install kernel-headers-`uname -r`
sudo yum install kernel-src-`uname -r`
sudo yum install kernel-devel-`uname -r`

Binding the available VFs to the DPDK UIO driver

Unbind the VFs from the stock driver so they can be bound to the uio driver.

For an Intel(R) QuickAssist Technology DH895xCC device

The unbind command below assumes BDFs of 03:01.00-03:04.07, if your VFs are different adjust the unbind command below:

for device in $(seq 1 4); do \
    for fn in $(seq 0 7); do \
        echo -n 0000:03:0${device}.${fn} > \
        /sys/bus/pci/devices/0000\:03\:0${device}.${fn}/driver/unbind; \
    done; \
done
For an Intel(R) QuickAssist Technology C62x device

The unbind command below assumes BDFs of 1a:01.00-1a:02.07, 3d:01.00-3d:02.07 and 3f:01.00-3f:02.07, if your VFs are different adjust the unbind command below:

for device in $(seq 1 2); do \
    for fn in $(seq 0 7); do \
        echo -n 0000:1a:0${device}.${fn} > \
        /sys/bus/pci/devices/0000\:1a\:0${device}.${fn}/driver/unbind; \

        echo -n 0000:3d:0${device}.${fn} > \
        /sys/bus/pci/devices/0000\:3d\:0${device}.${fn}/driver/unbind; \

        echo -n 0000:3f:0${device}.${fn} > \
        /sys/bus/pci/devices/0000\:3f\:0${device}.${fn}/driver/unbind; \
    done; \
done
For Intel(R) QuickAssist Technology C3xxx or D15xx device

The unbind command below assumes BDFs of 01:01.00-01:02.07, if your VFs are different adjust the unbind command below:

for device in $(seq 1 2); do \
    for fn in $(seq 0 7); do \
        echo -n 0000:01:0${device}.${fn} > \
        /sys/bus/pci/devices/0000\:01\:0${device}.${fn}/driver/unbind; \
    done; \
done
Bind to the DPDK uio driver

Install the DPDK igb_uio driver, bind the VF PCI Device id to it and use lspci to confirm the VF devices are now in use by igb_uio kernel driver, e.g. for the C62x device:

cd to the top-level DPDK directory
modprobe uio
insmod ./build/kmod/igb_uio.ko
echo "8086 37c9" > /sys/bus/pci/drivers/igb_uio/new_id
lspci -vvd:37c9

Another way to bind the VFs to the DPDK UIO driver is by using the dpdk-devbind.py script:

cd to the top-level DPDK directory
./usertools/dpdk-devbind.py -b igb_uio 0000:03:01.1

Extra notes on KASUMI F9

When using KASUMI F9 authentication algorithm, the input buffer must be constructed according to the 3GPP KASUMI specifications (section 4.4, page 13): http://cryptome.org/3gpp/35201-900.pdf. Input buffer has to have COUNT (4 bytes), FRESH (4 bytes), MESSAGE and DIRECTION (1 bit) concatenated. After the DIRECTION bit, a single ‘1’ bit is appended, followed by between 0 and 7 ‘0’ bits, so that the total length of the buffer is multiple of 8 bits. Note that the actual message can be any length, specified in bits.

Once this buffer is passed this way, when creating the crypto operation, length of data to authenticate (op.sym.auth.data.length) must be the length of all the items described above, including the padding at the end. Also, offset of data to authenticate (op.sym.auth.data.offset) must be such that points at the start of the COUNT bytes.

ZUC Crypto Poll Mode Driver

The ZUC PMD (librte_pmd_zuc) provides poll mode crypto driver support for utilizing Intel Libsso library, which implements F8 and F9 functions for ZUC EEA3 cipher and EIA3 hash algorithms.

Features

ZUC PMD has support for:

Cipher algorithm:

  • RTE_CRYPTO_CIPHER_ZUC_EEA3

Authentication algorithm:

  • RTE_CRYPTO_AUTH_ZUC_EIA3

Limitations

  • Chained mbufs are not supported.
  • ZUC (EIA3) supported only if hash offset field is byte-aligned.
  • ZUC (EEA3) supported only if cipher length, cipher offset fields are byte-aligned.
  • ZUC PMD cannot be built as a shared library, due to limitations in in the underlying library.

Installation

To build DPDK with the ZUC_PMD the user is required to download the export controlled libsso_zuc library, by requesting it from https://networkbuilders.intel.com/network-technologies/dpdk. Once approval has been granted, the user needs to log in https://networkbuilders.intel.com/dpdklogin and click on “ZUC Library” link, to download the library. After downloading the library, the user needs to unpack and compile it on their system before building DPDK:

make

Initialization

In order to enable this virtual crypto PMD, user must:

  • Export the environmental variable LIBSSO_ZUC_PATH with the path where the library was extracted (zuc folder).
  • Export the environmental variable LD_LIBRARY_PATH with the path where the built libsso library is (LIBSSO_ZUC_PATH/build).
  • Build the LIBSSO_ZUC library (explained in Installation section).
  • Build DPDK as follows:
make config T=x86_64-native-linuxapp-gcc
sed -i 's,\(CONFIG_RTE_LIBRTE_PMD_ZUC\)=n,\1=y,' build/.config
make

To use the PMD in an application, user must:

  • Call rte_vdev_init(“crypto_zuc”) within the application.
  • Use –vdev=”crypto_zuc” in the EAL options, which will call rte_vdev_init() internally.

The following parameters (all optional) can be provided in the previous two calls:

  • socket_id: Specify the socket where the memory for the device is going to be allocated (by default, socket_id will be the socket where the core that is creating the PMD is running on).
  • max_nb_queue_pairs: Specify the maximum number of queue pairs in the device (8 by default).
  • max_nb_sessions: Specify the maximum number of sessions that can be created (2048 by default).

Example:

./l2fwd-crypto -l 1 -n 4 --vdev="crypto_zuc,socket_id=0,max_nb_sessions=128" \
-- -p 1 --cdev SW --chain CIPHER_ONLY --cipher_algo "zuc-eea3"

Event Device Drivers

The following are a list of event device PMDs, which can be used from an application trough the eventdev API.

NXP DPAA2 Eventdev Driver

The dpaa2 eventdev is an implementation of the eventdev API, that provides a wide range of the eventdev features. The eventdev relies on a dpaa2 hw to perform event scheduling.

More information can be found at NXP Official Website.

Features

The DPAA2 EVENTDEV implements many features in the eventdev API;

  • Hardware based event scheduler
  • 8 event ports
  • 8 event queues
  • Parallel flows
  • Atomic flows

Supported DPAA2 SoCs

  • LS2080A/LS2040A
  • LS2084A/LS2044A
  • LS2088A/LS2048A
  • LS1088A/LS1048A

Prerequisites

There are three main pre-requisities for executing DPAA2 EVENTDEV on a DPAA2 compatible board:

  1. ARM 64 Tool Chain

    For example, the *aarch64* Linaro Toolchain.

  2. Linux Kernel

    It can be obtained from NXP’s Github hosting.

  3. Rootfile system

    Any aarch64 supporting filesystem can be used. For example, Ubuntu 15.10 (Wily) or 16.04 LTS (Xenial) userland which can be obtained from here.

As an alternative method, DPAA2 EVENTDEV can also be executed using images provided as part of SDK from NXP. The SDK includes all the above prerequisites necessary to bring up a DPAA2 board.

The following dependencies are not part of DPDK and must be installed separately:

  • NXP Linux SDK

    NXP Linux software development kit (SDK) includes support for family of QorIQ® ARM-Architecture-based system on chip (SoC) processors and corresponding boards.

    It includes the Linux board support packages (BSPs) for NXP SoCs, a fully operational tool chain, kernel and board specific modules.

    SDK and related information can be obtained from: NXP QorIQ SDK.

  • DPDK Extra Scripts

    DPAA2 based resources can be configured easily with the help of ready scripts as provided in the DPDK Extra repository.

    DPDK Extras Scripts.

Currently supported by DPDK:

  • NXP SDK 2.0+.
  • MC Firmware version 10.0.0 and higher.
  • Supported architectures: arm64 LE.
  • Follow the DPDK Getting Started Guide for Linux to setup the basic DPDK environment.

Note

Some part of fslmc bus code (mc flib - object library) routines are dual licensed (BSD & GPLv2).

Pre-Installation Configuration

Config File Options

The following options can be modified in the config file. Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_LIBRTE_PMD_DPAA2_EVENTDEV (default y)

    Toggle compilation of the lrte_pmd_dpaa2_event driver.

  • CONFIG_RTE_LIBRTE_PMD_DPAA2_EVENTDEV_DEBUG (default n)

    Toggle display of generic debugging messages

Driver Compilation

To compile the DPAA2 EVENTDEV PMD for Linux arm64 gcc target, run the following make command:

cd <DPDK-source-directory>
make config T=arm64-dpaa2-linuxapp-gcc install

Initialization

The dpaa2 eventdev is exposed as a vdev device which consists of a set of dpcon devices and dpci devices. On EAL initialization, dpcon and dpci devices will be probed and then vdev device can be created from the application code by

  • Invoking rte_vdev_init("event_dpaa2") from the application
  • Using --vdev="event_dpaa2" in the EAL options, which will call rte_vdev_init() internally

Example:

./your_eventdev_application --vdev="event_dpaa2"

Limitations

Platform Requirement

DPAA2 drivers for DPDK can only work on NXP SoCs as listed in the Supported DPAA2 SoCs.

Port-core binding

DPAA2 EVENTDEV driver requires event port ‘x’ to be used on core ‘x’.

Software Eventdev Poll Mode Driver

The software eventdev is an implementation of the eventdev API, that provides a wide range of the eventdev features. The eventdev relies on a CPU core to perform event scheduling. This PMD can use the service core library to run the scheduling function, allowing an application to utilize the power of service cores to multiplex other work on the same core if required.

Features

The software eventdev implements many features in the eventdev API;

Queues
  • Atomic
  • Ordered
  • Parallel
  • Single-Link
Ports
  • Load balanced (for Atomic, Ordered, Parallel queues)
  • Single Link (for single-link queues)
Event Priorities
  • Each event has a priority, which can be used to provide basic QoS

Configuration and Options

The software eventdev is a vdev device, and as such can be created from the application code, or from the EAL command line:

  • Call rte_vdev_init("event_sw0") from the application
  • Use --vdev="event_sw0" in the EAL options, which will call rte_vdev_init() internally

Example:

./your_eventdev_application --vdev="event_sw0"
Scheduling Quanta

The scheduling quanta sets the number of events that the device attempts to schedule in a single schedule call performed by the service core. Note that is a hint only, and that fewer or more events may be scheduled in a given iteration.

The scheduling quanta can be set using a string argument to the vdev create call:

--vdev="event_sw0,sched_quanta=64"
Credit Quanta

The credit quanta is the number of credits that a port will fetch at a time from the instance’s credit pool. Higher numbers will cause less overhead in the atomic credit fetch code, however it also reduces the overall number of credits in the system faster. A balanced number (eg 32) ensures that only small numbers of credits are pre-allocated at a time, while also mitigating performance impact of the atomics.

Experimentation with higher values may provide minor performance improvements, at the cost of the whole system having less credits. On the other hand, reducing the quanta may cause measurable performance impact but provide the system with a higher number of credits at all times.

A value of 32 seems a good balance however your specific application may benefit from a higher or reduced quanta size, experimentation is required to verify possible gains.

--vdev="event_sw0,credit_quanta=64"

Limitations

The software eventdev implementation has a few limitations. The reason for these limitations is usually that the performance impact of supporting the feature would be significant.

“All Types” Queues

The software eventdev does not support creating queues that handle all types of traffic. An eventdev with this capability allows enqueueing Atomic, Ordered and Parallel traffic to the same queue, but scheduling each of them appropriately.

The reason to not allow Atomic, Ordered and Parallel event types in the same queue is that it causes excessive branching in the code to enqueue packets to the queue, causing a significant performance impact.

The RTE_EVENT_DEV_CAP_QUEUE_ALL_TYPES flag is not set in the event_dev_cap field of the rte_event_dev_info struct for the software eventdev.

Distributed Scheduler

The software eventdev is a centralized scheduler, requiring a service core to perform the required event distribution. This is not really a limitation but rather a design decision.

The RTE_EVENT_DEV_CAP_DISTRIBUTED_SCHED flag is not set in the event_dev_cap field of the rte_event_dev_info struct for the software eventdev.

Dequeue Timeout

The eventdev API supports a timeout when dequeuing packets using the rte_event_dequeue_burst function. This allows a core to wait for an event to arrive, or until timeout number of ticks have passed. Timeout ticks is not supported by the software eventdev for performance reasons.

OCTEONTX SSOVF Eventdev Driver

The OCTEONTX SSOVF PMD (librte_pmd_octeontx_ssovf) provides poll mode eventdev driver support for the inbuilt event device found in the Cavium OCTEONTX SoC family as well as their virtual functions (VF) in SR-IOV context.

More information can be found at Cavium, Inc Official Website.

Features

Features of the OCTEONTX SSOVF PMD are:

  • 64 Event queues
  • 32 Event ports
  • HW event scheduler
  • Supports 1M flows per event queue
  • Flow based event pipelining
  • Flow pinning support in flow based event pipelining
  • Queue based event pipelining
  • Supports ATOMIC, ORDERED, PARALLEL schedule types per flow
  • Event scheduling QoS based on event queue priority
  • Open system with configurable amount of outstanding events
  • HW accelerated dequeue timeout support to enable power management
  • SR-IOV VF

Supported OCTEONTX SoCs

  • CN83xx

Prerequisites

See :doc: ../platform/octeontx for setup information.

Pre-Installation Configuration

Config File Options

The following options can be modified in the config file. Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_LIBRTE_PMD_OCTEONTX_SSOVF (default y)

    Toggle compilation of the librte_pmd_octeontx_ssovf driver.

  • CONFIG_RTE_LIBRTE_PMD_OCTEONTX_SSOVF_DEBUG (default n)

    Toggle display of generic debugging messages

Driver Compilation

To compile the OCTEONTX SSOVF PMD for Linux arm64 gcc target, run the following make command:

cd <DPDK-source-directory>
make config T=arm64-thunderx-linuxapp-gcc install

Initialization

The octeontx eventdev is exposed as a vdev device which consists of a set of SSO group and work-slot PCIe VF devices. On EAL initialization, SSO PCIe VF devices will be probed and then the vdev device can be created from the application code, or from the EAL command line based on the number of probed/bound SSO PCIe VF device to DPDK by

  • Invoking rte_vdev_init("event_octeontx") from the application
  • Using --vdev="event_octeontx" in the EAL options, which will call rte_vdev_init() internally

Example:

./your_eventdev_application --vdev="event_octeontx"

Limitations

Burst mode support

Burst mode is not supported. Dequeue and Enqueue functions accepts only single event at a time.

Rx adapter support

When eth_octeontx is used as Rx adapter event schedule type RTE_SCHED_TYPE_PARALLEL is not supported.

Mempool Device Driver

The following are a list of mempool PMDs, which can be used from an application through the mempool API.

OCTEONTX FPAVF Mempool Driver

The OCTEONTX FPAVF PMD (librte_mempool_octeontx) is a mempool driver for offload mempool device found in Cavium OCTEONTX SoC family.

More information can be found at Cavium, Inc Official Website.

Features

Features of the OCTEONTX FPAVF PMD are:

  • 32 SR-IOV Virtual functions
  • 32 Pools
  • HW mempool manager

Supported OCTEONTX SoCs

  • CN83xx

Prerequisites

See :doc: ../platform/octeontx.rst for setup information.

Pre-Installation Configuration

Config File Options

The following options can be modified in the config file. Please note that enabling debugging options may affect system performance.

  • CONFIG_RTE_MBUF_DEFAULT_MEMPOOL_OPS ( set to octeontx_fpavf)

    Set default mempool ops to octeontx_fpavf.

  • CONFIG_RTE_LIBRTE_OCTEONTX_MEMPOOL (default y)

    Toggle compilation of the librte_mempool_octeontx driver.

  • CONFIG_RTE_LIBRTE_OCTEONTX_MEMPOOL_DEBUG (default n)

    Toggle display of generic debugging messages

Driver Compilation

To compile the OCTEONTX FPAVF MEMPOOL PMD for Linux arm64 gcc target, run the following make command:

cd <DPDK-source-directory>
make config T=arm64-thunderx-linuxapp-gcc test-build

Initialization

The octeontx fpavf mempool initialization similar to other mempool drivers like ring. However user need to pass –base-virtaddr as command line input to application example test_mempool.c application.

Example:

./build/app/test -c 0xf --base-virtaddr=0x100000000000 \
                    --mbuf-pool-ops-name="octeontx_fpavf"

Platform Specific Guides

The following are platform specific guides and setup information.

OCTEONTX Board Support Package

This doc has information about steps to setup octeontx platform and information about common offload hw block drivers of Cavium OCTEONTX SoC family.

More information about SoC can be found at Cavium, Inc Official Website.

Common Offload HW Block Drivers

  1. Eventdev Driver See :doc: ../eventdevs/octeontx.rst for octeontx ssovf eventdev driver information.
  2. Mempool Driver See :doc: ../mempool/octeontx.rst for octeontx fpavf mempool driver information.

Steps To Setup Platform

There are three main pre-prerequisites for setting up Platform drivers on OCTEONTX compatible board:

  1. OCTEONTX Linux kernel PF driver for Network acceleration HW blocks

    The OCTEONTX Linux kernel drivers (includes the required PF driver for the Platform drivers) are available on Github at octeontx-kmod along with build, install and dpdk usage instructions.

  2. ARM64 Tool Chain

    For example, the aarch64 Linaro Toolchain, which can be obtained from here.

  3. Rootfile system

    Any aarch64 supporting filesystem can be used. For example, Ubuntu 15.10 (Wily) or 16.04 LTS (Xenial) userland which can be obtained from http://cdimage.ubuntu.com/ubuntu-base/releases/16.04/release/ubuntu-base-16.04.1-base-arm64.tar.gz.

    As an alternative method, Platform drivers can also be executed using images provided as part of SDK from Cavium. The SDK includes all the above prerequisites necessary to bring up a OCTEONTX board.

    SDK and related information can be obtained from: Cavium support site.

  • Follow the DPDK :doc: ../linux_gsg/index.rst to setup the basic DPDK environment.

Contributor’s Guidelines

DPDK Coding Style

Description

This document specifies the preferred style for source files in the DPDK source tree. It is based on the Linux Kernel coding guidelines and the FreeBSD 7.2 Kernel Developer’s Manual (see man style(9)), but was heavily modified for the needs of the DPDK.

General Guidelines

The rules and guidelines given in this document cannot cover every situation, so the following general guidelines should be used as a fallback:

  • The code style should be consistent within each individual file.
  • In the case of creating new files, the style should be consistent within each file in a given directory or module.
  • The primary reason for coding standards is to increase code readability and comprehensibility, therefore always use whatever option will make the code easiest to read.

Line length is recommended to be not more than 80 characters, including comments. [Tab stop size should be assumed to be 8-characters wide].

Note

The above is recommendation, and not a hard limit. However, it is expected that the recommendations should be followed in all but the rarest situations.

C Comment Style

Usual Comments

These comments should be used in normal cases. To document a public API, a doxygen-like format must be used: refer to Doxygen Guidelines.

/*
 * VERY important single-line comments look like this.
 */

/* Most single-line comments look like this. */

/*
 * Multi-line comments look like this.  Make them real sentences. Fill
 * them so they look like real paragraphs.
 */
License Header

Each file should begin with a special comment containing the appropriate copyright and license for the file. Generally this is the BSD License, except for code for Linux Kernel modules. After any copyright header, a blank line should be left before any other contents, e.g. include statements in a C file.

C Preprocessor Directives

Header Includes

In DPDK sources, the include files should be ordered as following:

  1. libc includes (system includes first)
  2. DPDK EAL includes
  3. DPDK misc libraries includes
  4. application-specific includes

Include files from the local application directory are included using quotes, while includes from other paths are included using angle brackets: “<>”.

Example:

#include <stdio.h>
#include <stdlib.h>

#include <rte_eal.h>

#include <rte_ring.h>
#include <rte_mempool.h>

#include "application.h"
Header File Guards

Headers should be protected against multiple inclusion with the usual:

#ifndef _FILE_H_
#define _FILE_H_

/* Code */

#endif /* _FILE_H_ */
Macros

Do not #define or declare names except with the standard DPDK prefix: RTE_. This is to ensure there are no collisions with definitions in the application itself.

The names of “unsafe” macros (ones that have side effects), and the names of macros for manifest constants, are all in uppercase.

The expansions of expression-like macros are either a single token or have outer parentheses. If a macro is an inline expansion of a function, the function name is all in lowercase and the macro has the same name all in uppercase. If the macro encapsulates a compound statement, enclose it in a do-while loop, so that it can be used safely in if statements. Any final statement-terminating semicolon should be supplied by the macro invocation rather than the macro, to make parsing easier for pretty-printers and editors.

For example:

#define MACRO(x, y) do {                                        \
        variable = (x) + (y);                                   \
        (y) += 2;                                               \
} while(0)

Note

Wherever possible, enums and inline functions should be preferred to macros, since they provide additional degrees of type-safety and can allow compilers to emit extra warnings about unsafe code.

Conditional Compilation
  • When code is conditionally compiled using #ifdef or #if, a comment may be added following the matching #endif or #else to permit the reader to easily discern where conditionally compiled code regions end.
  • This comment should be used only for (subjectively) long regions, regions greater than 20 lines, or where a series of nested #ifdef‘s may be confusing to the reader. Exceptions may be made for cases where code is conditionally not compiled for the purposes of lint(1), or other tools, even though the uncompiled region may be small.
  • The comment should be separated from the #endif or #else by a single space.
  • For short conditionally compiled regions, a closing comment should not be used.
  • The comment for #endif should match the expression used in the corresponding #if or #ifdef.
  • The comment for #else and #elif should match the inverse of the expression(s) used in the preceding #if and/or #elif statements.
  • In the comments, the subexpression defined(FOO) is abbreviated as “FOO”. For the purposes of comments, #ifndef FOO is treated as #if !defined(FOO).
#ifdef KTRACE
#include <sys/ktrace.h>
#endif

#ifdef COMPAT_43
/* A large region here, or other conditional code. */
#else /* !COMPAT_43 */
/* Or here. */
#endif /* COMPAT_43 */

#ifndef COMPAT_43
/* Yet another large region here, or other conditional code. */
#else /* COMPAT_43 */
/* Or here. */
#endif /* !COMPAT_43 */

Note

Conditional compilation should be used only when absolutely necessary, as it increases the number of target binaries that need to be built and tested.

C Types

Integers

For fixed/minimum-size integer values, the project uses the form uintXX_t (from stdint.h) instead of older BSD-style integer identifiers of the form u_intXX_t.

Enumerations
  • Enumeration values are all uppercase.
enum enumtype { ONE, TWO } et;
  • Enum types should be used in preference to macros #defining a set of (sequential) values.
  • Enum types should be prefixed with rte_ and the elements by a suitable prefix [generally starting RTE_<enum>_ - where <enum> is a shortname for the enum type] to avoid namespace collisions.
Bitfields

The developer should group bitfields that are included in the same integer, as follows:

struct grehdr {
  uint16_t rec:3,
      srr:1,
      seq:1,
      key:1,
      routing:1,
      csum:1,
      version:3,
      reserved:4,
      ack:1;
/* ... */
}
Variable Declarations

In declarations, do not put any whitespace between asterisks and adjacent tokens, except for tokens that are identifiers related to types. (These identifiers are the names of basic types, type qualifiers, and typedef-names other than the one being declared.) Separate these identifiers from asterisks using a single space.

For example:

int *x;         /* no space after asterisk */
int * const x;  /* space after asterisk when using a type qualifier */
  • All externally-visible variables should have an rte_ prefix in the name to avoid namespace collisions.
  • Do not use uppercase letters - either in the form of ALL_UPPERCASE, or CamelCase - in variable names. Lower-case letters and underscores only.
Structure Declarations
  • In general, when declaring variables in new structures, declare them sorted by use, then by size (largest to smallest), and then in alphabetical order. Sorting by use means that commonly used variables are used together and that the structure layout makes logical sense. Ordering by size then ensures that as little padding is added to the structure as possible.
  • For existing structures, additions to structures should be added to the end so for backward compatibility reasons.
  • Each structure element gets its own line.
  • Try to make the structure readable by aligning the member names using spaces as shown below.
  • Names following extremely long types, which therefore cannot be easily aligned with the rest, should be separated by a single space.
struct foo {
        struct foo      *next;          /* List of active foo. */
        struct mumble   amumble;        /* Comment for mumble. */
        int             bar;            /* Try to align the comments. */
        struct verylongtypename *baz;   /* Won't fit with other members */
};
  • Major structures should be declared at the top of the file in which they are used, or in separate header files if they are used in multiple source files.
  • Use of the structures should be by separate variable declarations and those declarations must be extern if they are declared in a header file.
  • Externally visible structure definitions should have the structure name prefixed by rte_ to avoid namespace collisions.
Queues

Use queue(3) macros rather than rolling your own lists, whenever possible. Thus, the previous example would be better written:

#include <sys/queue.h>

struct foo {
        LIST_ENTRY(foo) link;      /* Use queue macros for foo lists. */
        struct mumble   amumble;   /* Comment for mumble. */
        int             bar;       /* Try to align the comments. */
        struct verylongtypename *baz;   /* Won't fit with other members */
};
LIST_HEAD(, foo) foohead;          /* Head of global foo list. */

DPDK also provides an optimized way to store elements in lockless rings. This should be used in all data-path code, when there are several consumer and/or producers to avoid locking for concurrent access.

Typedefs

Avoid using typedefs for structure types.

For example, use:

struct my_struct_type {
/* ... */
};

struct my_struct_type my_var;

rather than:

typedef struct my_struct_type {
/* ... */
} my_struct_type;

my_struct_type my_var

Typedefs are problematic because they do not properly hide their underlying type; for example, you need to know if the typedef is the structure itself, as shown above, or a pointer to the structure. In addition, they must be declared exactly once, whereas an incomplete structure type can be mentioned as many times as necessary. Typedefs are difficult to use in stand-alone header files. The header that defines the typedef must be included before the header that uses it, or by the header that uses it (which causes namespace pollution), or there must be a back-door mechanism for obtaining the typedef.

Note that #defines used instead of typedefs also are problematic (since they do not propagate the pointer type correctly due to direct text replacement). For example, #define pint int * does not work as expected, while typedef int *pint does work. As stated when discussing macros, typedefs should be preferred to macros in cases like this.

When convention requires a typedef; make its name match the struct tag. Avoid typedefs ending in _t, except as specified in Standard C or by POSIX.

Note

It is recommended to use typedefs to define function pointer types, for reasons of code readability. This is especially true when the function type is used as a parameter to another function.

For example:

/**
 * Definition of a remote launch function.
 */
typedef int (lcore_function_t)(void *);

/* launch a function of lcore_function_t type */
int rte_eal_remote_launch(lcore_function_t *f, void *arg, unsigned slave_id);

C Indentation

General
  • Indentation is a hard tab, that is, a tab character, not a sequence of spaces,

Note

Global whitespace rule in DPDK, use tabs for indentation, spaces for alignment.

  • Do not put any spaces before a tab for indentation.
  • If you have to wrap a long statement, put the operator at the end of the line, and indent again.
  • For control statements (if, while, etc.), continuation it is recommended that the next line be indented by two tabs, rather than one, to prevent confusion as to whether the second line of the control statement forms part of the statement body or not. Alternatively, the line continuation may use additional spaces to line up to an appropriately point on the preceding line, for example, to align to an opening brace.

Note

As with all style guidelines, code should match style already in use in an existing file.

while (really_long_variable_name_1 == really_long_variable_name_2 &&
    var3 == var4){  /* confusing to read as */
    x = y + z;      /* control stmt body lines up with second line of */
    a = b + c;      /* control statement itself if single indent used */
}

if (really_long_variable_name_1 == really_long_variable_name_2 &&
        var3 == var4){  /* two tabs used */
    x = y + z;          /* statement body no longer lines up */
    a = b + c;
}

z = a + really + long + statement + that + needs +
        two + lines + gets + indented + on + the +
        second + and + subsequent + lines;
  • Do not add whitespace at the end of a line.
  • Do not add whitespace or a blank line at the end of a file.
Control Statements and Loops
  • Include a space after keywords (if, while, for, return, switch).
  • Do not use braces ({ and }) for control statements with zero or just a single statement, unless that statement is more than a single line in which case the braces are permitted.
for (p = buf; *p != '\0'; ++p)
        ;       /* nothing */
for (;;)
        stmt;
for (;;) {
        z = a + really + long + statement + that + needs +
                two + lines + gets + indented + on + the +
                second + and + subsequent + lines;
}
for (;;) {
        if (cond)
                stmt;
}
if (val != NULL)
        val = realloc(val, newsize);
  • Parts of a for loop may be left empty.
for (; cnt < 15; cnt++) {
        stmt1;
        stmt2;
}
  • Closing and opening braces go on the same line as the else keyword.
  • Braces that are not necessary should be left out.
if (test)
        stmt;
else if (bar) {
        stmt;
        stmt;
} else
        stmt;
Function Calls
  • Do not use spaces after function names.
  • Commas should have a space after them.
  • No spaces after ( or [ or preceding the ] or ) characters.
error = function(a1, a2);
if (error != 0)
        exit(error);
Operators
  • Unary operators do not require spaces, binary operators do.
  • Do not use parentheses unless they are required for precedence or unless the statement is confusing without them. However, remember that other people may be more easily confused than you.
Exit

Exits should be 0 on success, or 1 on failure.

        exit(0);        /*
                         * Avoid obvious comments such as
                         * "Exit 0 on success."
                         */
}
Local Variables
  • Variables should be declared at the start of a block of code rather than in the middle. The exception to this is when the variable is const in which case the declaration must be at the point of first use/assignment.
  • When declaring variables in functions, multiple variables per line are OK. However, if multiple declarations would cause the line to exceed a reasonable line length, begin a new set of declarations on the next line rather than using a line continuation.
  • Be careful to not obfuscate the code by initializing variables in the declarations, only the last variable on a line should be initialized. If multiple variables are to be initialized when defined, put one per line.
  • Do not use function calls in initializers, except for const variables.
int i = 0, j = 0, k = 0;  /* bad, too many initializer */

char a = 0;        /* OK, one variable per line with initializer */
char b = 0;

float x, y = 0.0;  /* OK, only last variable has initializer */
Casts and sizeof
  • Casts and sizeof statements are not followed by a space.
  • Always write sizeof statements with parenthesis. The redundant parenthesis rules do not apply to sizeof(var) instances.

C Function Definition, Declaration and Use

Prototypes
  • It is recommended (and generally required by the compiler) that all non-static functions are prototyped somewhere.
  • Functions local to one source module should be declared static, and should not be prototyped unless absolutely necessary.
  • Functions used from other parts of code (external API) must be prototyped in the relevant include file.
  • Function prototypes should be listed in a logical order, preferably alphabetical unless there is a compelling reason to use a different ordering.
  • Functions that are used locally in more than one module go into a separate header file, for example, “extern.h”.
  • Do not use the __P macro.
  • Functions that are part of an external API should be documented using Doxygen-like comments above declarations. See Doxygen Guidelines for details.
  • Functions that are part of the external API must have an rte_ prefix on the function name.
  • Do not use uppercase letters - either in the form of ALL_UPPERCASE, or CamelCase - in function names. Lower-case letters and underscores only.
  • When prototyping functions, associate names with parameter types, for example:
void function1(int fd); /* good */
void function2(int);    /* bad */
  • Short function prototypes should be contained on a single line. Longer prototypes, e.g. those with many parameters, can be split across multiple lines. The second and subsequent lines should be further indented as for line statement continuations as described in the previous section.
static char *function1(int _arg, const char *_arg2,
       struct foo *_arg3,
       struct bar *_arg4,
       struct baz *_arg5);
static void usage(void);

Note

Unlike function definitions, the function prototypes do not need to place the function return type on a separate line.

Definitions
  • The function type should be on a line by itself preceding the function.
  • The opening brace of the function body should be on a line by itself.
static char *
function(int a1, int a2, float fl, int a4)
{
  • Do not declare functions inside other functions. ANSI C states that such declarations have file scope regardless of the nesting of the declaration. Hiding file declarations in what appears to be a local scope is undesirable and will elicit complaints from a good compiler.
  • Old-style (K&R) function declaration should not be used, use ANSI function declarations instead as shown below.
  • Long argument lists should be wrapped as described above in the function prototypes section.
/*
 * All major routines should have a comment briefly describing what
 * they do. The comment before the "main" routine should describe
 * what the program does.
 */
int
main(int argc, char *argv[])
{
        char *ep;
        long num;
        int ch;

C Statement Style and Conventions

NULL Pointers
  • NULL is the preferred null pointer constant. Use NULL instead of (type *)0 or (type *)NULL, except where the compiler does not know the destination type e.g. for variadic args to a function.
  • Test pointers against NULL, for example, use:
if (p == NULL) /* Good, compare pointer to NULL */

if (!p) /* Bad, using ! on pointer */
  • Do not use ! for tests unless it is a boolean, for example, use:
if (*p == '\0') /* check character against (char)0 */
Return Value
  • Functions which create objects, or allocate memory, should return pointer types, and NULL on error. The error type should be indicated may setting the variable rte_errno appropriately.
  • Functions which work on bursts of packets, such as RX-like or TX-like functions, should return the number of packets handled.
  • Other functions returning int should generally behave like system calls: returning 0 on success and -1 on error, setting rte_errno to indicate the specific type of error.
  • Where already standard in a given library, the alternative error approach may be used where the negative value is not -1 but is instead -errno if relevant, for example, -EINVAL. Note, however, to allow consistency across functions returning integer or pointer types, the previous approach is preferred for any new libraries.
  • For functions where no error is possible, the function type should be void not int.
  • Routines returning void * should not have their return values cast to any pointer type. (Typecasting can prevent the compiler from warning about missing prototypes as any implicit definition of a function returns int, which, unlike void *, needs a typecast to assign to a pointer variable.)

Note

The above rule about not typecasting void * applies to malloc, as well as to DPDK functions.

  • Values in return statements should not be enclosed in parentheses.
Logging and Errors

In the DPDK environment, use the logging interface provided:

/* register log types for this application */
int my_logtype1 = rte_log_register("myapp.log1");
int my_logtype2 = rte_log_register("myapp.log2");

/* set global log level to INFO */
rte_log_set_global_level(RTE_LOG_INFO);

/* only display messages higher than NOTICE for log2 (default
 * is DEBUG) */
rte_log_set_level(my_logtype2, RTE_LOG_NOTICE);

/* enable all PMD logs (whose identifier string starts with "pmd") */
rte_log_set_level_regexp("pmd.*", RTE_LOG_DEBUG);

/* log in debug level */
rte_log_set_global_level(RTE_LOG_DEBUG);
RTE_LOG(DEBUG, my_logtype1, "this is is a debug level message\n");
RTE_LOG(INFO, my_logtype1, "this is is a info level message\n");
RTE_LOG(WARNING, my_logtype1, "this is is a warning level message\n");
RTE_LOG(WARNING, my_logtype2, "this is is a debug level message (not displayed)\n");

/* log in info level */
rte_log_set_global_level(RTE_LOG_INFO);
RTE_LOG(DEBUG, my_logtype1, "debug level message (not displayed)\n");
Branch Prediction
  • When a test is done in a critical zone (called often or in a data path) the code can use the likely() and unlikely() macros to indicate the expected, or preferred fast path. They are expanded as a compiler builtin and allow the developer to indicate if the branch is likely to be taken or not. Example:
#include <rte_branch_prediction.h>
if (likely(x > 1))
  do_stuff();

Note

The use of likely() and unlikely() should only be done in performance critical paths, and only when there is a clearly preferred path, or a measured performance increase gained from doing so. These macros should be avoided in non-performance-critical code.

Static Variables and Functions
  • All functions and variables that are local to a file must be declared as static because it can often help the compiler to do some optimizations (such as, inlining the code).
  • Functions that should be inlined should to be declared as static inline and can be defined in a .c or a .h file.

Note

Static functions defined in a header file must be declared as static inline in order to prevent compiler warnings about the function being unused.

Const Attribute

The const attribute should be used as often as possible when a variable is read-only.

Inline ASM in C code

The asm and volatile keywords do not have underscores. The AT&T syntax should be used. Input and output operands should be named to avoid confusion, as shown in the following example:

asm volatile("outb %[val], %[port]"
        : :
        [port] "dN" (port),
        [val] "a" (val));
Control Statements
  • Forever loops are done with for statements, not while statements.
  • Elements in a switch statement that cascade should have a FALLTHROUGH comment. For example:
switch (ch) {         /* Indent the switch. */
case 'a':             /* Don't indent the case. */
        aflag = 1;    /* Indent case body one tab. */
        /* FALLTHROUGH */
case 'b':
        bflag = 1;
        break;
case '?':
default:
        usage();
        /* NOTREACHED */
}

Python Code

All Python code should work with Python 2.7+ and 3.2+ and be compliant with PEP8 (Style Guide for Python Code).

The pep8 tool can be used for testing compliance with the guidelines.

Design

Environment or Architecture-specific Sources

In DPDK and DPDK applications, some code is specific to an architecture (i686, x86_64) or to an executive environment (bsdapp or linuxapp) and so on. As far as is possible, all such instances of architecture or env-specific code should be provided via standard APIs in the EAL.

By convention, a file is common if it is not located in a directory indicating that it is specific. For instance, a file located in a subdir of “x86_64” directory is specific to this architecture. A file located in a subdir of “linuxapp” is specific to this execution environment.

Note

Code in DPDK libraries and applications should be generic. The correct location for architecture or executive environment specific code is in the EAL.

When absolutely necessary, there are several ways to handle specific code:

  • Use a #ifdef with the CONFIG option in the C code. This can be done when the differences are small and they can be embedded in the same C file:

    #ifdef RTE_ARCH_I686
    toto();
    #else
    titi();
    #endif
    
  • Use the CONFIG option in the Makefile. This is done when the differences are more significant. In this case, the code is split into two separate files that are architecture or environment specific. This should only apply inside the EAL library.

Note

As in the linux kernel, the CONFIG_ prefix is not used in C code. This is only needed in Makefiles or shell scripts.

Per Architecture Sources

The following config options can be used:

  • CONFIG_RTE_ARCH is a string that contains the name of the architecture.
  • CONFIG_RTE_ARCH_I686, CONFIG_RTE_ARCH_X86_64, CONFIG_RTE_ARCH_X86_64_32 or CONFIG_RTE_ARCH_PPC_64 are defined only if we are building for those architectures.
Per Execution Environment Sources

The following config options can be used:

  • CONFIG_RTE_EXEC_ENV is a string that contains the name of the executive environment.
  • CONFIG_RTE_EXEC_ENV_BSDAPP or CONFIG_RTE_EXEC_ENV_LINUXAPP are defined only if we are building for this execution environment.

Library Statistics

Description

This document describes the guidelines for DPDK library-level statistics counter support. This includes guidelines for turning library statistics on and off and requirements for preventing ABI changes when implementing statistics.

Mechanism to allow the application to turn library statistics on and off

Each library that maintains statistics counters should provide a single build time flag that decides whether the statistics counter collection is enabled or not. This flag should be exposed as a variable within the DPDK configuration file. When this flag is set, all the counters supported by current library are collected for all the instances of every object type provided by the library. When this flag is cleared, none of the counters supported by the current library are collected for any instance of any object type provided by the library:

# DPDK file config/common_linuxapp, config/common_bsdapp, etc.
CONFIG_RTE_<LIBRARY_NAME>_STATS_COLLECT=y/n

The default value for this DPDK configuration file variable (either “yes” or “no”) is decided by each library.

Prevention of ABI changes due to library statistics support

The layout of data structures and prototype of functions that are part of the library API should not be affected by whether the collection of statistics counters is turned on or off for the current library. In practical terms, this means that space should always be allocated in the API data structures for statistics counters and the statistics related API functions are always built into the code, regardless of whether the statistics counter collection is turned on or off for the current library.

When the collection of statistics counters for the current library is turned off, the counters retrieved through the statistics related API functions should have a default value of zero.

Motivation to allow the application to turn library statistics on and off

It is highly recommended that each library provides statistics counters to allow an application to monitor the library-level run-time events. Typical counters are: number of packets received/dropped/transmitted, number of buffers allocated/freed, number of occurrences for specific events, etc.

However, the resources consumed for library-level statistics counter collection have to be spent out of the application budget and the counters collected by some libraries might not be relevant to the current application. In order to avoid any unwanted waste of resources and/or performance impacts, the application should decide at build time whether the collection of library-level statistics counters should be turned on or off for each library individually.

Library-level statistics counters can be relevant or not for specific applications:

  • For Application A, counters maintained by Library X are always relevant and the application needs to use them to implement certain features, such as traffic accounting, logging, application-level statistics, etc. In this case, the application requires that collection of statistics counters for Library X is always turned on.
  • For Application B, counters maintained by Library X are only useful during the application debug stage and are not relevant once debug phase is over. In this case, the application may decide to turn on the collection of Library X statistics counters during the debug phase and at a later stage turn them off.
  • For Application C, counters maintained by Library X are not relevant at all. It might be that the application maintains its own set of statistics counters that monitor a different set of run-time events (e.g. number of connection requests, number of active users, etc). It might also be that the application uses multiple libraries (Library X, Library Y, etc) and it is interested in the statistics counters of Library Y, but not in those of Library X. In this case, the application may decide to turn the collection of statistics counters off for Library X and on for Library Y.

The statistics collection consumes a certain amount of CPU resources (cycles, cache bandwidth, memory bandwidth, etc) that depends on:

  • Number of libraries used by the current application that have statistics counters collection turned on.
  • Number of statistics counters maintained by each library per object type instance (e.g. per port, table, pipeline, thread, etc).
  • Number of instances created for each object type supported by each library.
  • Complexity of the statistics logic collection for each counter: when only some occurrences of a specific event are valid, additional logic is typically needed to decide whether the current occurrence of the event should be counted or not. For example, in the event of packet reception, when only TCP packets with destination port within a certain range should be recorded, conditional branches are usually required. When processing a burst of packets that have been validated for header integrity, counting the number of bits set in a bitmask might be needed.

PF and VF Considerations

The primary goal of DPDK is to provide a userspace dataplane. Managing VFs from a PF driver is a control plane feature and developers should generally rely on the Linux Kernel for that.

Developers should work with the Linux Kernel community to get the required functionality upstream. PF functionality should only be added to DPDK for testing and prototyping purposes while the kernel work is ongoing. It should also be marked with an “EXPERIMENTAL” tag. If the functionality isn’t upstreamable then a case can be made to maintain the PF functionality in DPDK without the EXPERIMENTAL tag.

Managing ABI updates

Description

This document details some methods for handling ABI management in the DPDK. Note this document is not exhaustive, in that C library versioning is flexible allowing multiple methods to achieve various goals, but it will provide the user with some introductory methods

General Guidelines

  1. Whenever possible, ABI should be preserved
  2. Libraries or APIs marked in experimental state may change without constraint.
  3. New APIs will be marked as experimental for at least one release to allow any issues found by users of the new API to be fixed quickly
  4. The addition of symbols is generally not problematic
  5. The modification of symbols can generally be managed with versioning
  6. The removal of symbols generally is an ABI break and requires bumping of the LIBABIVER macro
  7. Updates to the minimum hardware requirements, which drop support for hardware which was previously supported, should be treated as an ABI change.

What is an ABI

An ABI (Application Binary Interface) is the set of runtime interfaces exposed by a library. It is similar to an API (Application Programming Interface) but is the result of compilation. It is also effectively cloned when applications link to dynamic libraries. That is to say when an application is compiled to link against dynamic libraries, it is assumed that the ABI remains constant between the time the application is compiled/linked, and the time that it runs. Therefore, in the case of dynamic linking, it is critical that an ABI is preserved, or (when modified), done in such a way that the application is unable to behave improperly or in an unexpected fashion.

The DPDK ABI policy

ABI versions are set at the time of major release labeling, and the ABI may change multiple times, without warning, between the last release label and the HEAD label of the git tree.

APIs marked as experimental are not considered part of the ABI and may change without warning at any time. Since changes to APIs are most likely immediately after their introduction, as users begin to take advantage of those new APIs and start finding issues with them, new DPDK APIs will be automatically marked as experimental to allow for a period of stabilization before they become part of a tracked ABI.

ABI versions, once released, are available until such time as their deprecation has been noted in the Release Notes for at least one major release cycle. For example consider the case where the ABI for DPDK 2.0 has been shipped and then a decision is made to modify it during the development of DPDK 2.1. The decision will be recorded in the Release Notes for the DPDK 2.1 release and the modification will be made available in the DPDK 2.2 release.

ABI versions may be deprecated in whole or in part as needed by a given update.

Some ABI changes may be too significant to reasonably maintain multiple versions. In those cases ABI’s may be updated without backward compatibility being provided. The requirements for doing so are:

  1. At least 3 acknowledgments of the need to do so must be made on the dpdk.org mailing list.
    • The acknowledgment of the maintainer of the component is mandatory, or if no maintainer is available for the component, the tree/sub-tree maintainer for that component must acknowledge the ABI change instead.
    • It is also recommended that acknowledgments from different “areas of interest” be sought for each deprecation, for example: from NIC vendors, CPU vendors, end-users, etc.
  2. The changes (including an alternative map file) must be gated with the RTE_NEXT_ABI option, and provided with a deprecation notice at the same time. It will become the default ABI in the next release.
  3. A full deprecation cycle, as explained above, must be made to offer downstream consumers sufficient warning of the change.
  4. At the beginning of the next release cycle, every RTE_NEXT_ABI conditions will be removed, the LIBABIVER variable in the makefile(s) where the ABI is changed will be incremented, and the map files will be updated.

Note that the above process for ABI deprecation should not be undertaken lightly. ABI stability is extremely important for downstream consumers of the DPDK, especially when distributed in shared object form. Every effort should be made to preserve the ABI whenever possible. The ABI should only be changed for significant reasons, such as performance enhancements. ABI breakage due to changes such as reorganizing public structure fields for aesthetic or readability purposes should be avoided.

Note

Updates to the minimum hardware requirements, which drop support for hardware which was previously supported, should be treated as an ABI change, and follow the relevant deprecation policy procedures as above: 3 acks and announcement at least one release in advance.

Examples of Deprecation Notices

The following are some examples of ABI deprecation notices which would be added to the Release Notes:

  • The Macro #RTE_FOO is deprecated and will be removed with version 2.0, to be replaced with the inline function rte_foo().
  • The function rte_mbuf_grok() has been updated to include a new parameter in version 2.0. Backwards compatibility will be maintained for this function until the release of version 2.1
  • The members of struct rte_foo have been reorganized in release 2.0 for performance reasons. Existing binary applications will have backwards compatibility in release 2.0, while newly built binaries will need to reference the new structure variant struct rte_foo2. Compatibility will be removed in release 2.2, and all applications will require updating and rebuilding to the new structure at that time, which will be renamed to the original struct rte_foo.
  • Significant ABI changes are planned for the librte_dostuff library. The upcoming release 2.0 will not contain these changes, but release 2.1 will, and no backwards compatibility is planned due to the extensive nature of these changes. Binaries using this library built prior to version 2.1 will require updating and recompilation.

Versioning Macros

When a symbol is exported from a library to provide an API, it also provides a calling convention (ABI) that is embodied in its name, return type and arguments. Occasionally that function may need to change to accommodate new functionality or behavior. When that occurs, it is desirable to allow for backward compatibility for a time with older binaries that are dynamically linked to the DPDK.

To support backward compatibility the lib/librte_compat/rte_compat.h header file provides macros to use when updating exported functions. These macros are used in conjunction with the rte_<library>_version.map file for a given library to allow multiple versions of a symbol to exist in a shared library so that older binaries need not be immediately recompiled.

The macros exported are:

  • VERSION_SYMBOL(b, e, n): Creates a symbol version table entry binding versioned symbol b@DPDK_n to the internal function b_e.
  • BIND_DEFAULT_SYMBOL(b, e, n): Creates a symbol version entry instructing the linker to bind references to symbol b to the internal symbol b_e.
  • MAP_STATIC_SYMBOL(f, p): Declare the prototype f, and map it to the fully qualified function p, so that if a symbol becomes versioned, it can still be mapped back to the public symbol name.

Setting a Major ABI version

Downstreams might want to provide different DPDK releases at the same time to support multiple consumers of DPDK linked against older and newer sonames.

Also due to the interdependencies that DPDK libraries can have applications might end up with an executable space in which multiple versions of a library are mapped by ld.so.

Think of LibA that got an ABI bump and LibB that did not get an ABI bump but is depending on LibA.

Note

Application -> LibA.old -> LibB.new -> LibA.new

That is a conflict which can be avoided by setting CONFIG_RTE_MAJOR_ABI. If set, the value of CONFIG_RTE_MAJOR_ABI overwrites all - otherwise per library - versions defined in the libraries LIBABIVER. An example might be CONFIG_RTE_MAJOR_ABI=16.11 which will make all libraries librte<?>.so.16.11 instead of librte<?>.so.<LIBABIVER>.

Examples of ABI Macro use

Updating a public API

Assume we have a function as follows

/*
 * Create an acl context object for apps to
 * manipulate
 */
struct rte_acl_ctx *
rte_acl_create(const struct rte_acl_param *param)
{
       ...
}

Assume that struct rte_acl_ctx is a private structure, and that a developer wishes to enhance the acl api so that a debugging flag can be enabled on a per-context basis. This requires an addition to the structure (which, being private, is safe), but it also requires modifying the code as follows

/*
 * Create an acl context object for apps to
 * manipulate
 */
struct rte_acl_ctx *
rte_acl_create(const struct rte_acl_param *param, int debug)
{
       ...
}

Note also that, being a public function, the header file prototype must also be changed, as must all the call sites, to reflect the new ABI footprint. We will maintain previous ABI versions that are accessible only to previously compiled binaries

The addition of a parameter to the function is ABI breaking as the function is public, and existing application may use it in its current form. However, the compatibility macros in DPDK allow a developer to use symbol versioning so that multiple functions can be mapped to the same public symbol based on when an application was linked to it. To see how this is done, we start with the requisite libraries version map file. Initially the version map file for the acl library looks like this

DPDK_2.0 {
     global:

     rte_acl_add_rules;
     rte_acl_build;
     rte_acl_classify;
     rte_acl_classify_alg;
     rte_acl_classify_scalar;
     rte_acl_create;
     rte_acl_dump;
     rte_acl_find_existing;
     rte_acl_free;
     rte_acl_ipv4vlan_add_rules;
     rte_acl_ipv4vlan_build;
     rte_acl_list_dump;
     rte_acl_reset;
     rte_acl_reset_rules;
     rte_acl_set_ctx_classify;

     local: *;
};

This file needs to be modified as follows

DPDK_2.0 {
     global:

     rte_acl_add_rules;
     rte_acl_build;
     rte_acl_classify;
     rte_acl_classify_alg;
     rte_acl_classify_scalar;
     rte_acl_create;
     rte_acl_dump;
     rte_acl_find_existing;
     rte_acl_free;
     rte_acl_ipv4vlan_add_rules;
     rte_acl_ipv4vlan_build;
     rte_acl_list_dump;
     rte_acl_reset;
     rte_acl_reset_rules;
     rte_acl_set_ctx_classify;

     local: *;
};

DPDK_2.1 {
     global:
     rte_acl_create;

} DPDK_2.0;

The addition of the new block tells the linker that a new version node is available (DPDK_2.1), which contains the symbol rte_acl_create, and inherits the symbols from the DPDK_2.0 node. This list is directly translated into a list of exported symbols when DPDK is compiled as a shared library

Next, we need to specify in the code which function map to the rte_acl_create symbol at which versions. First, at the site of the initial symbol definition, we need to update the function so that it is uniquely named, and not in conflict with the public symbol name

 struct rte_acl_ctx *
-rte_acl_create(const struct rte_acl_param *param)
+rte_acl_create_v20(const struct rte_acl_param *param)
{
       size_t sz;
       struct rte_acl_ctx *ctx;
       ...

Note that the base name of the symbol was kept intact, as this is conducive to the macros used for versioning symbols. That is our next step, mapping this new symbol name to the initial symbol name at version node 2.0. Immediately after the function, we add this line of code

VERSION_SYMBOL(rte_acl_create, _v20, 2.0);

Remembering to also add the rte_compat.h header to the requisite c file where these changes are being made. The above macro instructs the linker to create a new symbol rte_acl_create@DPDK_2.0, which matches the symbol created in older builds, but now points to the above newly named function. We have now mapped the original rte_acl_create symbol to the original function (but with a new name)

Next, we need to create the 2.1 version of the symbol. We create a new function name, with a different suffix, and implement it appropriately

struct rte_acl_ctx *
rte_acl_create_v21(const struct rte_acl_param *param, int debug);
{
     struct rte_acl_ctx *ctx = rte_acl_create_v20(param);

     ctx->debug = debug;

     return ctx;
}

This code serves as our new API call. Its the same as our old call, but adds the new parameter in place. Next we need to map this function to the symbol rte_acl_create@DPDK_2.1. To do this, we modify the public prototype of the call in the header file, adding the macro there to inform all including applications, that on re-link, the default rte_acl_create symbol should point to this function. Note that we could do this by simply naming the function above rte_acl_create, and the linker would chose the most recent version tag to apply in the version script, but we can also do this in the header file

struct rte_acl_ctx *
-rte_acl_create(const struct rte_acl_param *param);
+rte_acl_create(const struct rte_acl_param *param, int debug);
+BIND_DEFAULT_SYMBOL(rte_acl_create, _v21, 2.1);

The BIND_DEFAULT_SYMBOL macro explicitly tells applications that include this header, to link to the rte_acl_create_v21 function and apply the DPDK_2.1 version node to it. This method is more explicit and flexible than just re-implementing the exact symbol name, and allows for other features (such as linking to the old symbol version by default, when the new ABI is to be opt-in for a period.

One last thing we need to do. Note that we’ve taken what was a public symbol, and duplicated it into two uniquely and differently named symbols. We’ve then mapped each of those back to the public symbol rte_acl_create with different version tags. This only applies to dynamic linking, as static linking has no notion of versioning. That leaves this code in a position of no longer having a symbol simply named rte_acl_create and a static build will fail on that missing symbol.

To correct this, we can simply map a function of our choosing back to the public symbol in the static build with the MAP_STATIC_SYMBOL macro. Generally the assumption is that the most recent version of the symbol is the one you want to map. So, back in the C file where, immediately after rte_acl_create_v21 is defined, we add this

struct rte_acl_ctx *
rte_acl_create_v21(const struct rte_acl_param *param, int debug)
{
     ...
}
MAP_STATIC_SYMBOL(struct rte_acl_ctx *rte_acl_create(const struct rte_acl_param *param, int debug), rte_acl_create_v21);

That tells the compiler that, when building a static library, any calls to the symbol rte_acl_create should be linked to rte_acl_create_v21

That’s it, on the next shared library rebuild, there will be two versions of rte_acl_create, an old DPDK_2.0 version, used by previously built applications, and a new DPDK_2.1 version, used by future built applications.

Deprecating part of a public API

Lets assume that you’ve done the above update, and after a few releases have passed you decide you would like to retire the old version of the function. After having gone through the ABI deprecation announcement process, removal is easy. Start by removing the symbol from the requisite version map file:

  DPDK_2.0 {
       global:

       rte_acl_add_rules;
       rte_acl_build;
       rte_acl_classify;
       rte_acl_classify_alg;
       rte_acl_classify_scalar;
       rte_acl_dump;
-      rte_acl_create
       rte_acl_find_existing;
       rte_acl_free;
       rte_acl_ipv4vlan_add_rules;
       rte_acl_ipv4vlan_build;
       rte_acl_list_dump;
       rte_acl_reset;
       rte_acl_reset_rules;
       rte_acl_set_ctx_classify;

       local: *;
  };

  DPDK_2.1 {
       global:
       rte_acl_create;
  } DPDK_2.0;

Next remove the corresponding versioned export.

-VERSION_SYMBOL(rte_acl_create, _v20, 2.0);

Note that the internal function definition could also be removed, but its used in our example by the newer version _v21, so we leave it in place. This is a coding style choice.

Lastly, we need to bump the LIBABIVER number for this library in the Makefile to indicate to applications doing dynamic linking that this is a later, and possibly incompatible library version:

-LIBABIVER := 1
+LIBABIVER := 2
Deprecating an entire ABI version

While removing a symbol from and ABI may be useful, it is often more practical to remove an entire version node at once. If a version node completely specifies an API, then removing part of it, typically makes it incomplete. In those cases it is better to remove the entire node

To do this, start by modifying the version map file, such that all symbols from the node to be removed are merged into the next node in the map

In the case of our map above, it would transform to look as follows

  DPDK_2.1 {
       global:

       rte_acl_add_rules;
       rte_acl_build;
       rte_acl_classify;
       rte_acl_classify_alg;
       rte_acl_classify_scalar;
       rte_acl_dump;
       rte_acl_create
       rte_acl_find_existing;
       rte_acl_free;
       rte_acl_ipv4vlan_add_rules;
       rte_acl_ipv4vlan_build;
       rte_acl_list_dump;
       rte_acl_reset;
       rte_acl_reset_rules;
       rte_acl_set_ctx_classify;

       local: *;
};

Then any uses of BIND_DEFAULT_SYMBOL that pointed to the old node should be updated to point to the new version node in any header files for all affected symbols.

-BIND_DEFAULT_SYMBOL(rte_acl_create, _v20, 2.0);
+BIND_DEFAULT_SYMBOL(rte_acl_create, _v21, 2.1);

Lastly, any VERSION_SYMBOL macros that point to the old version node should be removed, taking care to keep, where need old code in place to support newer versions of the symbol.

Running the ABI Validator

The devtools directory in the DPDK source tree contains a utility program, validate-abi.sh, for validating the DPDK ABI based on the Linux ABI Compliance Checker.

This has a dependency on the abi-compliance-checker and and abi-dumper utilities which can be installed via a package manager. For example:

sudo yum install abi-compliance-checker
sudo yum install abi-dumper

The syntax of the validate-abi.sh utility is:

./devtools/validate-abi.sh <REV1> <REV2> <TARGET>

Where REV1 and REV2 are valid gitrevisions(7) https://www.kernel.org/pub/software/scm/git/docs/gitrevisions.html on the local repo and target is the usual DPDK compilation target.

For example:

# Check between the previous and latest commit:
./devtools/validate-abi.sh HEAD~1 HEAD x86_64-native-linuxapp-gcc

# Check between two tags:
./devtools/validate-abi.sh v2.0.0 v2.1.0 x86_64-native-linuxapp-gcc

# Check between git master and local topic-branch "vhost-hacking":
./devtools/validate-abi.sh master vhost-hacking x86_64-native-linuxapp-gcc

After the validation script completes (it can take a while since it need to compile both tags) it will create compatibility reports in the ./compat_report directory. Listed incompatibilities can be found as follows:

grep -lr Incompatible compat_reports/

DPDK Documentation Guidelines

This document outlines the guidelines for writing the DPDK Guides and API documentation in RST and Doxygen format.

It also explains the structure of the DPDK documentation and shows how to build the Html and PDF versions of the documents.

Structure of the Documentation

The DPDK source code repository contains input files to build the API documentation and User Guides.

The main directories that contain files related to documentation are shown below:

lib
|-- librte_acl
|-- librte_cfgfile
|-- librte_cmdline
|-- librte_compat
|-- librte_eal
|   |-- ...
...
doc
|-- api
+-- guides
    |-- freebsd_gsg
    |-- linux_gsg
    |-- prog_guide
    |-- sample_app_ug
    |-- guidelines
    |-- testpmd_app_ug
    |-- rel_notes
    |-- nics
    |-- ...

The API documentation is built from Doxygen comments in the header files. These files are mainly in the lib/librte_* directories although some of the Poll Mode Drivers in drivers/net are also documented with Doxygen.

The configuration files that are used to control the Doxygen output are in the doc/api directory.

The user guides such as The Programmers Guide and the FreeBSD and Linux Getting Started Guides are generated from RST markup text files using the Sphinx Documentation Generator.

These files are included in the doc/guides/ directory. The output is controlled by the doc/guides/conf.py file.

Role of the Documentation

The following items outline the roles of the different parts of the documentation and when they need to be updated or added to by the developer.

  • Release Notes

    The Release Notes document which features have been added in the current and previous releases of DPDK and highlight any known issues. The Releases Notes also contain notifications of features that will change ABI compatibility in the next major release.

    Developers should include updates to the Release Notes with patch sets that relate to any of the following sections:

    • New Features
    • Resolved Issues (see below)
    • Known Issues
    • API Changes
    • ABI Changes
    • Shared Library Versions

    Resolved Issues should only include issues from previous releases that have been resolved in the current release. Issues that are introduced and then fixed within a release cycle do not have to be included here.

    Refer to the Release Notes from the previous DPDK release for the correct format of each section.

  • API documentation

    The API documentation explains how to use the public DPDK functions. The API index page shows the generated API documentation with related groups of functions.

    The API documentation should be updated via Doxygen comments when new functions are added.

  • Getting Started Guides

    The Getting Started Guides show how to install and configure DPDK and how to run DPDK based applications on different OSes.

    A Getting Started Guide should be added when DPDK is ported to a new OS.

  • The Programmers Guide

    The Programmers Guide explains how the API components of DPDK such as the EAL, Memzone, Rings and the Hash Library work. It also explains how some higher level functionality such as Packet Distributor, Packet Framework and KNI work. It also shows the build system and explains how to add applications.

    The Programmers Guide should be expanded when new functionality is added to DPDK.

  • App Guides

    The app guides document the DPDK applications in the app directory such as testpmd.

    The app guides should be updated if functionality is changed or added.

  • Sample App Guides

    The sample app guides document the DPDK example applications in the examples directory. Generally they demonstrate a major feature such as L2 or L3 Forwarding, Multi Process or Power Management. They explain the purpose of the sample application, how to run it and step through some of the code to explain the major functionality.

    A new sample application should be accompanied by a new sample app guide. The guide for the Skeleton Forwarding app is a good starting reference.

  • Network Interface Controller Drivers

    The NIC Drivers document explains the features of the individual Poll Mode Drivers, such as software requirements, configuration and initialization.

    New documentation should be added for new Poll Mode Drivers.

  • Guidelines

    The guideline documents record community process, expectations and design directions.

    They can be extended, amended or discussed by submitting a patch and getting community approval.

Building the Documentation

Dependencies

The following dependencies must be installed to build the documentation:

  • Doxygen.
  • Sphinx (also called python-sphinx).
  • TexLive (at least TexLive-core and the extra Latex support).
  • Inkscape.

Doxygen generates documentation from commented source code. It can be installed as follows:

# Ubuntu/Debian.
sudo apt-get -y install doxygen

# Red Hat/Fedora.
sudo dnf     -y install doxygen

Sphinx is a Python documentation tool for converting RST files to Html or to PDF (via LaTeX). For full support with figure and table captioning the latest version of Sphinx can be installed as follows:

# Ubuntu/Debian.
sudo apt-get -y install python-pip
sudo pip install --upgrade sphinx
sudo pip install --upgrade sphinx_rtd_theme

# Red Hat/Fedora.
sudo dnf     -y install python-pip
sudo pip install --upgrade sphinx
sudo pip install --upgrade sphinx_rtd_theme

For further information on getting started with Sphinx see the Sphinx Tutorial.

Note

To get full support for Figure and Table numbering it is best to install Sphinx 1.3.1 or later.

Inkscape is a vector based graphics program which is used to create SVG images and also to convert SVG images to PDF images. It can be installed as follows:

# Ubuntu/Debian.
sudo apt-get -y install inkscape

# Red Hat/Fedora.
sudo dnf     -y install inkscape

TexLive is an installation package for Tex/LaTeX. It is used to generate the PDF versions of the documentation. The main required packages can be installed as follows:

# Ubuntu/Debian.
sudo apt-get -y install texlive-latex-extra

# Red Hat/Fedora, selective install.
sudo dnf     -y install texlive-collection-latexextra
Build commands

The documentation is built using the standard DPDK build system. Some examples are shown below:

  • Generate all the documentation targets:

    make doc
    
  • Generate the Doxygen API documentation in Html:

    make doc-api-html
    
  • Generate the guides documentation in Html:

    make doc-guides-html
    
  • Generate the guides documentation in Pdf:

    make doc-guides-pdf
    

The output of these commands is generated in the build directory:

build/doc
      |-- html
      |   |-- api
      |   +-- guides
      |
      +-- pdf
          +-- guides

Note

Make sure to fix any Sphinx or Doxygen warnings when adding or updating documentation.

The documentation output files can be removed as follows:

make doc-clean

Document Guidelines

Here are some guidelines in relation to the style of the documentation:

  • Document the obvious as well as the obscure since it won’t always be obvious to the reader. For example an instruction like “Set up 64 2MB Hugepages” is better when followed by a sample commandline or a link to the appropriate section of the documentation.

  • Use American English spellings throughout. This can be checked using the aspell utility:

    aspell --lang=en_US --check doc/guides/sample_app_ug/mydoc.rst
    

RST Guidelines

The RST (reStructuredText) format is a plain text markup format that can be converted to Html, PDF or other formats. It is most closely associated with Python but it can be used to document any language. It is used in DPDK to document everything apart from the API.

The Sphinx documentation contains a very useful RST Primer which is a good place to learn the minimal set of syntax required to format a document.

The official reStructuredText website contains the specification for the RST format and also examples of how to use it. However, for most developers the RST Primer is a better resource.

The most common guidelines for writing RST text are detailed in the Documenting Python guidelines. The additional guidelines below reiterate or expand upon those guidelines.

Line Length
  • Lines in sentences should be less than 80 characters and wrapped at words. Multiple sentences which are not separated by a blank line are joined automatically into paragraphs.

  • Lines in literal blocks must be less than 80 characters since they are not wrapped by the document formatters and can exceed the page width in PDF documents.

    Long literal command lines can be shown wrapped with backslashes. For example:

    testpmd -l 2-3 -n 4 \
            --vdev=virtio_user0,path=/dev/vhost-net,queues=2,queue_size=1024 \
            -- -i --txqflags=0x0 --disable-hw-vlan --enable-lro \
            --enable-rx-cksum --txq=2 --rxq=2 --rxd=1024  --txd=1024
    
Whitespace
  • Standard RST indentation is 3 spaces. Code can be indented 4 spaces, especially if it is copied from source files.
  • No tabs. Convert tabs in embedded code to 4 or 8 spaces.
  • No trailing whitespace.
  • Add 2 blank lines before each section header.
  • Add 1 blank line after each section header.
  • Add 1 blank line between each line of a list.
Section Headers
  • Section headers should use the following underline formats:

    Level 1 Heading
    ===============
    
    
    Level 2 Heading
    ---------------
    
    
    Level 3 Heading
    ~~~~~~~~~~~~~~~
    
    
    Level 4 Heading
    ^^^^^^^^^^^^^^^
    
  • Level 4 headings should be used sparingly.

  • The underlines should match the length of the text.

  • In general, the heading should be less than 80 characters, for conciseness.

  • As noted above:

    • Add 2 blank lines before each section header.
    • Add 1 blank line after each section header.
Lists
  • Bullet lists should be formatted with a leading * as follows:

    * Item one.
    
    * Item two is a long line that is wrapped and then indented to match
      the start of the previous line.
    
    * One space character between the bullet and the text is preferred.
    
  • Numbered lists can be formatted with a leading number but the preference is to use #. which will give automatic numbering. This is more convenient when adding or removing items:

    #. Item one.
    
    #. Item two is a long line that is wrapped and then indented to match
       the start of the previous line.
    
    #. Item three.
    
  • Definition lists can be written with or without a bullet:

    * Item one.
    
      Some text about item one.
    
    * Item two.
    
      Some text about item two.
    
  • All lists, and sub-lists, must be separated from the preceding text by a blank line. This is a syntax requirement.

  • All list items should be separated by a blank line for readability.

Code and Literal block sections
  • Inline text that is required to be rendered with a fixed width font should be enclosed in backquotes like this: ``text``, so that it appears like this: text.

  • Fixed width, literal blocks of texts should be indented at least 3 spaces and prefixed with :: like this:

    Here is some fixed width text::
    
       0x0001 0x0001 0x00FF 0x00FF
    
  • It is also possible to specify an encoding for a literal block using the .. code-block:: directive so that syntax highlighting can be applied. Examples of supported highlighting are:

    .. code-block:: console
    .. code-block:: c
    .. code-block:: python
    .. code-block:: diff
    .. code-block:: none
    

    That can be applied as follows:

    .. code-block:: c
    
       #include<stdio.h>
    
       int main() {
    
          printf("Hello World\n");
    
          return 0;
       }
    

    Which would be rendered as:

    #include<stdio.h>
    
    int main() {
    
       printf("Hello World\n");
    
       return 0;
    }
    
  • The default encoding for a literal block using the simplified :: directive is none.

  • Lines in literal blocks must be less than 80 characters since they can exceed the page width when converted to PDF documentation. For long literal lines that exceed that limit try to wrap the text at sensible locations. For example a long command line could be documented like this and still work if copied directly from the docs:

    build/app/testpmd -l 0-2 -n3 --vdev=net_pcap0,iface=eth0     \
                              --vdev=net_pcap1,iface=eth1     \
                              -- -i --nb-cores=2 --nb-ports=2 \
                                 --total-num-mbufs=2048
    
  • Long lines that cannot be wrapped, such as application output, should be truncated to be less than 80 characters.

Images
  • All images should be in SVG scalar graphics format. They should be true SVG XML files and should not include binary formats embedded in a SVG wrapper.

  • The DPDK documentation contains some legacy images in PNG format. These will be converted to SVG in time.

  • Inkscape is the recommended graphics editor for creating the images. Use some of the older images in doc/guides/prog_guide/img/ as a template, for example mbuf1.svg or ring-enqueue1.svg.

  • The SVG images should include a copyright notice, as an XML comment.

  • Images in the documentation should be formatted as follows:

    • The image should be preceded by a label in the format .. _figure_XXXX: with a leading underscore and where XXXX is a unique descriptive name.
    • Images should be included using the .. figure:: directive and the file type should be set to * (not .svg). This allows the format of the image to be changed if required, without updating the documentation.
    • Images must have a caption as part of the .. figure:: directive.
  • Here is an example of the previous three guidelines:

    .. _figure_mempool:
    
    .. figure:: img/mempool.*
    
       A mempool in memory with its associated ring.
    
  • Images can then be linked to using the :numref: directive:

    The mempool layout is shown in :numref:`figure_mempool`.
    

    This would be rendered as: The mempool layout is shown in Fig 6.3.

    Note: The :numref: directive requires Sphinx 1.3.1 or later. With earlier versions it will still be rendered as a link but won’t have an automatically generated number.

  • The caption of the image can be generated, with a link, using the :ref: directive:

    :ref:`figure_mempool`
    

    This would be rendered as: A mempool in memory with its associated ring.

Tables
  • RST tables should be used sparingly. They are hard to format and to edit, they are often rendered incorrectly in PDF format, and the same information can usually be shown just as clearly with a definition or bullet list.

  • Tables in the documentation should be formatted as follows:

    • The table should be preceded by a label in the format .. _table_XXXX: with a leading underscore and where XXXX is a unique descriptive name.
    • Tables should be included using the .. table:: directive and must have a caption.
  • Here is an example of the previous two guidelines:

    .. _table_qos_pipes:
    
    .. table:: Sample configuration for QOS pipes.
    
       +----------+----------+----------+
       | Header 1 | Header 2 | Header 3 |
       |          |          |          |
       +==========+==========+==========+
       | Text     | Text     | Text     |
       +----------+----------+----------+
       | ...      | ...      | ...      |
       +----------+----------+----------+
    
  • Tables can be linked to using the :numref: and :ref: directives, as shown in the previous section for images. For example:

    The QOS configuration is shown in :numref:`table_qos_pipes`.
    
  • Tables should not include merged cells since they are not supported by the PDF renderer.

Doxygen Guidelines

The DPDK API is documented using Doxygen comment annotations in the header files. Doxygen is a very powerful tool, it is extremely configurable and with a little effort can be used to create expressive documents. See the Doxygen website for full details on how to use it.

The following are some guidelines for use of Doxygen in the DPDK API documentation:

  • New libraries that are documented with Doxygen should be added to the Doxygen configuration file: doc/api/doxy-api.conf. It is only required to add the directory that contains the files. It isn’t necessary to explicitly name each file since the configuration matches all rte_*.h files in the directory.

  • Use proper capitalization and punctuation in the Doxygen comments since they will become sentences in the documentation. This in particular applies to single line comments, which is the case the is most often forgotten.

  • Use @ style Doxygen commands instead of \ style commands.

  • Add a general description of each library at the head of the main header files:

    /**
     * @file
     * RTE Mempool.
     *
     * A memory pool is an allocator of fixed-size object. It is
     * identified by its name, and uses a ring to store free objects.
     * ...
     */
    
  • Document the purpose of a function, the parameters used and the return value:

    /**
     * Attach a new Ethernet device specified by arguments.
     *
     * @param devargs
     *  A pointer to a strings array describing the new device
     *  to be attached. The strings should be a pci address like
     *  `0000:01:00.0` or **virtual** device name like `net_pcap0`.
     * @param port_id
     *  A pointer to a port identifier actually attached.
     *
     * @return
     *  0 on success and port_id is filled, negative on error.
     */
    int rte_eth_dev_attach(const char *devargs, uint8_t *port_id);
    
  • Doxygen supports Markdown style syntax such as bold, italics, fixed width text and lists. For example the second line in the devargs parameter in the previous example will be rendered as:

    The strings should be a pci address like 0000:01:00.0 or virtual device name like net_pcap0.

  • Use - instead of * for lists within the Doxygen comment since the latter can get confused with the comment delimiter.

  • Add an empty line between the function description, the @params and @return for readability.

  • Place the @params description on separate line and indent it by 2 spaces. (It would be better to use no indentation since this is more common and also because checkpatch complains about leading whitespace in comments. However this is the convention used in the existing DPDK code.)

  • Documented functions can be linked to simply by adding () to the function name:

    /**
     * The functions exported by the application Ethernet API to setup
     * a device designated by its port identifier must be invoked in
     * the following order:
     *     - rte_eth_dev_configure()
     *     - rte_eth_tx_queue_setup()
     *     - rte_eth_rx_queue_setup()
     *     - rte_eth_dev_start()
     */
    

    In the API documentation the functions will be rendered as links, see the online section of the rte_ethdev.h docs that contains the above text.

  • The @see keyword can be used to create a see also link to another file or library. This directive should be placed on one line at the bottom of the documentation section.

    /**
     * ...
     *
     * Some text that references mempools.
     *
     * @see eal_memzone.c
     */
    
  • Doxygen supports two types of comments for documenting variables, constants and members: prefix and postfix:

    /** This is a prefix comment. */
    #define RTE_FOO_ERROR  0x023.
    
    #define RTE_BAR_ERROR  0x024. /**< This is a postfix comment. */
    
  • Postfix comments are preferred for struct members and constants if they can be documented in the same way:

    struct rte_eth_stats {
        uint64_t ipackets; /**< Total number of received packets. */
        uint64_t opackets; /**< Total number of transmitted packets.*/
        uint64_t ibytes;   /**< Total number of received bytes. */
        uint64_t obytes;   /**< Total number of transmitted bytes. */
        uint64_t imissed;  /**< Total of RX missed packets. */
        uint64_t ibadcrc;  /**< Total of RX packets with CRC error. */
        uint64_t ibadlen;  /**< Total of RX packets with bad length. */
    }
    

    Note: postfix comments should be aligned with spaces not tabs in accordance with the DPDK Coding Style.

  • If a single comment type can’t be used, due to line length limitations then prefix comments should be preferred. For example this section of the code contains prefix comments, postfix comments on the same line and postfix comments on a separate line:

    /** Number of elements in the elt_pa array. */
    uint32_t    pg_num __rte_cache_aligned;
    uint32_t    pg_shift;     /**< LOG2 of the physical pages. */
    uintptr_t   pg_mask;      /**< Physical page mask value. */
    uintptr_t   elt_va_start;
    /**< Virtual address of the first mempool object. */
    uintptr_t   elt_va_end;
    /**< Virtual address of the <size + 1> mempool object. */
    phys_addr_t elt_pa[MEMPOOL_PG_NUM_DEFAULT];
    /**< Array of physical page addresses for the mempool buffer. */
    

    This doesn’t have an effect on the rendered documentation but it is confusing for the developer reading the code. It this case it would be clearer to use prefix comments throughout:

    /** Number of elements in the elt_pa array. */
    uint32_t    pg_num __rte_cache_aligned;
    /** LOG2 of the physical pages. */
    uint32_t    pg_shift;
    /** Physical page mask value. */
    uintptr_t   pg_mask;
    /** Virtual address of the first mempool object. */
    uintptr_t   elt_va_start;
    /** Virtual address of the <size + 1> mempool object. */
    uintptr_t   elt_va_end;
    /** Array of physical page addresses for the mempool buffer. */
    phys_addr_t elt_pa[MEMPOOL_PG_NUM_DEFAULT];
    
  • Check for Doxygen warnings in new code by checking the API documentation build:

    make doc-api-html >/dev/null
    
  • Read the rendered section of the documentation that you have added for correctness, clarity and consistency with the surrounding text.

Contributing Code to DPDK

This document outlines the guidelines for submitting code to DPDK.

The DPDK development process is modelled (loosely) on the Linux Kernel development model so it is worth reading the Linux kernel guide on submitting patches: How to Get Your Change Into the Linux Kernel. The rationale for many of the DPDK guidelines is explained in greater detail in the kernel guidelines.

The DPDK Development Process

The DPDK development process has the following features:

  • The code is hosted in a public git repository.
  • There is a mailing list where developers submit patches.
  • There are maintainers for hierarchical components.
  • Patches are reviewed publicly on the mailing list.
  • Successfully reviewed patches are merged to the repository.
  • Patches should be sent to the target repository or sub-tree, see below.
  • All sub-repositories are merged into main repository for -rc1 and -rc2 versions of the release.
  • After the -rc2 release all patches should target the main repository.

The mailing list for DPDK development is dev@dpdk.org. Contributors will need to register for the mailing list in order to submit patches. It is also worth registering for the DPDK Patchwork

The development process requires some familiarity with the git version control system. Refer to the Pro Git Book for further information.

Maintainers and Sub-trees

The DPDK maintenance hierarchy is divided into a main repository dpdk and sub-repositories dpdk-next-*.

There are maintainers for the trees and for components within the tree.

Trees and maintainers are listed in the MAINTAINERS file. For example:

Crypto Drivers
--------------
M: Some Name <some.name@email.com>
B: Another Name <another.name@email.com>
T: git://dpdk.org/next/dpdk-next-crypto

Intel AES-NI GCM PMD
M: Some One <some.one@email.com>
F: drivers/crypto/aesni_gcm/
F: doc/guides/cryptodevs/aesni_gcm.rst

Where:

  • M is a tree or component maintainer.
  • B is a tree backup maintainer.
  • T is a repository tree.
  • F is a maintained file or directory.

Additional details are given in the MAINTAINERS file.

The role of the component maintainers is to:

  • Review patches for the component or delegate the review. The review should be done, ideally, within 1 week of submission to the mailing list.
  • Add an acked-by to patches, or patchsets, that are ready for committing to a tree.
  • Reply to questions asked about the component.

Component maintainers can be added or removed by submitting a patch to the MAINTAINERS file. Maintainers should have demonstrated a reasonable level of contributions or reviews to the component area. The maintainer should be confirmed by an ack from an established contributor. There can be more than one component maintainer if desired.

The role of the tree maintainers is to:

  • Maintain the overall quality of their tree. This can entail additional review, compilation checks or other tests deemed necessary by the maintainer.
  • Commit patches that have been reviewed by component maintainers and/or other contributors. The tree maintainer should determine if patches have been reviewed sufficiently.
  • Ensure that patches are reviewed in a timely manner.
  • Prepare the tree for integration.
  • Ensure that there is a designated back-up maintainer and coordinate a handover for periods where the tree maintainer can’t perform their role.

Tree maintainers can be added or removed by submitting a patch to the MAINTAINERS file. The proposer should justify the need for a new sub-tree and should have demonstrated a sufficient level of contributions in the area or to a similar area. The maintainer should be confirmed by an ack from an existing tree maintainer. Disagreements on trees or maintainers can be brought to the Technical Board.

The backup maintainer for the master tree should be selected from the existing sub-tree maintainers from the project. The backup maintainer for a sub-tree should be selected from among the component maintainers within that sub-tree.

Getting the Source Code

The source code can be cloned using either of the following:

main repository:

git clone git://dpdk.org/dpdk
git clone http://dpdk.org/git/dpdk

sub-repositories (list):

git clone git://dpdk.org/next/dpdk-next-*
git clone http://dpdk.org/git/next/dpdk-next-*

Make your Changes

Make your planned changes in the cloned dpdk repo. Here are some guidelines and requirements:

  • Follow the DPDK Coding Style guidelines.
  • If you add new files or directories you should add your name to the MAINTAINERS file.
  • New external functions should be added to the local version.map file. See the Guidelines for ABI policy and versioning. New external functions should also be added in alphabetical order.
  • Important changes will require an addition to the release notes in doc/guides/rel_notes/. See the Release Notes section of the Documentation Guidelines for details.
  • Test the compilation works with different targets, compilers and options, see Checking Compilation.
  • Don’t break compilation between commits with forward dependencies in a patchset. Each commit should compile on its own to allow for git bisect and continuous integration testing.
  • Add tests to the the app/test unit test framework where possible.
  • Add documentation, if relevant, in the form of Doxygen comments or a User Guide in RST format. See the Documentation Guidelines.

Once the changes have been made you should commit them to your local repo.

For small changes, that do not require specific explanations, it is better to keep things together in the same patch. Larger changes that require different explanations should be separated into logical patches in a patchset. A good way of thinking about whether a patch should be split is to consider whether the change could be applied without dependencies as a backport.

As a guide to how patches should be structured run git log on similar files.

Commit Messages: Subject Line

The first, summary, line of the git commit message becomes the subject line of the patch email. Here are some guidelines for the summary line:

  • The summary line must capture the area and the impact of the change.

  • The summary line should be around 50 characters.

  • The summary line should be lowercase apart from acronyms.

  • It should be prefixed with the component name (use git log to check existing components). For example:

    ixgbe: fix offload config option name
    
    config: increase max queues per port
    
  • Use the imperative of the verb (like instructions to the code base).

  • Don’t add a period/full stop to the subject line or you will end up two in the patch name: dpdk_description..patch.

The actual email subject line should be prefixed by [PATCH] and the version, if greater than v1, for example: PATCH v2. The is generally added by git send-email or git format-patch, see below.

If you are submitting an RFC draft of a feature you can use [RFC] instead of [PATCH]. An RFC patch doesn’t have to be complete. It is intended as a way of getting early feedback.

Commit Messages: Body

Here are some guidelines for the body of a commit message:

  • The body of the message should describe the issue being fixed or the feature being added. It is important to provide enough information to allow a reviewer to understand the purpose of the patch.

  • When the change is obvious the body can be blank, apart from the signoff.

  • The commit message must end with a Signed-off-by: line which is added using:

    git commit --signoff # or -s
    

    The purpose of the signoff is explained in the Developer’s Certificate of Origin section of the Linux kernel guidelines.

    Note

    All developers must ensure that they have read and understood the Developer’s Certificate of Origin section of the documentation prior to applying the signoff and submitting a patch.

  • The signoff must be a real name and not an alias or nickname. More than one signoff is allowed.

  • The text of the commit message should be wrapped at 72 characters.

  • When fixing a regression, it is required to reference the id of the commit which introduced the bug, and put the original author of that commit on CC. You can generate the required lines using the following git alias, which prints the commit SHA and the author of the original code:

    git config alias.fixline "log -1 --abbrev=12 --format='Fixes: %h (\"%s\")%nCc: %ae'"
    

    The output of git fixline <SHA> must then be added to the commit message:

    doc: fix some parameter description
    
    Update the docs, fixing description of some parameter.
    
    Fixes: abcdefgh1234 ("doc: add some parameter")
    Cc: author@example.com
    
    Signed-off-by: Alex Smith <alex.smith@example.com>
    
  • When fixing an error or warning it is useful to add the error message and instructions on how to reproduce it.

  • Use correct capitalization, punctuation and spelling.

In addition to the Signed-off-by: name the commit messages can also have tags for who reported, suggested, tested and reviewed the patch being posted. Please refer to the Tested, Acked and Reviewed by section.

Creating Patches

It is possible to send patches directly from git but for new contributors it is recommended to generate the patches with git format-patch and then when everything looks okay, and the patches have been checked, to send them with git send-email.

Here are some examples of using git format-patch to generate patches:

# Generate a patch from the last commit.
git format-patch -1

# Generate a patch from the last 3 commits.
git format-patch -3

# Generate the patches in a directory.
git format-patch -3 -o ~/patch/

# Add a cover letter to explain a patchset.
git format-patch -3 -o ~/patch/ --cover-letter

# Add a prefix with a version number.
git format-patch -3 -o ~/patch/ -v 2

Cover letters are useful for explaining a patchset and help to generate a logical threading to the patches. Smaller notes can be put inline in the patch after the --- separator, for example:

Subject: [PATCH] fm10k/base: add FM10420 device ids

Add the device ID for Boulder Rapids and Atwood Channel to enable
drivers to support those devices.

Signed-off-by: Alex Smith <alex.smith@example.com>
---

ADD NOTES HERE.

 drivers/net/fm10k/base/fm10k_api.c  | 6 ++++++
 drivers/net/fm10k/base/fm10k_type.h | 6 ++++++
 2 files changed, 12 insertions(+)
...

Version 2 and later of a patchset should also include a short log of the changes so the reviewer knows what has changed. This can be added to the cover letter or the annotations. For example:

---
v3:
* Fixed issued with version.map.

v2:
* Added i40e support.
* Renamed ethdev functions from rte_eth_ieee15888_*() to rte_eth_timesync_*()
  since 802.1AS can be supported through the same interfaces.

Checking the Patches

Patches should be checked for formatting and syntax issues using the checkpatches.sh script in the devtools directory of the DPDK repo. This uses the Linux kernel development tool checkpatch.pl which can be obtained by cloning, and periodically, updating the Linux kernel sources.

The path to the original Linux script must be set in the environment variable DPDK_CHECKPATCH_PATH. This, and any other configuration variables required by the development tools, are loaded from the following files, in order of preference:

.develconfig
~/.config/dpdk/devel.config
/etc/dpdk/devel.config.

Once the environment variable the script can be run as follows:

devtools/checkpatches.sh ~/patch/

The script usage is:

checkpatches.sh [-h] [-q] [-v] [patch1 [patch2] ...]]"

Where:

  • -h: help, usage.
  • -q: quiet. Don’t output anything for files without issues.
  • -v: verbose.
  • patchX: path to one or more patches.

Then the git logs should be checked using the check-git-log.sh script.

The script usage is:

check-git-log.sh [range]

Where the range is a git log option.

Checking Compilation

Compilation of patches and changes should be tested using the the test-build.sh script in the devtools directory of the DPDK repo:

devtools/test-build.sh x86_64-native-linuxapp-gcc+next+shared

The script usage is:

test-build.sh [-h] [-jX] [-s] [config1 [config2] ...]]

Where:

  • -h: help, usage.
  • -jX: use X parallel jobs in “make”.
  • -s: short test with only first config and without examples/doc.
  • config: default config name plus config switches delimited with a + sign.

Examples of configs are:

x86_64-native-linuxapp-gcc
x86_64-native-linuxapp-gcc+next+shared
x86_64-native-linuxapp-clang+shared

The builds can be modified via the following environmental variables:

  • DPDK_BUILD_TEST_CONFIGS (target1+option1+option2 target2)
  • DPDK_DEP_CFLAGS
  • DPDK_DEP_LDFLAGS
  • DPDK_DEP_PCAP (y/[n])
  • DPDK_NOTIFY (notify-send)

These can be set from the command line or in the config files shown above in the Checking the Patches.

The recommended configurations and options to test compilation prior to submitting patches are:

x86_64-native-linuxapp-gcc+shared+next
x86_64-native-linuxapp-clang+shared
i686-native-linuxapp-gcc

export DPDK_DEP_ZLIB=y
export DPDK_DEP_PCAP=y
export DPDK_DEP_SSL=y

Sending Patches

Patches should be sent to the mailing list using git send-email. You can configure an external SMTP with something like the following:

[sendemail]
    smtpuser = name@domain.com
    smtpserver = smtp.domain.com
    smtpserverport = 465
    smtpencryption = ssl

See the Git send-email documentation for more details.

The patches should be sent to dev@dpdk.org. If the patches are a change to existing files then you should send them TO the maintainer(s) and CC dev@dpdk.org. The appropriate maintainer can be found in the MAINTAINERS file:

git send-email --to maintainer@some.org --cc dev@dpdk.org 000*.patch

Script get-maintainer.sh can be used to select maintainers automatically:

git send-email --to-cmd ./devtools/get-maintainer.sh --cc dev@dpdk.org 000*.patch

New additions can be sent without a maintainer:

git send-email --to dev@dpdk.org 000*.patch

You can test the emails by sending it to yourself or with the --dry-run option.

If the patch is in relation to a previous email thread you can add it to the same thread using the Message ID:

git send-email --to dev@dpdk.org --in-reply-to <1234-foo@bar.com> 000*.patch

The Message ID can be found in the raw text of emails or at the top of each Patchwork patch, for example. Shallow threading (--thread --no-chain-reply-to) is preferred for a patch series.

Once submitted your patches will appear on the mailing list and in Patchwork.

Experienced committers may send patches directly with git send-email without the git format-patch step. The options --annotate and confirm = always are recommended for checking patches before sending.

The Review Process

Patches are reviewed by the community, relying on the experience and collaboration of the members to double-check each other’s work. There are a number of ways to indicate that you have checked a patch on the mailing list.

Tested, Acked and Reviewed by

To indicate that you have interacted with a patch on the mailing list you should respond to the patch in an email with one of the following tags:

  • Reviewed-by:
  • Acked-by:
  • Tested-by:
  • Reported-by:
  • Suggested-by:

The tag should be on a separate line as follows:

tag-here: Name Surname <email@address.com>

Each of these tags has a specific meaning. In general, the DPDK community follows the kernel usage of the tags. A short summary of the meanings of each tag is given here for reference:

Reviewed-by: is a strong statement that the patch is an appropriate state for merging without any remaining serious technical issues. Reviews from community members who are known to understand the subject area and to perform thorough reviews will increase the likelihood of the patch getting merged.

Acked-by: is a record that the person named was not directly involved in the preparation of the patch but wishes to signify and record their acceptance and approval of it.

Tested-by: indicates that the patch has been successfully tested (in some environment) by the person named.

Reported-by: is used to acknowledge person who found or reported the bug.

Suggested-by: indicates that the patch idea was suggested by the named person.

Steps to getting your patch merged

The more work you put into the previous steps the easier it will be to get a patch accepted. The general cycle for patch review and acceptance is:

  1. Submit the patch.

  2. Check the automatic test reports in the coming hours.

  3. Wait for review comments. While you are waiting review some other patches.

  4. Fix the review comments and submit a v n+1 patchset:

    git format-patch -3 -v 2
    
  5. Update Patchwork to mark your previous patches as “Superseded”.

  6. If the patch is deemed suitable for merging by the relevant maintainer(s) or other developers they will ack the patch with an email that includes something like:

    Acked-by: Alex Smith <alex.smith@example.com>
    

    Note: When acking patches please remove as much of the text of the patch email as possible. It is generally best to delete everything after the Signed-off-by: line.

  7. Having the patch Reviewed-by: and/or Tested-by: will also help the patch to be accepted.

  8. If the patch isn’t deemed suitable based on being out of scope or conflicting with existing functionality it may receive a nack. In this case you will need to make a more convincing technical argument in favor of your patches.

  9. In addition a patch will not be accepted if it doesn’t address comments from a previous version with fixes or valid arguments.

  10. It is the responsibility of a maintainer to ensure that patches are reviewed and to provide an ack or nack of those patches as appropriate.

  11. Once a patch has been acked by the relevant maintainer, reviewers may still comment on it for a further two weeks. After that time, the patch should be merged into the relevant git tree for the next release. Additional notes and restrictions:

    • Patches should be acked by a maintainer at least two days before the release merge deadline, in order to make that release.
    • For patches acked with less than two weeks to go to the merge deadline, all additional comments should be made no later than two days before the merge deadline.
    • After the appropriate time for additional feedback has passed, if the patch has not yet been merged to the relevant tree by the committer, it should be treated as though it had, in that any additional changes needed to it must be addressed by a follow-on patch, rather than rework of the original.
    • Trivial patches may be merged sooner than described above at the tree committer’s discretion.

DPDK Stable Releases and Long Term Support

This section sets out the guidelines for the DPDK Stable Releases and the DPDK Long Term Support releases (LTS).

Introduction

The purpose of the DPDK Stable Releases is to maintain releases of DPDK with backported fixes over an extended period of time. This provides downstream consumers of DPDK with a stable target on which to base applications or packages.

The Long Term Support release (LTS) is a designation applied to a Stable Release to indicate longer term support.

Stable Releases

Any major release of DPDK can be designated as a Stable Release if a maintainer volunteers to maintain it.

A Stable Release is used to backport fixes from an N release back to an N-1 release, for example, from 16.11 to 16.07.

The duration of a stable is one complete release cycle (3 months). It can be longer, up to 1 year, if a maintainer continues to support the stable branch, or if users supply backported fixes, however the explicit commitment should be for one release cycle.

The release cadence is determined by the maintainer based on the number of bugfixes and the criticality of the bugs. Releases should be coordinated with the validation engineers to ensure that a tagged release has been tested.

LTS Release

A stable release can be designated as an LTS release based on community agreement and a commitment from a maintainer. The current policy is that each year’s November release will be maintained as an LTS for 2 years.

The current DPDK LTS releases are 16.11 and 17.11.

It is anticipated that there will be at least 4 releases per year of the LTS or approximately 1 every 3 months. However, the cadence can be shorter or longer depending on the number and criticality of the backported fixes. Releases should be coordinated with the validation engineers to ensure that a tagged release has been tested.

What changes should be backported

Backporting should be limited to bug fixes.

Features should not be backported to stable releases. It may be acceptable, in limited cases, to back port features for the LTS release where:

  • There is a justifiable use case (for example a new PMD).
  • The change is non-invasive.
  • The work of preparing the backport is done by the proposer.
  • There is support within the community.

The Stable Mailing List

The Stable and LTS release are coordinated on the stable@dpdk.org mailing list.

All fix patches to the master branch that are candidates for backporting should also be CCed to the stable@dpdk.org mailing list.

Releasing

A Stable Release will be released by:

  • Tagging the release with YY.MM.n (year, month, number).
  • Uploading a tarball of the release to dpdk.org.
  • Sending an announcement to the announce@dpdk.org list.

Stable releases are available on the dpdk.org download page.

ABI

The Stable Release should not be seen as a way of breaking or circumventing the DPDK ABI policy.

Patch Cheatsheet

_images/patch_cheatsheet.svg

Cheat sheet for submitting patches to dev@dpdk.org

Release Notes

Description of Release

This document contains the release notes for Data Plane Development Kit (DPDK) release version 17.11.0 and previous releases.

It lists new features, fixed bugs, API and ABI changes and known issues.

For instructions on compiling and running the release, see the DPDK Getting Started Guide.

DPDK Release 17.11

New Features

  • Extended port_id range from uint8_t to uint16_t.

    Increased the port_id range from 8 bits to 16 bits in order to support more than 256 ports in DPDK. All ethdev APIs which have port_id as parameter have been changed.

  • Modified the return type of rte_eth_stats_reset.

    Changed return type of rte_eth_stats_reset from void to int so that the caller can determine whether a device supports the operation or not and if the operation was carried out.

  • Added a new driver for Marvell Armada 7k/8k devices.

    Added the new mrvl net driver for Marvell Armada 7k/8k devices. See the MRVL Poll Mode Driver NIC guide for more details on this new driver.

  • Updated mlx4 driver.

    Updated the mlx4 driver including the following changes:

    • Isolated mode (rte_flow) can now be enabled anytime, not only during initial device configuration.
    • Flow rules now support up to 4096 priority levels usable at will by applications.
    • Enhanced error message to help debugging invalid/unsupported flow rules.
    • Flow rules matching all multicast and promiscuous traffic are now allowed.
    • No more software restrictions on flow rules with the RSS action, their configuration is much more flexible.
    • Significantly reduced memory footprint for Rx and Tx queue objects.
    • While supported, UDP RSS is temporarily disabled due to a remaining issue with its support in the Linux kernel.
    • The new RSS implementation does not automatically spread traffic according to the inner packet of VXLAN frames anymore, only the outer one (like other PMDs).
    • Partial (Tx only) support for secondary processes was broken and had to be removed.
    • Refactored driver to get rid of dependency on the components provided by Mellanox OFED and instead rely on the current and public rdma-core package and Linux version from now on.
    • Removed compile-time limitation on number of device instances the PMD can support.
  • Updated mlx5 driver.

    Updated the mlx5 driver including the following changes:

    • Enabled the PMD to run on top of upstream Linux kernel and rdma-core libs, removing the dependency on specific Mellanox OFED libraries.
    • Improved PMD latency performance.
    • Improved PMD memory footprint.
    • Added support for vectorized Rx/Tx burst for ARMv8.
    • Added support for secondary process.
    • Added support for flow counters.
    • Added support for Rx hardware timestamp offload.
    • Added support for device removal event.
  • Added SoftNIC PMD.

    Added a new SoftNIC PMD. This virtual device provides applications with software fallback support for traffic management.

  • Added support for NXP DPAA Devices.

    Added support for NXP’s DPAA devices - LS104x series. This includes:

    • DPAA Bus driver
    • DPAA Mempool driver for supporting offloaded packet memory pool
    • DPAA PMD for DPAA devices

    See the DPAA Poll Mode Driver document for more details of this new driver.

  • Updated support for Cavium OCTEONTX Device.

    Updated support for Cavium’s OCTEONTX device (CN83xx). This includes:

    • OCTEONTX Mempool driver for supporting offloaded packet memory pool
    • OCTEONTX Ethdev PMD
    • OCTEONTX Eventdev-Ethdev Rx adapter

    See the OCTEONTX Poll Mode driver document for more details of this new driver.

  • Added PF support to the Netronome NFP PMD.

    Added PF support to the Netronome NFP PMD. Previously the NFP PMD only supported VFs. PF support is just as a basic DPDK port and has no VF management yet.

    PF support comes with firmware upload support which allows the PMD to independently work from kernel netdev NFP drivers.

    NFP 4000 devices are also now supported along with previous 6000 devices.

  • Updated bnxt PMD.

    Major enhancements include:

    • Support for Flow API
    • Support for Tx and Rx descriptor status functions
  • Added bus agnostic functions to cryptodev for PMD initialization

    Added new PMD assist, bus independent, functions rte_cryptodev_pmd_parse_input_args(), rte_cryptodev_pmd_create() and rte_cryptodev_pmd_destroy() for drivers to manage creation and destruction of new device instances.

  • Updated QAT crypto PMD.

    Added several performance enhancements:

    • Removed atomics from the internal queue pair structure.
    • Added coalesce writes to HEAD CSR on response processing.
    • Added coalesce writes to TAIL CSR on request processing.

    In addition support was added for the AES CCM algorithm.

  • Updated the AESNI MB PMD.

    The AESNI MB PMD has been updated with additional support for:

    • The DES CBC algorithm.
    • The DES DOCSIS BPI algorithm.

    This change requires version 0.47 of the IPSec Multi-buffer library. For more details see the AESN-NI Multi Buffer Crypto Poll Mode Driver documentation.

  • Updated the OpenSSL PMD.

    The OpenSSL PMD has been updated with additional support for:

    • The DES CBC algorithm.
    • The AES CCM algorithm.
  • Added NXP DPAA SEC crypto PMD.

    A new dpaa_sec hardware based crypto PMD for NXP DPAA devices has been added. See the NXP DPAA CAAM (DPAA_SEC) document for more details.

  • Added MRVL crypto PMD.

    A new crypto PMD has been added, which provides several ciphering and hashing algorithms. All cryptography operations use the MUSDK library crypto API. See the MRVL Crypto Poll Mode Driver document for more details.

  • Add new benchmarking mode to dpdk-test-crypto-perf application.

    Added a new “PMD cyclecount” benchmark mode to the dpdk-test-crypto-perf application to display a detailed breakdown of CPU cycles used by hardware acceleration.

  • Added the Security Offload Library.

    Added an experimental library - rte_security. This provide security APIs for protocols like IPSec using inline ipsec offload to ethernet devices or full protocol offload with lookaside crypto devices.

    See the Security Library section of the DPDK Programmers Guide document for more information.

  • Updated the DPAA2_SEC crypto driver to support rte_security.

    Updated the dpaa2_sec crypto PMD to support rte_security lookaside protocol offload for IPSec.

  • Updated the IXGBE ethernet driver to support rte_security.

    Updated ixgbe ethernet PMD to support rte_security inline IPSec offload.

  • Updated ipsec-secgw application to support rte_security.

    Updated the ipsec-secgw sample application to support rte_security actions for ipsec inline and full protocol offload using lookaside crypto offload.

  • Added IOMMU support to libvhost-user

    Implemented device IOTLB in the Vhost-user backend, and enabled Virtio’s IOMMU feature. The feature is disabled by default, and can be enabled by setting RTE_VHOST_USER_IOMMU_SUPPORT flag at vhost device registration time.

  • Added the Event Ethernet Adapter Library.

    Added the Event Ethernet Adapter library. This library provides APIs for eventdev applications to configure the ethdev for eventdev packet flow.

  • Updated DPAA2 Event PMD for the Event Ethernet Adapter.

    Added support for the eventdev ethernet adapter for DPAA2.

  • Added Membership library (rte_member).

    Added a new data structure library called the Membership Library.

    The Membership Library is an extension and generalization of a traditional filter (for example Bloom Filter) structure that has multiple usages in a wide variety of workloads and applications. In general, the Membership Library is a data structure that provides a “set-summary” and responds to set-membership queries whether a certain member belongs to a set(s).

    The library provides APIs for DPDK applications to insert a new member, delete an existing member, and query the existence of a member in a given set, or a group of sets. For the case of a group of sets the library will return not only whether the element has been inserted in one of the sets but also which set it belongs to.

    See the Membership Library documentation in the Programmers Guide, for more information.

  • Added the Generic Segmentation Offload Library.

    Added the Generic Segmentation Offload (GSO) library to enable applications to split large packets (e.g. MTU is 64KB) into small ones (e.g. MTU is 1500B). Supported packet types are:

    • TCP/IPv4 packets.
    • VxLAN packets, which must have an outer IPv4 header, and contain an inner TCP/IPv4 packet.
    • GRE packets, which must contain an outer IPv4 header, and inner TCP/IPv4 headers.

    The GSO library doesn’t check if the input packets have correct checksums, and doesn’t update checksums for output packets. Additionally, the GSO library doesn’t process IP fragmented packets.

  • Added the Flow Classification Library.

    Added an experimental Flow Classification library to provide APIs for DPDK applications to classify an input packet by matching it against a set of flow rules. It uses the librte_table API to manage the flow rules.

Resolved Issues

  • Service core fails to call service callback due to atomic lock

    In a specific configuration of multi-thread unsafe services and service cores, a service core previously did not correctly release the atomic lock on the service. This would result in the cores polling the service, but it looked like another thread was executing the service callback. The logic for atomic locking of the services has been fixed and refactored for readability.

API Changes

  • Ethdev device name length increased.

    The size of internal device name has been increased to 64 characters to allow for storing longer bus specific names.

  • Removed the Ethdev RTE_ETH_DEV_DETACHABLE flag.

    Removed the Ethdev RTE_ETH_DEV_DETACHABLE flag. This flag is not required anymore, with the new hotplug implementation. It has been removed from the ether library. Its semantics are now expressed at the bus and PMD level.

  • Service cores API updated for usability

    The service cores API has been changed, removing pointers from the API where possible, and instead using integer IDs to identify each service. This simplifies application code, aids debugging, and provides better encapsulation. A summary of the main changes made is as follows:

    • Services identified by ID not by rte_service_spec pointer
    • Reduced API surface by using set functions instead of enable/disable
    • Reworked rte_service_register to provide the service ID to registrar
    • Reworked start and stop APIs into rte_service_runstate_set
    • Added API to set runstate of service implementation to indicate readiness
  • The following changes have been made in the mempool library

    • Moved flags datatype from int to unsigned int for rte_mempool.
    • Removed __rte_unused int flag param from rte_mempool_generic_put and rte_mempool_generic_get API.
    • Added flags param in rte_mempool_xmem_size and rte_mempool_xmem_usage.
    • rte_mem_phy2mch was used in Xen dom0 to obtain the physical address; remove this API as Xen dom0 support was removed.
  • Added IOVA aliases related to physical address handling.

    Some data types, structure members and functions related to physical address handling are deprecated and have new aliases with IOVA wording. For example:

    • phys_addr_t can be often replaced by rte_iova_t of same size.
    • RTE_BAD_PHYS_ADDR is often replaced by RTE_BAD_IOVA of same value.
    • rte_memseg.phys_addr is aliased with rte_memseg.iova_addr.
    • rte_mem_virt2phy() can often be replaced by rte_mem_virt2iova.
    • rte_malloc_virt2phy is aliased with rte_malloc_virt2iova.
    • rte_memzone.phys_addr is aliased with rte_memzone.iova.
    • rte_mempool_objhdr.physaddr is aliased with rte_mempool_objhdr.iova.
    • rte_mempool_memhdr.phys_addr is aliased with rte_mempool_memhdr.iova.
    • rte_mempool_virt2phy() can be replaced by rte_mempool_virt2iova().
    • rte_mempool_populate_phys*() are aliased with rte_mempool_populate_iova*()
    • rte_mbuf.buf_physaddr is aliased with rte_mbuf.buf_iova.
    • rte_mbuf_data_dma_addr*() are aliased with rte_mbuf_data_iova*().
    • rte_pktmbuf_mtophys* are aliased with rte_pktmbuf_iova*().
  • PCI bus API moved outside of the EAL

    The PCI bus previously implemented within the EAL has been moved. A first part has been added as an RTE library providing PCI helpers to parse device locations or other such utilities. A second part consisting of the actual bus driver has been moved to its proper subdirectory, without changing its functionalities.

    As such, several PCI-related functions are not exposed by the EAL anymore:

    • rte_pci_detach
    • rte_pci_dump
    • rte_pci_ioport_map
    • rte_pci_ioport_read
    • rte_pci_ioport_unmap
    • rte_pci_ioport_write
    • rte_pci_map_device
    • rte_pci_probe
    • rte_pci_probe_one
    • rte_pci_read_config
    • rte_pci_register
    • rte_pci_scan
    • rte_pci_unmap_device
    • rte_pci_unregister
    • rte_pci_write_config

    These functions are made available either as part of librte_pci or librte_bus_pci.

  • Moved vdev bus APIs outside of the EAL

    Moved the following APIs from librte_eal to librte_bus_vdev:

    • rte_vdev_init
    • rte_vdev_register
    • rte_vdev_uninit
    • rte_vdev_unregister
  • Add return value to stats_get dev op API

    The stats_get dev op API return value has been changed to be int. In this way PMDs can return an error value in case of failure at stats getting process time.

  • Modified the rte_cryptodev_allocate_driver function.

    Modified the rte_cryptodev_allocate_driver() function in the cryptodev library. An extra parameter struct cryptodev_driver *crypto_drv has been added.

  • Removed virtual device bus specific functions from librte_cryptodev.

    The functions rte_cryptodev_vdev_parse_init_params() and rte_cryptodev_vdev_pmd_init() have been removed from librte_cryptodev and have been replaced by non bus specific functions rte_cryptodev_pmd_parse_input_args() and rte_cryptodev_pmd_create().

    The rte_cryptodev_create_vdev() function was removed to avoid the dependency on vdev in librte_cryptodev; instead, users can call rte_vdev_init() directly.

  • Removed PCI device bus specific functions from librte_cryptodev.

    The functions rte_cryptodev_pci_generic_probe() and rte_cryptodev_pci_generic_remove() have been removed from librte_cryptodev and have been replaced by non bus specific functions rte_cryptodev_pmd_create() and rte_cryptodev_pmd_destroy().

  • Removed deprecated functions to manage log level or type.

    The functions rte_set_log_level(), rte_get_log_level(), rte_set_log_type() and rte_get_log_type() have been removed.

    They are respectively replaced by rte_log_set_global_level(), rte_log_get_global_level(), rte_log_set_level() and rte_log_get_level().

  • Removed mbuf flags PKT_RX_VLAN_PKT and PKT_RX_QINQ_PKT.

    The mbuf flags PKT_RX_VLAN_PKT and PKT_RX_QINQ_PKT have been removed since their behavior was not properly described.

  • Added mbuf flags PKT_RX_VLAN and PKT_RX_QINQ.

    Two mbuf flags have been added to indicate that the VLAN identifier has been saved in in the mbuf structure. For instance:

    • If VLAN is not stripped and TCI is saved: PKT_RX_VLAN
    • If VLAN is stripped and TCI is saved: PKT_RX_VLAN | PKT_RX_VLAN_STRIPPED
  • Modified the vlan_offload_set_t function prototype in the ethdev library.

    Modified the vlan_offload_set_t function prototype in the ethdev library. The return value has been changed from void to int so the caller can determine whether the backing device supports the operation or if the operation was successfully performed.

ABI Changes

  • Extended port_id range.

    The size of the field port_id in the rte_eth_dev_data structure has changed, as described in the New Features section above.

  • New parameter added to rte_eth_dev.

    A new parameter security_ctx has been added to rte_eth_dev to support security operations like IPSec inline.

  • New parameter added to rte_cryptodev.

    A new parameter security_ctx has been added to rte_cryptodev to support security operations like lookaside crypto.

Removed Items

  • Xen dom0 in EAL has been removed, as well as the xenvirt PMD and vhost_xen.
  • The crypto performance unit tests have been removed, replaced by the dpdk-test-crypto-perf application.

Shared Library Versions

The libraries prepended with a plus sign were incremented in this version.

  librte_acl.so.2
+ librte_bitratestats.so.2
+ librte_bus_dpaa.so.1
+ librte_bus_fslmc.so.1
+ librte_bus_pci.so.1
+ librte_bus_vdev.so.1
  librte_cfgfile.so.2
  librte_cmdline.so.2
+ librte_cryptodev.so.4
  librte_distributor.so.1
+ librte_eal.so.6
+ librte_ethdev.so.8
+ librte_eventdev.so.3
+ librte_flow_classify.so.1
  librte_gro.so.1
+ librte_gso.so.1
  librte_hash.so.2
  librte_ip_frag.so.1
  librte_jobstats.so.1
  librte_kni.so.2
  librte_kvargs.so.1
  librte_latencystats.so.1
  librte_lpm.so.2
  librte_mbuf.so.3
+ librte_mempool.so.3
  librte_meter.so.1
  librte_metrics.so.1
  librte_net.so.1
+ librte_pci.so.1
+ librte_pdump.so.2
  librte_pipeline.so.3
+ librte_pmd_bnxt.so.2
+ librte_pmd_bond.so.2
+ librte_pmd_i40e.so.2
+ librte_pmd_ixgbe.so.2
  librte_pmd_ring.so.2
+ librte_pmd_softnic.so.1
+ librte_pmd_vhost.so.2
  librte_port.so.3
  librte_power.so.1
  librte_reorder.so.1
  librte_ring.so.1
  librte_sched.so.1
+ librte_security.so.1
+ librte_table.so.3
  librte_timer.so.1
  librte_vhost.so.3

Tested Platforms

  • Intel(R) platforms with Intel(R) NICs combinations

    • CPU
      • Intel(R) Atom(TM) CPU C2758 @ 2.40GHz
      • Intel(R) Xeon(R) CPU D-1540 @ 2.00GHz
      • Intel(R) Xeon(R) CPU D-1541 @ 2.10GHz
      • Intel(R) Xeon(R) CPU E5-4667 v3 @ 2.00GHz
      • Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
      • Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz
      • Intel(R) Xeon(R) CPU E5-2695 v4 @ 2.10GHz
      • Intel(R) Xeon(R) CPU E5-2658 v2 @ 2.40GHz
      • Intel(R) Xeon(R) CPU E5-2658 v3 @ 2.20GHz
    • OS:
      • CentOS 7.2
      • Fedora 25
      • Fedora 26
      • FreeBSD 11
      • Red Hat Enterprise Linux Server release 7.3
      • SUSE Enterprise Linux 12
      • Wind River Linux 8
      • Ubuntu 16.04
      • Ubuntu 16.10
    • NICs:
      • Intel(R) 82599ES 10 Gigabit Ethernet Controller
        • Firmware version: 0x61bf0001
        • Device id (pf/vf): 8086:10fb / 8086:10ed
        • Driver version: 5.2.3 (ixgbe)
      • Intel(R) Corporation Ethernet Connection X552/X557-AT 10GBASE-T
        • Firmware version: 0x800003e7
        • Device id (pf/vf): 8086:15ad / 8086:15a8
        • Driver version: 4.4.6 (ixgbe)
      • Intel(R) Ethernet Converged Network Adapter X710-DA4 (4x10G)
        • Firmware version: 6.01 0x80003205
        • Device id (pf/vf): 8086:1572 / 8086:154c
        • Driver version: 2.1.26 (i40e)
      • Intel(R) Ethernet Converged Network Adapter X710-DA2 (2x10G)
        • Firmware version: 6.01 0x80003204
        • Device id (pf/vf): 8086:1572 / 8086:154c
        • Driver version: 2.1.26 (i40e)
      • Intel(R) Ethernet Converged Network Adapter XXV710-DA2 (2x25G)
        • Firmware version: 6.01 0x80003221
        • Device id (pf/vf): 8086:158b
        • Driver version: 2.1.26 (i40e)
      • Intel(R) Ethernet Converged Network Adapter XL710-QDA2 (2X40G)
        • Firmware version: 6.01 0x8000321c
        • Device id (pf/vf): 8086:1583 / 8086:154c
        • Driver version: 2.1.26 (i40e)
      • Intel(R) Corporation I350 Gigabit Network Connection
        • Firmware version: 1.63, 0x80000dda
        • Device id (pf/vf): 8086:1521 / 8086:1520
        • Driver version: 5.3.0-k (igb)
  • Intel(R) platforms with Mellanox(R) NICs combinations

    • Platform details:
      • Intel(R) Xeon(R) CPU E5-2697A v4 @ 2.60GHz
      • Intel(R) Xeon(R) CPU E5-2697 v3 @ 2.60GHz
      • Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
      • Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz
      • Intel(R) Xeon(R) CPU E5-2640 @ 2.50GHz
      • Intel(R) Xeon(R) CPU E5-2620 v4 @ 2.10GHz
    • OS:
      • Red Hat Enterprise Linux Server release 7.3 (Maipo)
      • Red Hat Enterprise Linux Server release 7.2 (Maipo)
      • Ubuntu 16.10
      • Ubuntu 16.04
      • Ubuntu 14.04
    • MLNX_OFED: 4.2-1.0.0.0
    • NICs:
      • Mellanox(R) ConnectX(R)-3 Pro 40G MCX354A-FCC_Ax (2x40G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1007
        • Firmware version: 2.42.5000
      • Mellanox(R) ConnectX(R)-4 10G MCX4111A-XCAT (1x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.21.1000
      • Mellanox(R) ConnectX(R)-4 10G MCX4121A-XCAT (2x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.21.1000
      • Mellanox(R) ConnectX(R)-4 25G MCX4111A-ACAT (1x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.21.1000
      • Mellanox(R) ConnectX(R)-4 25G MCX4121A-ACAT (2x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.21.1000
      • Mellanox(R) ConnectX(R)-4 40G MCX4131A-BCAT/MCX413A-BCAT (1x40G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.21.1000
      • Mellanox(R) ConnectX(R)-4 40G MCX415A-BCAT (1x40G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.21.1000
      • Mellanox(R) ConnectX(R)-4 50G MCX4131A-GCAT/MCX413A-GCAT (1x50G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.21.1000
      • Mellanox(R) ConnectX(R)-4 50G MCX414A-BCAT (2x50G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.21.1000
      • Mellanox(R) ConnectX(R)-4 50G MCX415A-GCAT/MCX416A-BCAT/MCX416A-GCAT (2x50G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.21.1000
      • Mellanox(R) ConnectX(R)-4 50G MCX415A-CCAT (1x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.21.1000
      • Mellanox(R) ConnectX(R)-4 100G MCX416A-CCAT (2x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.21.1000
      • Mellanox(R) ConnectX(R)-4 Lx 10G MCX4121A-XCAT (2x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1015
        • Firmware version: 14.21.1000
      • Mellanox(R) ConnectX(R)-4 Lx 25G MCX4121A-ACAT (2x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1015
        • Firmware version: 14.21.1000
      • Mellanox(R) ConnectX(R)-5 100G MCX556A-ECAT (2x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1017
        • Firmware version: 16.21.1000
      • Mellanox(R) ConnectX-5 Ex EN 100G MCX516A-CDAT (2x100G)
        • Host interface: PCI Express 4.0 x16
        • Device ID: 15b3:1019
        • Firmware version: 16.21.1000
  • ARM platforms with Mellanox(R) NICs combinations

    • Platform details:
      • Qualcomm ARM 1.1 2500MHz
    • OS:
      • Ubuntu 16.04
    • MLNX_OFED: 4.2-1.0.0.0
    • NICs:
      • Mellanox(R) ConnectX(R)-4 Lx 25G MCX4121A-ACAT (2x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1015
        • Firmware version: 14.21.1000
      • Mellanox(R) ConnectX(R)-5 100G MCX556A-ECAT (2x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1017
        • Firmware version: 16.21.1000

DPDK Release 17.08

New Features

  • Increase minimum x86 ISA version to SSE4.2.

    Starting with version 17.08, DPDK requires SSE4.2 to run on x86. Previous versions required SSE3.

  • Added Service Core functionality.

    The service core functionality added to EAL allows DPDK to run services such as software PMDs on lcores without the application manually running them. The service core infrastructure allows flexibility of running multiple services on the same service lcore, and provides the application with powerful APIs to configure the mapping from service lcores to services.

  • Added Generic Receive Offload API.

    Added Generic Receive Offload (GRO) API support to reassemble TCP/IPv4 packets. The GRO API assumes all input packets have the correct checksums. GRO API doesn’t update checksums for merged packets. If input packets are IP fragmented, the GRO API assumes they are complete packets (i.e. with L4 headers).

  • Added Fail-Safe PMD

    Added the new Fail-Safe PMD. This virtual device allows applications to support seamless hotplug of devices. See the Fail-safe poll mode driver library guide for more details about this driver.

  • Added support for generic flow API (rte_flow) on igb NICs.

    This API provides a generic means of configuring hardware to match specific ingress or egress traffic, altering its behavior and querying related counters according to any number of user-defined rules.

    Added generic flow API support for Ethernet, IPv4, UDP, TCP and RAW pattern items with QUEUE actions. There are four types of filter support for this feature on igb.

  • Added support for generic flow API (rte_flow) on enic.

    Added flow API support for outer Ethernet, VLAN, IPv4, IPv6, UDP, TCP, SCTP, VxLAN and inner Ethernet, VLAN, IPv4, IPv6, UDP and TCP pattern items with QUEUE, MARK, FLAG and VOID actions for ingress traffic.

  • Added support for Chelsio T6 family of adapters

    The CXGBE PMD was updated to run Chelsio T6 family of adapters.

  • Added latency and performance improvements for cxgbe

    the Tx and Rx path in cxgbe were reworked to improve performance. In addition the latency was reduced for slow traffic.

  • Updated the bnxt PMD.

    Updated the bnxt PMD. The major enhancements include:

    • Support MTU modification.
    • Add support for LRO.
    • Add support for VLAN filter and strip functionality.
    • Additional enhancements to add support for more dev_ops.
    • Added PMD specific APIs mainly to control VF from PF.
    • Update HWRM version to 1.7.7
  • Added support for Rx interrupts on mlx4 driver.

    Rx queues can be now be armed with an interrupt which will trigger on the next packet arrival.

  • Updated mlx5 driver.

    Updated the mlx5 driver including the following changes:

    • Added vectorized Rx/Tx burst for x86.
    • Added support for isolated mode from flow API.
    • Reworked the flow drop action to implement in hardware classifier.
    • Improved Rx interrupts management.
  • Updated szedata2 PMD.

    Added support for firmware with multiple Ethernet ports per physical port.

  • Updated dpaa2 PMD.

    Updated dpaa2 PMD. Major enhancements include:

    • Added support for MAC Filter configuration.
    • Added support for Segmented Buffers.
    • Added support for VLAN filter and strip functionality.
    • Additional enhancements to add support for more dev_ops.
    • Optimized the packet receive path
  • Reorganized the symmetric crypto operation structure.

    The crypto operation (rte_crypto_sym_op) has been reorganized as follows:

    • Removed the rte_crypto_sym_op_sess_type field.
    • Replaced the pointer and physical address of IV with offset from the start of the crypto operation.
    • Moved length and offset of cipher IV to rte_crypto_cipher_xform.
    • Removed “Additional Authentication Data” (AAD) length.
    • Removed digest length.
    • Removed AAD pointer and physical address from auth structure.
    • Added aead structure, containing parameters for AEAD algorithms.
  • Reorganized the crypto operation structure.

    The crypto operation (rte_crypto_op) has been reorganized as follows:

    • Added the rte_crypto_op_sess_type field.
    • The enumerations rte_crypto_op_status and rte_crypto_op_type have been modified to be uint8_t values.
    • Removed the field opaque_data.
    • Pointer to rte_crypto_sym_op has been replaced with a zero length array.
  • Reorganized the crypto symmetric session structure.

    The crypto symmetric session structure (rte_cryptodev_sym_session) has been reorganized as follows:

    • The dev_id field has been removed.
    • The driver_id field has been removed.
    • The mempool pointer mp has been removed.
    • Replaced private marker with array of pointers to private data sessions sess_private_data.
  • Updated cryptodev library.

    • Added AEAD algorithm specific functions and structures, so it is not necessary to use a combination of cipher and authentication structures anymore.
    • Added helper functions for crypto device driver identification.
    • Added support for multi-device sessions, so a single session can be used in multiple drivers.
    • Added functions to initialize and free individual driver private data with the same session.
  • Updated dpaa2_sec crypto PMD.

    Added support for AES-GCM and AES-CTR.

  • Updated the AESNI MB PMD.

    The AESNI MB PMD has been updated with additional support for:

    • 12-byte IV on AES Counter Mode, apart from the previous 16-byte IV.
  • Updated the AES-NI GCM PMD.

    The AES-NI GCM PMD was migrated from the ISA-L library to the Multi Buffer library, as the latter library has Scatter Gather List support now. The migration entailed adding additional support for 192-bit keys.

  • Updated the Cryptodev Scheduler PMD.

    Added a multicore based distribution mode, which distributes the enqueued crypto operations among several slaves, running on different logical cores.

  • Added NXP DPAA2 Eventdev PMD.

    Added the new dpaa2 eventdev driver for NXP DPAA2 devices. See the “Event Device Drivers” document for more details on this new driver.

  • Added dpdk-test-eventdev test application.

    The dpdk-test-eventdev tool is a Data Plane Development Kit (DPDK) application that allows exercising various eventdev use cases. This application has a generic framework to add new eventdev based test cases to verify functionality and measure the performance parameters of DPDK eventdev devices.

Known Issues

  • Starting with version 17.08, libnuma is required to build DPDK.

API Changes

  • Modified the _rte_eth_dev_callback_process function in the ethdev library.

    The function _rte_eth_dev_callback_process() has been modified. The return value has been changed from void to int and an extra parameter void *ret_param has been added.

  • Moved bypass functions from the rte_ethdev library to ixgbe PMD

    • The following rte_ethdev library functions were removed:
      • rte_eth_dev_bypass_event_show()
      • rte_eth_dev_bypass_event_store()
      • rte_eth_dev_bypass_init()
      • rte_eth_dev_bypass_state_set()
      • rte_eth_dev_bypass_state_show()
      • rte_eth_dev_bypass_ver_show()
      • rte_eth_dev_bypass_wd_reset()
      • rte_eth_dev_bypass_wd_timeout_show()
      • rte_eth_dev_wd_timeout_store()
    • The following ixgbe PMD functions were added:
      • rte_pmd_ixgbe_bypass_event_show()
      • rte_pmd_ixgbe_bypass_event_store()
      • rte_pmd_ixgbe_bypass_init()
      • rte_pmd_ixgbe_bypass_state_set()
      • rte_pmd_ixgbe_bypass_state_show()
      • rte_pmd_ixgbe_bypass_ver_show()
      • rte_pmd_ixgbe_bypass_wd_reset()
      • rte_pmd_ixgbe_bypass_wd_timeout_show()
      • rte_pmd_ixgbe_bypass_wd_timeout_store()
  • Reworked rte_cryptodev library.

    The rte_cryptodev library has been reworked and updated. The following changes have been made to it:

    • The crypto device type enumeration has been removed from cryptodev library.
    • The function rte_crypto_count_devtype() has been removed, and replaced by the new function rte_crypto_count_by_driver().
    • Moved crypto device driver names definitions to the particular PMDs. These names are not public anymore.
    • The rte_cryptodev_configure() function does not create the session mempool for the device anymore.
    • The rte_cryptodev_queue_pair_attach_sym_session() and rte_cryptodev_queue_pair_dettach_sym_session() functions require the new parameter device id.
    • Parameters of rte_cryptodev_sym_session_create() were modified to accept mempool, instead of device id and rte_crypto_sym_xform.
    • Removed device id parameter from rte_cryptodev_sym_session_free().
    • Added a new field session_pool to rte_cryptodev_queue_pair_setup().
    • Removed aad_size parameter from rte_cryptodev_sym_capability_check_auth().
    • Added iv_size parameter to rte_cryptodev_sym_capability_check_auth().
    • Removed RTE_CRYPTO_OP_STATUS_ENQUEUED from enum rte_crypto_op_status.

ABI Changes

  • Changed type of domain field in rte_pci_addr to uint32_t to follow the PCI standard.

  • Added new rte_bus experimental APIs available as operators within the rte_bus structure.

  • Made rte_devargs structure internal device representation generic to prepare for a bus-agnostic EAL.

  • Reorganized the crypto operation structures.

    Some fields have been modified in the rte_crypto_op and rte_crypto_sym_op structures, as described in the New Features section.

  • Reorganized the crypto symmetric session structure.

    Some fields have been modified in the rte_cryptodev_sym_session structure, as described in the New Features section.

  • Reorganized the rte_crypto_sym_cipher_xform structure.

    • Added cipher IV length and offset parameters.
    • Changed field size of key length from size_t to uint16_t.
  • Reorganized the rte_crypto_sym_auth_xform structure.

    • Added authentication IV length and offset parameters.
    • Changed field size of AAD length from uint32_t to uint16_t.
    • Changed field size of digest length from uint32_t to uint16_t.
    • Removed AAD length.
    • Changed field size of key length from size_t to uint16_t.
  • Replaced dev_type enumeration with uint8_t driver_id in rte_cryptodev_info and rte_cryptodev structures.

  • Removed session_mp from rte_cryptodev_config.

Shared Library Versions

The libraries prepended with a plus sign were incremented in this version.

  librte_acl.so.2
  librte_bitratestats.so.1
  librte_cfgfile.so.2
  librte_cmdline.so.2
+ librte_cryptodev.so.3
  librte_distributor.so.1
+ librte_eal.so.5
+ librte_ethdev.so.7
+ librte_eventdev.so.2
+ librte_gro.so.1
  librte_hash.so.2
  librte_ip_frag.so.1
  librte_jobstats.so.1
  librte_kni.so.2
  librte_kvargs.so.1
  librte_latencystats.so.1
  librte_lpm.so.2
  librte_mbuf.so.3
  librte_mempool.so.2
  librte_meter.so.1
  librte_metrics.so.1
  librte_net.so.1
  librte_pdump.so.1
  librte_pipeline.so.3
  librte_pmd_bond.so.1
  librte_pmd_ring.so.2
  librte_port.so.3
  librte_power.so.1
  librte_reorder.so.1
  librte_ring.so.1
  librte_sched.so.1
  librte_table.so.2
  librte_timer.so.1
  librte_vhost.so.3

Tested Platforms

  • Intel(R) platforms with Mellanox(R) NICs combinations

    • Platform details:
      • Intel(R) Xeon(R) CPU E5-2697A v4 @ 2.60GHz
      • Intel(R) Xeon(R) CPU E5-2697 v3 @ 2.60GHz
      • Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
      • Intel(R) Xeon(R) CPU E5-2640 @ 2.50GHz
    • OS:
      • Red Hat Enterprise Linux Server release 7.3 (Maipo)
      • Red Hat Enterprise Linux Server release 7.2 (Maipo)
      • Ubuntu 16.10
      • Ubuntu 16.04
      • Ubuntu 14.04
    • MLNX_OFED: 4.1-1.0.2.0
    • NICs:
      • Mellanox(R) ConnectX(R)-3 Pro 40G MCX354A-FCC_Ax (2x40G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1007
        • Firmware version: 2.40.5030
      • Mellanox(R) ConnectX(R)-4 10G MCX4111A-XCAT (1x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 10G MCX4121A-XCAT (2x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 25G MCX4111A-ACAT (1x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 25G MCX4121A-ACAT (2x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 40G MCX4131A-BCAT/MCX413A-BCAT (1x40G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 40G MCX415A-BCAT (1x40G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 50G MCX4131A-GCAT/MCX413A-GCAT (1x50G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 50G MCX414A-BCAT (2x50G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 50G MCX415A-GCAT/MCX416A-BCAT/MCX416A-GCAT (2x50G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 50G MCX415A-CCAT (1x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 100G MCX416A-CCAT (2x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 Lx 10G MCX4121A-XCAT (2x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1015
        • Firmware version: 14.18.2000
      • Mellanox(R) ConnectX(R)-4 Lx 25G MCX4121A-ACAT (2x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1015
        • Firmware version: 14.18.2000
      • Mellanox(R) ConnectX(R)-5 100G MCX556A-ECAT (2x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1017
        • Firmware version: 16.19.1200
      • Mellanox(R) ConnectX-5 Ex EN 100G MCX516A-CDAT (2x100G)
        • Host interface: PCI Express 4.0 x16
        • Device ID: 15b3:1019
        • Firmware version: 16.19.1200
  • Intel(R) platforms with Intel(R) NICs combinations

    • CPU
      • Intel(R) Atom(TM) CPU C2758 @ 2.40GHz
      • Intel(R) Xeon(R) CPU D-1540 @ 2.00GHz
      • Intel(R) Xeon(R) CPU D-1541 @ 2.10GHz
      • Intel(R) Xeon(R) CPU E5-4667 v3 @ 2.00GHz
      • Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
      • Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz
      • Intel(R) Xeon(R) CPU E5-2695 v4 @ 2.10GHz
      • Intel(R) Xeon(R) CPU E5-2658 v2 @ 2.40GHz
      • Intel(R) Xeon(R) CPU E5-2658 v3 @ 2.20GHz
    • OS:
      • CentOS 7.2
      • Fedora 25
      • FreeBSD 11
      • Red Hat Enterprise Linux Server release 7.3
      • SUSE Enterprise Linux 12
      • Wind River Linux 8
      • Ubuntu 16.04
      • Ubuntu 16.10
    • NICs:
      • Intel(R) 82599ES 10 Gigabit Ethernet Controller
        • Firmware version: 0x61bf0001
        • Device id (pf/vf): 8086:10fb / 8086:10ed
        • Driver version: 4.0.1-k (ixgbe)
      • Intel(R) Corporation Ethernet Connection X552/X557-AT 10GBASE-T
        • Firmware version: 0x800001cf
        • Device id (pf/vf): 8086:15ad / 8086:15a8
        • Driver version: 4.2.5 (ixgbe)
      • Intel(R) Ethernet Converged Network Adapter X710-DA4 (4x10G)
        • Firmware version: 6.01 0x80003205
        • Device id (pf/vf): 8086:1572 / 8086:154c
        • Driver version: 2.0.19 (i40e)
      • Intel(R) Ethernet Converged Network Adapter X710-DA2 (2x10G)
        • Firmware version: 6.01 0x80003204
        • Device id (pf/vf): 8086:1572 / 8086:154c
        • Driver version: 2.0.19 (i40e)
      • Intel(R) Ethernet Converged Network Adapter XXV710-DA2 (2x25G)
        • Firmware version: 6.01 0x80003221
        • Device id (pf/vf): 8086:158b
        • Driver version: 2.0.19 (i40e)
      • Intel(R) Ethernet Converged Network Adapter XL710-QDA2 (2X40G)
        • Firmware version: 6.01 0x8000321c
        • Device id (pf/vf): 8086:1583 / 8086:154c
        • Driver version: 2.0.19 (i40e)
      • Intel(R) Corporation I350 Gigabit Network Connection
        • Firmware version: 1.48, 0x800006e7
        • Device id (pf/vf): 8086:1521 / 8086:1520
        • Driver version: 5.2.13-k (igb)

DPDK Release 17.05

New Features

  • Reorganized mbuf structure.

    The mbuf structure has been reorganized as follows:

    • Align fields to facilitate the writing of data_off, refcnt, and nb_segs in one operation.
    • Use 2 bytes for port and number of segments.
    • Move the sequence number to the second cache line.
    • Add a timestamp field.
    • Set default value for refcnt, next and nb_segs at mbuf free.
  • Added mbuf raw free API.

    Moved rte_mbuf_raw_free() and rte_pktmbuf_prefree_seg() functions to the public API.

  • Added free Tx mbuf on demand API.

    Added a new function rte_eth_tx_done_cleanup() which allows an application to request the driver to release mbufs that are no longer in use from a Tx ring, independent of whether or not the tx_rs_thresh has been crossed.

  • Added device removal interrupt.

    Added a new ethdev event RTE_ETH_DEV_INTR_RMV to signify the sudden removal of a device. This event can be advertised by PCI drivers and enabled accordingly.

  • Added EAL dynamic log framework.

    Added new APIs to dynamically register named log types, and control the level of each type independently.

  • Added descriptor status ethdev API.

    Added a new API to get the status of a descriptor.

    For Rx, it is almost similar to the rx_descriptor_done API, except it differentiates descriptors which are held by the driver and not returned to the hardware. For Tx, it is a new API.

  • Increased number of next hops for LPM IPv6 to 2^21.

    The next_hop field has been extended from 8 bits to 21 bits for IPv6.

  • Added VFIO hotplug support.

    Added hotplug support for VFIO in addition to the existing UIO support.

  • Added PowerPC support to pci probing for vfio-pci devices.

    Enabled sPAPR IOMMU based pci probing for vfio-pci devices.

  • Kept consistent PMD batching behavior.

    Removed the limit of fm10k/i40e/ixgbe Tx burst size and vhost Rx/Tx burst size in order to support the same policy of “make an best effort to Rx/Tx pkts” for PMDs.

  • Updated the ixgbe base driver.

    Updated the ixgbe base driver, including the following changes:

    • Add link block check for KR.
    • Complete HW initialization even if SFP is not present.
    • Add VF xcast promiscuous mode.
  • Added PowerPC support for i40e and its vector PMD.

    Enabled i40e PMD and its vector PMD by default in PowerPC.

  • Added VF max bandwidth setting in i40e.

    Enabled capability to set the max bandwidth for a VF in i40e.

  • Added VF TC min and max bandwidth setting in i40e.

    Enabled capability to set the min and max allocated bandwidth for a TC on a VF in i40.

  • Added TC strict priority mode setting on i40e.

    There are 2 Tx scheduling modes supported for TCs by i40e HW: round robin mode and strict priority mode. By default the round robin mode is used. It is now possible to change the Tx scheduling mode for a TC. This is a global setting on a physical port.

  • Added i40e dynamic device personalization support.

    • Added dynamic device personalization processing to i40e firmware.
  • Added Cloud Filter for QinQ steering to i40e.

    • Added a QinQ cloud filter on the i40e PMD, for steering traffic to a VM using both VLAN tags. Note, this feature is not supported in Vector Mode.
  • Updated mlx5 PMD.

    Updated the mlx5 driver, including the following changes:

    • Added Generic flow API support for classification according to ether type.
    • Extended Generic flow API support for classification of IPv6 flow according to Vtc flow, Protocol and Hop limit.
    • Added Generic flow API support for FLAG action.
    • Added Generic flow API support for RSS action.
    • Added support for TSO for non-tunneled and VXLAN packets.
    • Added support for hardware Tx checksum offloads for VXLAN packets.
    • Added support for user space Rx interrupt mode.
    • Improved ConnectX-5 single core and maximum performance.
  • Updated mlx4 PMD.

    Updated the mlx4 driver, including the following changes:

    • Added support for Generic flow API basic flow items and actions.
    • Added support for device removal event.
  • Updated the sfc_efx driver.

    • Added Generic Flow API support for Ethernet, VLAN, IPv4, IPv6, UDP and TCP pattern items with QUEUE action for ingress traffic.
    • Added support for virtual functions (VFs).
  • Added LiquidIO network PMD.

    Added poll mode driver support for Cavium LiquidIO II server adapter VFs.

  • Added Atomic Rules Arkville PMD.

    Added a new poll mode driver for the Arkville family of devices from Atomic Rules. The net/ark PMD supports line-rate agnostic, multi-queue data movement on Arkville core FPGA instances.

  • Added support for NXP DPAA2 - FSLMC bus.

    Added the new bus “fslmc” driver for NXP DPAA2 devices. See the “Network Interface Controller Drivers” document for more details of this new driver.

  • Added support for NXP DPAA2 Network PMD.

    Added the new “dpaa2” net driver for NXP DPAA2 devices. See the “Network Interface Controller Drivers” document for more details of this new driver.

  • Added support for the Wind River Systems AVP PMD.

    Added a new networking driver for the AVP device type. Theses devices are specific to the Wind River Systems virtualization platforms.

  • Added vmxnet3 version 3 support.

    Added support for vmxnet3 version 3 which includes several performance enhancements such as configurable Tx data ring, Receive Data Ring, and the ability to register memory regions.

  • Updated the TAP driver.

    Updated the TAP PMD to:

    • Support MTU modification.
    • Support packet type for Rx.
    • Support segmented packets on Rx and Tx.
    • Speed up Rx on TAP when no packets are available.
    • Support capturing traffic from another netdevice.
    • Dynamically change link status when the underlying interface state changes.
    • Added Generic Flow API support for Ethernet, VLAN, IPv4, IPv6, UDP and TCP pattern items with DROP, QUEUE and PASSTHRU actions for ingress traffic.
  • Added MTU feature support to Virtio and Vhost.

    Implemented new Virtio MTU feature in Vhost and Virtio:

    • Add rte_vhost_mtu_get() API to Vhost library.
    • Enable Vhost PMD’s MTU get feature.
    • Get max MTU value from host in Virtio PMD
  • Added interrupt mode support for virtio-user.

    Implemented Rxq interrupt mode and LSC support for virtio-user as a virtual device. Supported cases:

    • Rxq interrupt for virtio-user + vhost-user as the backend.
    • Rxq interrupt for virtio-user + vhost-kernel as the backend.
    • LSC interrupt for virtio-user + vhost-user as the backend.
  • Added event driven programming model library (rte_eventdev).

    This API introduces an event driven programming model.

    In a polling model, lcores poll ethdev ports and associated Rx queues directly to look for a packet. By contrast in an event driven model, lcores call the scheduler that selects packets for them based on programmer-specified criteria. The Eventdev library adds support for an event driven programming model, which offers applications automatic multicore scaling, dynamic load balancing, pipelining, packet ingress order maintenance and synchronization services to simplify application packet processing.

    By introducing an event driven programming model, DPDK can support both polling and event driven programming models for packet processing, and applications are free to choose whatever model (or combination of the two) best suits their needs.

  • Added Software Eventdev PMD.

    Added support for the software eventdev PMD. The software eventdev is a software based scheduler device that implements the eventdev API. This PMD allows an application to configure a pipeline using the eventdev library, and run the scheduling workload on a CPU core.

  • Added Cavium OCTEONTX Eventdev PMD.

    Added the new octeontx ssovf eventdev driver for OCTEONTX devices. See the “Event Device Drivers” document for more details on this new driver.

  • Added information metrics library.

    Added a library that allows information metrics to be added and updated by producers, typically other libraries, for later retrieval by consumers such as applications. It is intended to provide a reporting mechanism that is independent of other libraries such as ethdev.

  • Added bit-rate calculation library.

    Added a library that can be used to calculate device bit-rates. Calculated bitrates are reported using the metrics library.

  • Added latency stats library.

    Added a library that measures packet latency. The collected statistics are jitter and latency. For latency the minimum, average, and maximum is measured.

  • Added NXP DPAA2 SEC crypto PMD.

    A new “dpaa2_sec” hardware based crypto PMD for NXP DPAA2 devices has been added. See the “Crypto Device Drivers” document for more details on this driver.

  • Updated the Cryptodev Scheduler PMD.

    • Added a packet-size based distribution mode, which distributes the enqueued crypto operations among two slaves, based on their data lengths.
    • Added fail-over scheduling mode, which enqueues crypto operations to a primary slave first. Then, any operation that cannot be enqueued is enqueued to a secondary slave.
    • Added mode specific option support, so each scheduling mode can now be configured individually by the new API.
  • Updated the QAT PMD.

    The QAT PMD has been updated with additional support for:

    • AES DOCSIS BPI algorithm.
    • DES DOCSIS BPI algorithm.
    • ZUC EEA3/EIA3 algorithms.
  • Updated the AESNI MB PMD.

    The AESNI MB PMD has been updated with additional support for:

    • AES DOCSIS BPI algorithm.
  • Updated the OpenSSL PMD.

    The OpenSSL PMD has been updated with additional support for:

    • DES DOCSIS BPI algorithm.

Resolved Issues

  • l2fwd-keepalive: Fixed unclean shutdowns.

    Added clean shutdown to l2fwd-keepalive so that it can free up stale resources used for inter-process communication.

Known Issues

  • LSC interrupt doesn’t work for virtio-user + vhost-kernel.

    LSC interrupt cannot be detected when setting the backend, tap device, up/down as we fail to find a way to monitor such event.

API Changes

  • The LPM next_hop field is extended from 8 bits to 21 bits for IPv6 while keeping ABI compatibility.

  • Reworked rte_ring library.

    The rte_ring library has been reworked and updated. The following changes have been made to it:

    • Removed the build-time setting CONFIG_RTE_RING_SPLIT_PROD_CONS.

    • Removed the build-time setting CONFIG_RTE_LIBRTE_RING_DEBUG.

    • Removed the build-time setting CONFIG_RTE_RING_PAUSE_REP_COUNT.

    • Removed the function rte_ring_set_water_mark as part of a general removal of watermarks support in the library.

    • Added an extra parameter to the burst/bulk enqueue functions to return the number of free spaces in the ring after enqueue. This can be used by an application to implement its own watermark functionality.

    • Added an extra parameter to the burst/bulk dequeue functions to return the number elements remaining in the ring after dequeue.

    • Changed the return value of the enqueue and dequeue bulk functions to match that of the burst equivalents. In all cases, ring functions which operate on multiple packets now return the number of elements enqueued or dequeued, as appropriate. The updated functions are:

      • rte_ring_mp_enqueue_bulk
      • rte_ring_sp_enqueue_bulk
      • rte_ring_enqueue_bulk
      • rte_ring_mc_dequeue_bulk
      • rte_ring_sc_dequeue_bulk
      • rte_ring_dequeue_bulk

      NOTE: the above functions all have different parameters as well as different return values, due to the other listed changes above. This means that all instances of the functions in existing code will be flagged by the compiler. The return value usage should be checked while fixing the compiler error due to the extra parameter.

  • Reworked rte_vhost library.

    The rte_vhost library has been reworked to make it generic enough so that the user could build other vhost-user drivers on top of it. To achieve this the following changes have been made:

    • The following vhost-pmd APIs are removed:
      • rte_eth_vhost_feature_disable
      • rte_eth_vhost_feature_enable
      • rte_eth_vhost_feature_get
    • The vhost API rte_vhost_driver_callback_register(ops) is reworked to be per vhost-user socket file. Thus, it takes one more argument: rte_vhost_driver_callback_register(path, ops).
    • The vhost API rte_vhost_get_queue_num is deprecated, instead, rte_vhost_get_vring_num should be used.
    • The following macros are removed in rte_virtio_net.h
      • VIRTIO_RXQ
      • VIRTIO_TXQ
      • VIRTIO_QNUM
    • The following net specific header files are removed in rte_virtio_net.h
      • linux/virtio_net.h
      • sys/socket.h
      • linux/if.h
      • rte_ether.h
    • The vhost struct virtio_net_device_ops is renamed to vhost_device_ops
    • The vhost API rte_vhost_driver_session_start is removed. Instead, rte_vhost_driver_start should be used, and there is no need to create a thread to call it.
    • The vhost public header file rte_virtio_net.h is renamed to rte_vhost.h

ABI Changes

  • Reorganized the mbuf structure.

    The order and size of the fields in the mbuf structure changed, as described in the New Features section.

  • The rte_cryptodev_info.sym structure has a new field max_nb_sessions_per_qp to support drivers which may support a limited number of sessions per queue_pair.

Removed Items

  • KNI vhost support has been removed.
  • The dpdk_qat sample application has been removed.

Shared Library Versions

The libraries prepended with a plus sign were incremented in this version.

  librte_acl.so.2
+ librte_bitratestats.so.1
  librte_cfgfile.so.2
  librte_cmdline.so.2
  librte_cryptodev.so.2
  librte_distributor.so.1
+ librte_eal.so.4
  librte_ethdev.so.6
+ librte_eventdev.so.1
  librte_hash.so.2
  librte_ip_frag.so.1
  librte_jobstats.so.1
  librte_kni.so.2
  librte_kvargs.so.1
+ librte_latencystats.so.1
  librte_lpm.so.2
+ librte_mbuf.so.3
  librte_mempool.so.2
  librte_meter.so.1
+ librte_metrics.so.1
  librte_net.so.1
  librte_pdump.so.1
  librte_pipeline.so.3
  librte_pmd_bond.so.1
  librte_pmd_ring.so.2
  librte_port.so.3
  librte_power.so.1
  librte_reorder.so.1
  librte_ring.so.1
  librte_sched.so.1
  librte_table.so.2
  librte_timer.so.1
  librte_vhost.so.3

Tested Platforms

  • Intel(R) platforms with Intel(R) NICs combinations

    • CPU
      • Intel(R) Atom(TM) CPU C2758 @ 2.40GHz
      • Intel(R) Xeon(R) CPU D-1540 @ 2.00GHz
      • Intel(R) Xeon(R) CPU E5-4667 v3 @ 2.00GHz
      • Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
      • Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz
      • Intel(R) Xeon(R) CPU E5-2695 v4 @ 2.10GHz
      • Intel(R) Xeon(R) CPU E5-2658 v2 @ 2.40GHz
      • Intel(R) Xeon(R) CPU E5-2658 v3 @ 2.20GHz
    • OS:
      • CentOS 7.2
      • Fedora 25
      • FreeBSD 11
      • Red Hat Enterprise Linux Server release 7.3
      • SUSE Enterprise Linux 12
      • Wind River Linux 8
      • Ubuntu 16.04
      • Ubuntu 16.10
    • NICs:
      • Intel(R) 82599ES 10 Gigabit Ethernet Controller
        • Firmware version: 0x61bf0001
        • Device id (pf/vf): 8086:10fb / 8086:10ed
        • Driver version: 4.0.1-k (ixgbe)
      • Intel(R) Corporation Ethernet Connection X552/X557-AT 10GBASE-T
        • Firmware version: 0x800001cf
        • Device id (pf/vf): 8086:15ad / 8086:15a8
        • Driver version: 4.2.5 (ixgbe)
      • Intel(R) Ethernet Converged Network Adapter X710-DA4 (4x10G)
        • Firmware version: 5.05
        • Device id (pf/vf): 8086:1572 / 8086:154c
        • Driver version: 1.5.23 (i40e)
      • Intel(R) Ethernet Converged Network Adapter X710-DA2 (2x10G)
        • Firmware version: 5.05
        • Device id (pf/vf): 8086:1572 / 8086:154c
        • Driver version: 1.5.23 (i40e)
      • Intel(R) Ethernet Converged Network Adapter XL710-QDA1 (1x40G)
        • Firmware version: 5.05
        • Device id (pf/vf): 8086:1584 / 8086:154c
        • Driver version: 1.5.23 (i40e)
      • Intel(R) Ethernet Converged Network Adapter XL710-QDA2 (2X40G)
        • Firmware version: 5.05
        • Device id (pf/vf): 8086:1583 / 8086:154c
        • Driver version: 1.5.23 (i40e)
      • Intel(R) Corporation I350 Gigabit Network Connection
        • Firmware version: 1.48, 0x800006e7
        • Device id (pf/vf): 8086:1521 / 8086:1520
        • Driver version: 5.2.13-k (igb)
  • Intel(R) platforms with Mellanox(R) NICs combinations

    • Platform details:
      • Intel(R) Xeon(R) CPU E5-2697A v4 @ 2.60GHz
      • Intel(R) Xeon(R) CPU E5-2697 v3 @ 2.60GHz
      • Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
      • Intel(R) Xeon(R) CPU E5-2640 @ 2.50GHz
    • OS:
      • Red Hat Enterprise Linux Server release 7.3 (Maipo)
      • Red Hat Enterprise Linux Server release 7.2 (Maipo)
      • Ubuntu 16.10
      • Ubuntu 16.04
      • Ubuntu 14.04
    • MLNX_OFED: 4.0-2.0.0.0
    • NICs:
      • Mellanox(R) ConnectX(R)-3 Pro 40G MCX354A-FCC_Ax (2x40G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1007
        • Firmware version: 2.40.5030
      • Mellanox(R) ConnectX(R)-4 10G MCX4111A-XCAT (1x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 10G MCX4121A-XCAT (2x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 25G MCX4111A-ACAT (1x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 25G MCX4121A-ACAT (2x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 40G MCX4131A-BCAT/MCX413A-BCAT (1x40G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 40G MCX415A-BCAT (1x40G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 50G MCX4131A-GCAT/MCX413A-GCAT (1x50G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 50G MCX414A-BCAT (2x50G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 50G MCX415A-GCAT/MCX416A-BCAT/MCX416A-GCAT (2x50G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 50G MCX415A-CCAT (1x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 100G MCX416A-CCAT (2x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 Lx 10G MCX4121A-XCAT (2x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1015
        • Firmware version: 14.18.2000
      • Mellanox(R) ConnectX(R)-4 Lx 25G MCX4121A-ACAT (2x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1015
        • Firmware version: 14.18.2000
      • Mellanox(R) ConnectX(R)-5 100G MCX556A-ECAT (2x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1017
        • Firmware version: 16.19.1200
      • Mellanox(R) ConnectX-5 Ex EN 100G MCX516A-CDAT (2x100G)
        • Host interface: PCI Express 4.0 x16
        • Device ID: 15b3:1019
        • Firmware version: 16.19.1200
  • IBM(R) Power8(R) with Mellanox(R) NICs combinations

    • Platform details:

      • Processor: POWER8E (raw), AltiVec supported
      • type-model: 8247-22L
      • Firmware FW810.21 (SV810_108)
    • OS: Ubuntu 16.04 LTS PPC le

    • MLNX_OFED: 4.0-2.0.0.0

    • NICs:

      • Mellanox(R) ConnectX(R)-4 10G MCX4111A-XCAT (1x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 10G MCX4121A-XCAT (2x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 25G MCX4111A-ACAT (1x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 25G MCX4121A-ACAT (2x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 40G MCX4131A-BCAT/MCX413A-BCAT (1x40G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 40G MCX415A-BCAT (1x40G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 50G MCX4131A-GCAT/MCX413A-GCAT (1x50G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 50G MCX414A-BCAT (2x50G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 50G MCX415A-GCAT/MCX416A-BCAT/MCX416A-GCAT (2x50G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 50G MCX415A-CCAT (1x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000
      • Mellanox(R) ConnectX(R)-4 100G MCX416A-CCAT (2x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.2000

DPDK Release 17.02

New Features

  • Added support for representing buses in EAL

    The rte_bus structure was introduced into the EAL. This allows for devices to be represented by buses they are connected to. A new bus can be added to DPDK by extending the rte_bus structure and implementing the scan and probe functions. Once a new bus is registered using the provided APIs, new devices can be detected and initialized using bus scan and probe callbacks.

    With this change, devices other than PCI or VDEV type can be represented in the DPDK framework.

  • Added generic EAL API for I/O device memory read/write operations.

    This API introduces 8 bit, 16 bit, 32 bit and 64 bit I/O device memory read/write operations along with “relaxed” versions.

    Weakly-ordered architectures like ARM need an additional I/O barrier for device memory read/write access over PCI bus. By introducing the EAL abstraction for I/O device memory read/write access, the drivers can access I/O device memory in an architecture-agnostic manner. The relaxed version does not have an additional I/O memory barrier, which is useful in accessing the device registers of integrated controllers which is implicitly strongly ordered with respect to memory access.

  • Added generic flow API (rte_flow).

    This API provides a generic means to configure hardware to match specific ingress or egress traffic, alter its behavior and query related counters according to any number of user-defined rules.

    In order to expose a single interface with an unambiguous behavior that is common to all poll-mode drivers (PMDs) the rte_flow API is slightly higher-level than the legacy filtering framework, which it encompasses and supersedes (including all functions and filter types) .

    See the Generic flow API documentation for more information.

  • Added firmware version get API.

    Added a new function rte_eth_dev_fw_version_get() to fetch the firmware version for a given device.

  • Added APIs for MACsec offload support to the ixgbe PMD.

    Six new APIs have been added to the ixgbe PMD for MACsec offload support. The declarations for the APIs can be found in rte_pmd_ixgbe.h.

  • Added I219 NICs support.

    Added support for I219 Intel 1GbE NICs.

  • Added VF Daemon (VFD) for i40e. - EXPERIMENTAL

    This is an EXPERIMENTAL feature to enhance the capability of the DPDK PF as many VF management features are not currently supported by the kernel PF driver. Some new private APIs are implemented directly in the PMD without an abstraction layer. They can be used directly by some users who have the need.

    The new APIs to control VFs directly from PF include:

    • Set VF MAC anti-spoofing.
    • Set VF VLAN anti-spoofing.
    • Set TX loopback.
    • Set VF unicast promiscuous mode.
    • Set VF multicast promiscuous mode.
    • Set VF MTU.
    • Get/reset VF stats.
    • Set VF MAC address.
    • Set VF VLAN stripping.
    • Vf VLAN insertion.
    • Set VF broadcast mode.
    • Set VF VLAN tag.
    • Set VF VLAN filter.

    VFD also includes VF to PF mailbox message management from an application. When the PF receives mailbox messages from the VF the PF should call the callback provided by the application to know if they’re permitted to be processed.

    As an EXPERIMENTAL feature, please be aware it can be changed or even removed without prior notice.

  • Updated the i40e base driver.

    Updated the i40e base driver, including the following changes:

    • Replace existing legacy memcpy() calls with i40e_memcpy() calls.
    • Use BIT() macro instead of bit fields.
    • Add clear all WoL filters implementation.
    • Add broadcast promiscuous control per VLAN.
    • Remove unused X722_SUPPORT and I40E_NDIS_SUPPORT macros.
  • Updated the enic driver.

    • Set new Rx checksum flags in mbufs to indicate unknown, good or bad checksums.
    • Fix set/remove of MAC addresses. Allow up to 64 addresses per device.
    • Enable TSO on outer headers.
  • Added Solarflare libefx-based network PMD.

    Added a new network PMD which supports Solarflare SFN7xxx and SFN8xxx family of 10/40 Gbps adapters.

  • Updated the mlx4 driver.

    • Addressed a few bugs.
  • Added support for Mellanox ConnectX-5 adapters (mlx5).

    Added support for Mellanox ConnectX-5 family of 10/25/40/50/100 Gbps adapters to the existing mlx5 PMD.

  • Updated the mlx5 driver.

    • Improve Tx performance by using vector logic.
    • Improve RSS balancing when number of queues is not a power of two.
    • Generic flow API support for Ethernet, IPv4, IPv4, UDP, TCP, VLAN and VXLAN pattern items with DROP and QUEUE actions.
    • Support for extended statistics.
    • Addressed several data path bugs.
    • As of MLNX_OFED 4.0-1.0.1.0, the Toeplitz RSS hash function is not symmetric anymore for consistency with other PMDs.
  • virtio-user with vhost-kernel as another exceptional path.

    Previously, we upstreamed a virtual device, virtio-user with vhost-user as the backend as a way of enabling IPC (Inter-Process Communication) and user space container networking.

    Virtio-user with vhost-kernel as the backend is a solution for the exception path, such as KNI, which exchanges packets with the kernel networking stack. This solution is very promising in:

    • Maintenance: vhost and vhost-net (kernel) is an upstreamed and extensively used kernel module.
    • Features: vhost-net is designed to be a networking solution, which has lots of networking related features, like multi-queue, TSO, multi-seg mbuf, etc.
    • Performance: similar to KNI, this solution would use one or more kthreads to send/receive packets from user space DPDK applications, which has little impact on user space polling thread (except that it might enter into kernel space to wake up those kthreads if necessary).
  • Added virtio Rx interrupt support.

    Added a feature to enable Rx interrupt mode for virtio pci net devices as bound to VFIO (noiommu mode) and driven by virtio PMD.

    With this feature, the virtio PMD can switch between polling mode and interrupt mode, to achieve best performance, and at the same time save power. It can work on both legacy and modern virtio devices. In this mode, each rxq is mapped with an excluded MSIx interrupt.

    See the Virtio Interrupt Mode documentation for more information.

  • Added ARMv8 crypto PMD.

    A new crypto PMD has been added, which provides combined mode cryptographic operations optimized for ARMv8 processors. The driver can be used to enhance performance in processing chained operations such as cipher + HMAC.

  • Updated the QAT PMD.

    The QAT PMD has been updated with additional support for:

    • DES algorithm.
    • Scatter-gather list (SGL) support.
  • Updated the AESNI MB PMD.

    • The Intel(R) Multi Buffer Crypto for IPsec library used in AESNI MB PMD has been moved to a new repository, in GitHub.
    • Support has been added for single operations (cipher only and authentication only).
  • Updated the AES-NI GCM PMD.

    The AES-NI GCM PMD was migrated from the Multi Buffer library to the ISA-L library. The migration entailed adding additional support for:

    • GMAC algorithm.
    • 256-bit cipher key.
    • Session-less mode.
    • Out-of place processing
    • Scatter-gather support for chained mbufs (only out-of place and destination mbuf must be contiguous)
  • Added crypto performance test application.

    Added a new performance test application for measuring performance parameters of PMDs available in the crypto tree.

  • Added Elastic Flow Distributor library (rte_efd).

    Added a new library which uses perfect hashing to determine a target/value for a given incoming flow key.

    The library does not store the key itself for lookup operations, and therefore, lookup performance is not dependent on the key size. Also, the target/value can be any arbitrary value (8 bits by default). Finally, the storage requirement is much smaller than a hash-based flow table and therefore, it can better fit in CPU cache and scale to millions of flow keys.

    See the Elastic Flow Distributor Library documentation in the Programmers Guide document, for more information.

Resolved Issues

Drivers
  • net/virtio: Fixed multiple process support.

    Fixed a few regressions introduced in recent releases that break the virtio multiple process support.

Examples
  • examples/ethtool: Fixed crash with non-PCI devices.

    Fixed issue where querying a non-PCI device was dereferencing non-existent PCI data resulting in a segmentation fault.

API Changes

  • Moved five APIs for VF management from the ethdev to the ixgbe PMD.

    The following five APIs for VF management from the PF have been removed from the ethdev, renamed, and added to the ixgbe PMD:

    rte_eth_dev_set_vf_rate_limit()
    rte_eth_dev_set_vf_rx()
    rte_eth_dev_set_vf_rxmode()
    rte_eth_dev_set_vf_tx()
    rte_eth_dev_set_vf_vlan_filter()
    

    The API’s have been renamed to the following:

    rte_pmd_ixgbe_set_vf_rate_limit()
    rte_pmd_ixgbe_set_vf_rx()
    rte_pmd_ixgbe_set_vf_rxmode()
    rte_pmd_ixgbe_set_vf_tx()
    rte_pmd_ixgbe_set_vf_vlan_filter()
    

    The declarations for the API’s can be found in rte_pmd_ixgbe.h.

ABI Changes

Shared Library Versions

The libraries prepended with a plus sign were incremented in this version.

  librte_acl.so.2
  librte_cfgfile.so.2
  librte_cmdline.so.2
  librte_cryptodev.so.2
  librte_distributor.so.1
  librte_eal.so.3
+ librte_ethdev.so.6
  librte_hash.so.2
  librte_ip_frag.so.1
  librte_jobstats.so.1
  librte_kni.so.2
  librte_kvargs.so.1
  librte_lpm.so.2
  librte_mbuf.so.2
  librte_mempool.so.2
  librte_meter.so.1
  librte_net.so.1
  librte_pdump.so.1
  librte_pipeline.so.3
  librte_pmd_bond.so.1
  librte_pmd_ring.so.2
  librte_port.so.3
  librte_power.so.1
  librte_reorder.so.1
  librte_ring.so.1
  librte_sched.so.1
  librte_table.so.2
  librte_timer.so.1
  librte_vhost.so.3

Tested Platforms

This release has been tested with the below list of CPU/device/firmware/OS. Each section describes a different set of combinations.

  • Intel(R) platforms with Mellanox(R) NICs combinations

    • Platform details
      • Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz
      • Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
      • Intel(R) Xeon(R) CPU E5-2697 v3 @ 2.60GHz
    • OS:
      • CentOS 7.0
      • Fedora 23
      • Fedora 24
      • FreeBSD 10.3
      • Red Hat Enterprise Linux 7.2
      • SUSE Enterprise Linux 12
      • Ubuntu 14.04 LTS
      • Ubuntu 15.10
      • Ubuntu 16.04 LTS
      • Wind River Linux 8
    • MLNX_OFED: 4.0-1.0.1.0
    • NICs:
      • Mellanox(R) ConnectX(R)-3 Pro 40G MCX354A-FCC_Ax (2x40G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1007
        • Firmware version: 2.40.5030
      • Mellanox(R) ConnectX(R)-4 10G MCX4111A-XCAT (1x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 10G MCX4121A-XCAT (2x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 25G MCX4111A-ACAT (1x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 25G MCX4121A-ACAT (2x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 40G MCX4131A-BCAT/MCX413A-BCAT (1x40G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 40G MCX415A-BCAT (1x40G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 50G MCX4131A-GCAT/MCX413A-GCAT (1x50G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 50G MCX414A-BCAT (2x50G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 50G MCX415A-GCAT/MCX416A-BCAT/MCX416A-GCAT (2x50G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 50G MCX415A-CCAT (1x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 100G MCX416A-CCAT (2x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 Lx 10G MCX4121A-XCAT (2x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1015
        • Firmware version: 14.18.1000
      • Mellanox(R) ConnectX(R)-4 Lx 25G MCX4121A-ACAT (2x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1015
        • Firmware version: 14.18.1000
      • Mellanox(R) ConnectX(R)-5 100G MCX556A-ECAT (2x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1017
        • Firmware version: 16.18.1000
      • Mellanox(R) ConnectX-5 Ex EN 100G MCX516A-CDAT (2x100G)
        • Host interface: PCI Express 4.0 x16
        • Device ID: 15b3:1019
        • Firmware version: 16.18.1000
  • IBM(R) Power8(R) with Mellanox(R) NICs combinations

    • Machine:
      • Processor: POWER8E (raw), AltiVec supported
        • type-model: 8247-22L
        • Firmware FW810.21 (SV810_108)
    • OS: Ubuntu 16.04 LTS PPC le
    • MLNX_OFED: 4.0-1.0.1.0
    • NICs:
      • Mellanox(R) ConnectX(R)-4 10G MCX4111A-XCAT (1x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 10G MCX4121A-XCAT (2x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 25G MCX4111A-ACAT (1x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 25G MCX4121A-ACAT (2x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 40G MCX4131A-BCAT/MCX413A-BCAT (1x40G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 40G MCX415A-BCAT (1x40G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 50G MCX4131A-GCAT/MCX413A-GCAT (1x50G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 50G MCX414A-BCAT (2x50G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 50G MCX415A-GCAT/MCX416A-BCAT/MCX416A-GCAT (2x50G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 50G MCX415A-CCAT (1x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 100G MCX416A-CCAT (2x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1013
        • Firmware version: 12.18.1000
      • Mellanox(R) ConnectX(R)-4 Lx 10G MCX4121A-XCAT (2x10G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1015
        • Firmware version: 14.18.1000
      • Mellanox(R) ConnectX(R)-4 Lx 25G MCX4121A-ACAT (2x25G)
        • Host interface: PCI Express 3.0 x8
        • Device ID: 15b3:1015
        • Firmware version: 14.18.1000
      • Mellanox(R) ConnectX(R)-5 100G MCX556A-ECAT (2x100G)
        • Host interface: PCI Express 3.0 x16
        • Device ID: 15b3:1017
        • Firmware version: 16.18.1000
  • Intel(R) platforms with Intel(R) NICs combinations

    • Platform details
      • Intel(R) Atom(TM) CPU C2758 @ 2.40GHz
      • Intel(R) Xeon(R) CPU D-1540 @ 2.00GHz
      • Intel(R) Xeon(R) CPU E5-4667 v3 @ 2.00GHz
      • Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
      • Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz
      • Intel(R) Xeon(R) CPU E5-2695 v4 @ 2.10GHz
      • Intel(R) Xeon(R) CPU E5-2658 v2 @ 2.40GHz
    • OS:
      • CentOS 7.2
      • Fedora 25
      • FreeBSD 11
      • Red Hat Enterprise Linux Server release 7.3
      • SUSE Enterprise Linux 12
      • Wind River Linux 8
      • Ubuntu 16.04
      • Ubuntu 16.10
    • NICs:
      • Intel(R) 82599ES 10 Gigabit Ethernet Controller
        • Firmware version: 0x61bf0001
        • Device id (pf/vf): 8086:10fb / 8086:10ed
        • Driver version: 4.0.1-k (ixgbe)
      • Intel(R) Corporation Ethernet Connection X552/X557-AT 10GBASE-T
        • Firmware version: 0x800001cf
        • Device id (pf/vf): 8086:15ad / 8086:15a8
        • Driver version: 4.2.5 (ixgbe)
      • Intel(R) Ethernet Converged Network Adapter X710-DA4 (4x10G)
        • Firmware version: 5.05
        • Device id (pf/vf): 8086:1572 / 8086:154c
        • Driver version: 1.5.23 (i40e)
      • Intel(R) Ethernet Converged Network Adapter X710-DA2 (2x10G)
        • Firmware version: 5.05
        • Device id (pf/vf): 8086:1572 / 8086:154c
        • Driver version: 1.5.23 (i40e)
      • Intel(R) Ethernet Converged Network Adapter XL710-QDA1 (1x40G)
        • Firmware version: 5.05
        • Device id (pf/vf): 8086:1584 / 8086:154c
        • Driver version: 1.5.23 (i40e)
      • Intel(R) Ethernet Converged Network Adapter XL710-QDA2 (2X40G)
        • Firmware version: 5.05
        • Device id (pf/vf): 8086:1583 / 8086:154c
        • Driver version: 1.5.23 (i40e)
      • Intel(R) Corporation I350 Gigabit Network Connection
        • Firmware version: 1.48, 0x800006e7
        • Device id (pf/vf): 8086:1521 / 8086:1520
        • Driver version: 5.2.13-k (igb)

DPDK Release 16.11

New Features

  • Added software parser for packet type.

    • Added a new function rte_pktmbuf_read() to read the packet data from an mbuf chain, linearizing if required.
    • Added a new function rte_net_get_ptype() to parse an Ethernet packet in an mbuf chain and retrieve its packet type from software.
    • Added new functions rte_get_ptype_*() to dump a packet type as a string.
  • Improved offloads support in mbuf.

    • Added a new function rte_raw_cksum_mbuf() to process the checksum of data embedded in an mbuf chain.
    • Added new Rx checksum flags in mbufs to describe more states: unknown, good, bad, or not present (useful for virtual drivers). This modification was done for IP and L4.
    • Added a new Rx LRO mbuf flag, used when packets are coalesced. This flag indicates that the segment size of original packets is known.
  • Added vhost-user dequeue zero copy support.

    The copy in the dequeue path is avoided in order to improve the performance. In the VM2VM case, the boost is quite impressive. The bigger the packet size, the bigger performance boost you may get. However, for the VM2NIC case, there are some limitations, so the boost is not as impressive as the VM2VM case. It may even drop quite a bit for small packets.

    For that reason, this feature is disabled by default. It can be enabled when the RTE_VHOST_USER_DEQUEUE_ZERO_COPY flag is set. Check the VHost section of the Programming Guide for more information.

  • Added vhost-user indirect descriptors support.

    If the indirect descriptor feature is enabled, each packet sent by the guest will take exactly one slot in the enqueue virtqueue. Without this feature, as in the current version, even 64 bytes packets take two slots with Virtio PMD on guest side.

    The main impact is better performance for 0% packet loss use-cases, as it behaves as if the virtqueue size was enlarged, so more packets can be buffered in the case of system perturbations. On the downside, small performance degradations were measured when running micro-benchmarks.

  • Added vhost PMD xstats.

    Added extended statistics to vhost PMD from a per port perspective.

  • Supported offloads with virtio.

    Added support for the following offloads in virtio:

    • Rx/Tx checksums.
    • LRO.
    • TSO.
  • Added virtio NEON support for ARM.

    Added NEON support for ARM based virtio.

  • Updated the ixgbe base driver.

    Updated the ixgbe base driver, including the following changes:

    • Added X550em_a 10G PHY support.
    • Added support for flow control auto negotiation for X550em_a 1G PHY.
    • Added X550em_a FW ALEF support.
    • Increased mailbox version to ixgbe_mbox_api_13.
    • Added two MAC operations for Hyper-V support.
  • Added APIs for VF management to the ixgbe PMD.

    Eight new APIs have been added to the ixgbe PMD for VF management from the PF. The declarations for the API’s can be found in rte_pmd_ixgbe.h.

  • Updated the enic driver.

    • Added update to use interrupt for link status checking instead of polling.
    • Added more flow director modes on UCS Blade with firmware version >= 2.0(13e).
    • Added full support for MTU update.
    • Added support for the rte_eth_rx_queue_count function.
  • Updated the mlx5 driver.

    • Added support for RSS hash results.
    • Added several performance improvements.
    • Added several bug fixes.
  • Updated the QAT PMD.

    The QAT PMD was updated with additional support for:

    • MD5_HMAC algorithm.
    • SHA224-HMAC algorithm.
    • SHA384-HMAC algorithm.
    • GMAC algorithm.
    • KASUMI (F8 and F9) algorithm.
    • 3DES algorithm.
    • NULL algorithm.
    • C3XXX device.
    • C62XX device.
  • Added openssl PMD.

    A new crypto PMD has been added, which provides several ciphering and hashing algorithms. All cryptography operations use the Openssl library crypto API.

  • Updated the IPsec example.

    Updated the IPsec example with the following support:

    • Configuration file support.
    • AES CBC IV generation with cipher forward function.
    • AES GCM/CTR mode.
  • Added support for new gcc -march option.

    The GCC 4.9 -march option supports the Intel processor code names. The config option RTE_MACHINE can be used to pass code names to the compiler via the -march flag.

Resolved Issues

Drivers
  • enic: Fixed several flow director issues.
  • enic: Fixed inadvertent setting of L4 checksum ptype on ICMP packets.
  • enic: Fixed high driver overhead when servicing Rx queues beyond the first.

Known Issues

  • L3fwd-power app does not work properly when Rx vector is enabled.

    The L3fwd-power app doesn’t work properly with some drivers in vector mode since the queue monitoring works differently between scalar and vector modes leading to incorrect frequency scaling. In addition, L3fwd-power application requires the mbuf to have correct packet type set but in some drivers the vector mode must be disabled for this.

    Therefore, in order to use L3fwd-power, vector mode should be disabled via the config file.

  • Digest address must be supplied for crypto auth operation on QAT PMD.

    The cryptodev API specifies that if the rte_crypto_sym_op.digest.data field, and by inference the digest.phys_addr field which points to the same location, is not set for an auth operation the driver is to understand that the digest result is located immediately following the region over which the digest is computed. The QAT PMD doesn’t correctly handle this case and reads and writes to an incorrect location.

    Callers can workaround this by always supplying the digest virtual and physical address fields in the rte_crypto_sym_op for an auth operation.

API Changes

  • The driver naming convention has been changed to make them more consistent. It especially impacts --vdev arguments. For example eth_pcap becomes net_pcap and cryptodev_aesni_mb_pmd becomes crypto_aesni_mb.

    For backward compatibility an alias feature has been enabled to support the original names.

  • The log history has been removed.

  • The rte_ivshmem feature (including library and EAL code) has been removed in 16.11 because it had some design issues which were not planned to be fixed.

  • The file_name data type of struct rte_port_source_params and struct rte_port_sink_params is changed from char * to const char *.

  • Improved device/driver hierarchy and generalized hotplugging.

    The device and driver relationship has been restructured by introducing generic classes. This paves the way for having PCI, VDEV and other device types as instantiated objects rather than classes in themselves. Hotplugging has also been generalized into EAL so that Ethernet or crypto devices can use the common infrastructure.

    • Removed pmd_type as a way of segregation of devices.
    • Moved numa_node and devargs into rte_driver from rte_pci_driver. These can now be used by any instantiated object of rte_driver.
    • Added rte_device class and all PCI and VDEV devices inherit from it
    • Renamed devinit/devuninit handlers to probe/remove to make it more semantically correct with respect to the device <=> driver relationship.
    • Moved hotplugging support to EAL. Hereafter, PCI and vdev can use the APIs rte_eal_dev_attach and rte_eal_dev_detach.
    • Renamed helpers and support macros to make them more synonymous with their device types (e.g. PMD_REGISTER_DRIVER => RTE_PMD_REGISTER_PCI).
    • Device naming functions have been generalized from ethdev and cryptodev to EAL. rte_eal_pci_device_name has been introduced for obtaining unique device name from PCI Domain-BDF description.
    • Virtual device registration APIs have been added: rte_eal_vdrv_register and rte_eal_vdrv_unregister.

ABI Changes

Shared Library Versions

The libraries prepended with a plus sign were incremented in this version.

  librte_acl.so.2
  librte_cfgfile.so.2
  librte_cmdline.so.2
+ librte_cryptodev.so.2
  librte_distributor.so.1
+ librte_eal.so.3
+ librte_ethdev.so.5
  librte_hash.so.2
  librte_ip_frag.so.1
  librte_jobstats.so.1
  librte_kni.so.2
  librte_kvargs.so.1
  librte_lpm.so.2
  librte_mbuf.so.2
  librte_mempool.so.2
  librte_meter.so.1
  librte_net.so.1
  librte_pdump.so.1
  librte_pipeline.so.3
  librte_pmd_bond.so.1
  librte_pmd_ring.so.2
  librte_port.so.3
  librte_power.so.1
  librte_reorder.so.1
  librte_ring.so.1
  librte_sched.so.1
  librte_table.so.2
  librte_timer.so.1
  librte_vhost.so.3

Tested Platforms

  1. SuperMicro 1U
    • BIOS: 1.0c
    • Processor: Intel(R) Atom(TM) CPU C2758 @ 2.40GHz
  2. SuperMicro 1U
    • BIOS: 1.0a
    • Processor: Intel(R) Xeon(R) CPU D-1540 @ 2.00GHz
    • Onboard NIC: Intel(R) X552/X557-AT (2x10G)
      • Firmware-version: 0x800001cf
      • Device ID (PF/VF): 8086:15ad /8086:15a8
    • kernel driver version: 4.2.5 (ixgbe)
  3. SuperMicro 2U
    • BIOS: 1.0a
    • Processor: Intel(R) Xeon(R) CPU E5-4667 v3 @ 2.00GHz
  4. Intel(R) Server board S2600GZ
    • BIOS: SE5C600.86B.02.02.0002.122320131210
    • Processor: Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
  5. Intel(R) Server board W2600CR
    • BIOS: SE5C600.86B.02.01.0002.082220131453
    • Processor: Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
  6. Intel(R) Server board S2600CWT
    • BIOS: SE5C610.86B.01.01.0009.060120151350
    • Processor: Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz
  7. Intel(R) Server board S2600WTT
    • BIOS: SE5C610.86B.01.01.0005.101720141054
    • Processor: Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz
  8. Intel(R) Server board S2600WTT
    • BIOS: SE5C610.86B.11.01.0044.090120151156
    • Processor: Intel(R) Xeon(R) CPU E5-2695 v4 @ 2.10GHz
  9. Intel(R) Server board S2600WTT
    • Processor: Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz
  10. Intel(R) Server
    • Intel(R) Xeon(R) CPU E5-2697 v3 @ 2.60GHz
  11. IBM(R) Power8(R)
    • Machine type-model: 8247-22L
    • Firmware FW810.21 (SV810_108)
    • Processor: POWER8E (raw), AltiVec supported

Tested NICs

  1. Intel(R) Ethernet Controller X540-AT2
    • Firmware version: 0x80000389
    • Device id (pf): 8086:1528
    • Driver version: 3.23.2 (ixgbe)
  2. Intel(R) 82599ES 10 Gigabit Ethernet Controller
    • Firmware version: 0x61bf0001
    • Device id (pf/vf): 8086:10fb / 8086:10ed
    • Driver version: 4.0.1-k (ixgbe)
  3. Intel(R) Corporation Ethernet Connection X552/X557-AT 10GBASE-T
    • Firmware version: 0x800001cf
    • Device id (pf/vf): 8086:15ad / 8086:15a8
    • Driver version: 4.2.5 (ixgbe)
  4. Intel(R) Ethernet Converged Network Adapter X710-DA4 (4x10G)
    • Firmware version: 5.05
    • Device id (pf/vf): 8086:1572 / 8086:154c
    • Driver version: 1.5.23 (i40e)
  5. Intel(R) Ethernet Converged Network Adapter X710-DA2 (2x10G)
    • Firmware version: 5.05
    • Device id (pf/vf): 8086:1572 / 8086:154c
    • Driver version: 1.5.23 (i40e)
  6. Intel(R) Ethernet Converged Network Adapter XL710-QDA1 (1x40G)
    • Firmware version: 5.05
    • Device id (pf/vf): 8086:1584 / 8086:154c
    • Driver version: 1.5.23 (i40e)
  7. Intel(R) Ethernet Converged Network Adapter XL710-QDA2 (2X40G)
    • Firmware version: 5.05
    • Device id (pf/vf): 8086:1583 / 8086:154c
    • Driver version: 1.5.23 (i40e)
  8. Intel(R) Corporation I350 Gigabit Network Connection
    • Firmware version: 1.48, 0x800006e7
    • Device id (pf/vf): 8086:1521 / 8086:1520
    • Driver version: 5.2.13-k (igb)
  9. Intel(R) Ethernet Multi-host Controller FM10000
    • Firmware version: N/A
    • Device id (pf/vf): 8086:15d0
    • Driver version: 0.17.0.9 (fm10k)
  10. Mellanox(R) ConnectX(R)-4 10G MCX4111A-XCAT (1x10G)
    • Host interface: PCI Express 3.0 x8
    • Device ID: 15b3:1013
    • MLNX_OFED: 3.4-1.0.0.0
    • Firmware version: 12.17.1010
  11. Mellanox(R) ConnectX(R)-4 10G MCX4121A-XCAT (2x10G)
    • Host interface: PCI Express 3.0 x8
    • Device ID: 15b3:1013
    • MLNX_OFED: 3.4-1.0.0.0
    • Firmware version: 12.17.1010
  12. Mellanox(R) ConnectX(R)-4 25G MCX4111A-ACAT (1x25G)
    • Host interface: PCI Express 3.0 x8
    • Device ID: 15b3:1013
    • MLNX_OFED: 3.4-1.0.0.0
    • Firmware version: 12.17.1010
  13. Mellanox(R) ConnectX(R)-4 25G MCX4121A-ACAT (2x25G)
    • Host interface: PCI Express 3.0 x8
    • Device ID: 15b3:1013
    • MLNX_OFED: 3.4-1.0.0.0
    • Firmware version: 12.17.1010
  14. Mellanox(R) ConnectX(R)-4 40G MCX4131A-BCAT/MCX413A-BCAT (1x40G)
    • Host interface: PCI Express 3.0 x8
    • Device ID: 15b3:1013
    • MLNX_OFED: 3.4-1.0.0.0
    • Firmware version: 12.17.1010
  15. Mellanox(R) ConnectX(R)-4 40G MCX415A-BCAT (1x40G)
    • Host interface: PCI Express 3.0 x16
    • Device ID: 15b3:1013
    • MLNX_OFED: 3.4-1.0.0.0
    • Firmware version: 12.17.1010
  16. Mellanox(R) ConnectX(R)-4 50G MCX4131A-GCAT/MCX413A-GCAT (1x50G)
    • Host interface: PCI Express 3.0 x8
    • Device ID: 15b3:1013
    • MLNX_OFED: 3.4-1.0.0.0
    • Firmware version: 12.17.1010
  17. Mellanox(R) ConnectX(R)-4 50G MCX414A-BCAT (2x50G)
    • Host interface: PCI Express 3.0 x8
    • Device ID: 15b3:1013
    • MLNX_OFED: 3.4-1.0.0.0
    • Firmware version: 12.17.1010
  18. Mellanox(R) ConnectX(R)-4 50G MCX415A-GCAT/MCX416A-BCAT/MCX416A-GCAT (2x50G)
    • Host interface: PCI Express 3.0 x16
    • Device ID: 15b3:1013
    • MLNX_OFED: 3.4-1.0.0.0
    • Firmware version: 12.17.1010
  19. Mellanox(R) ConnectX(R)-4 50G MCX415A-CCAT (1x100G)
    • Host interface: PCI Express 3.0 x16
    • Device ID: 15b3:1013
    • MLNX_OFED: 3.4-1.0.0.0
    • Firmware version: 12.17.1010
  20. Mellanox(R) ConnectX(R)-4 100G MCX416A-CCAT (2x100G)
    • Host interface: PCI Express 3.0 x16
    • Device ID: 15b3:1013
    • MLNX_OFED: 3.4-1.0.0.0
    • Firmware version: 12.17.1010
  21. Mellanox(R) ConnectX(R)-4 Lx 10G MCX4121A-XCAT (2x10G)
    • Host interface: PCI Express 3.0 x8
    • Device ID: 15b3:1015
    • MLNX_OFED: 3.4-1.0.0.0
    • Firmware version: 14.17.1010
  22. Mellanox(R) ConnectX(R)-4 Lx 25G MCX4121A-ACAT (2x25G)
    • Host interface: PCI Express 3.0 x8
    • Device ID: 15b3:1015
    • MLNX_OFED: 3.4-1.0.0.0
    • Firmware version: 14.17.1010

Tested OSes

  • CentOS 7.2
  • Fedora 23
  • Fedora 24
  • FreeBSD 10.3
  • FreeBSD 11
  • Red Hat Enterprise Linux Server release 6.7 (Santiago)
  • Red Hat Enterprise Linux Server release 7.0 (Maipo)
  • Red Hat Enterprise Linux Server release 7.2 (Maipo)
  • SUSE Enterprise Linux 12
  • Wind River Linux 6.0.0.26
  • Wind River Linux 8
  • Ubuntu 14.04
  • Ubuntu 15.04
  • Ubuntu 16.04

DPDK Release 16.07

New Features

  • Removed the mempool cache memory if caching is not being used.

    The size of the mempool structure is reduced if the per-lcore cache is disabled.

  • Added mempool external cache for non-EAL thread.

    Added new functions to create, free or flush a user-owned mempool cache for non-EAL threads. Previously the caching was always disabled on these threads.

  • Changed the memory allocation scheme in the mempool library.

    • Added the ability to allocate a large mempool in fragmented virtual memory.
    • Added new APIs to populate a mempool with memory.
    • Added an API to free a mempool.
    • Modified the API of the rte_mempool_obj_iter() function.
    • Dropped the specific Xen Dom0 code.
    • Dropped the specific anonymous mempool code in testpmd.
  • Added a new driver for Broadcom NetXtreme-C devices.

    Added the new bnxt driver for Broadcom NetXtreme-C devices. See the “Network Interface Controller Drivers” document for more details on this new driver.

  • Added a new driver for ThunderX nicvf devices.

    Added the new thunderx net driver for ThunderX nicvf devices. See the “Network Interface Controller Drivers” document for more details on this new driver.

  • Added mailbox interrupt support for ixgbe and igb VFs.

    When the physical NIC link comes up or down, the PF driver will send a mailbox message to notify each VF. To handle this link up/down event, support have been added for a mailbox interrupt to receive the message and allow the application to register a callback for it.

  • Updated the ixgbe base driver.

    The ixgbe base driver was updated with changes including the following:

    • Added sgmii link for X550.
    • Added MAC link setup for X550a SFP and SFP+.
    • Added KR support for X550em_a.
    • Added new PHY definitions for M88E1500.
    • Added support for the VLVF to be bypassed when adding/removing a VFTA entry.
    • Added X550a flow control auto negotiation support.
  • Updated the i40e base driver.

    Updated the i40e base driver including support for new devices IDs.

  • Updated the enic driver.

    The enic driver was updated with changes including the following:

    • Optimized the Tx function.
    • Added Scattered Rx capability.
    • Improved packet type identification.
    • Added MTU update in non Scattered Rx mode and enabled MTU of up to 9208 with UCS Software release 2.2 on 1300 series VICs.
  • Updated the mlx5 driver.

    The mlx5 driver was updated with changes including the following:

    • Data path was refactored to bypass Verbs to improve RX and TX performance.
    • Removed compilation parameters for inline send, MLX5_MAX_INLINE, and added command line parameter instead, txq_inline.
    • Improved TX scatter gather support: Removed compilation parameter MLX5_PMD_SGE_WR_N. Scatter-gather elements is set to the maximum value the NIC supports. Removed linearization logic, this decreases the memory consumption of the PMD.
    • Improved jumbo frames support, by dynamically setting RX scatter gather elements according to the MTU and mbuf size, no need for compilation parameter MLX5_PMD_SGE_WR_N
  • Added support for virtio on IBM POWER8.

    The ioports are mapped in memory when using Linux UIO.

  • Added support for Virtio in containers.

    Add a new virtual device, named virtio_user, to support virtio for containers.

    Known limitations:

    • Control queue and multi-queue are not supported yet.
    • Doesn’t work with --huge-unlink.
    • Doesn’t work with --no-huge.
    • Doesn’t work when there are more than VHOST_MEMORY_MAX_NREGIONS(8) hugepages.
    • Root privilege is required for sorting hugepages by physical address.
    • Can only be used with the vhost user backend.
  • Added vhost-user client mode.

    DPDK vhost-user now supports client mode as well as server mode. Client mode is enabled when the RTE_VHOST_USER_CLIENT flag is set while calling rte_vhost_driver_register.

    When DPDK vhost-user restarts from an normal or abnormal exit (such as a crash), the client mode allows DPDK to establish the connection again. Note that QEMU version v2.7 or above is required for this feature.

    DPDK vhost-user will also try to reconnect by default when:

    • The first connect fails (for example when QEMU is not started yet).
    • The connection is broken (for example when QEMU restarts).

    It can be turned off by setting the RTE_VHOST_USER_NO_RECONNECT flag.

  • Added NSH packet recognition in i40e.

  • Added AES-CTR support to AESNI MB PMD.

    Now AESNI MB PMD supports 128/192/256-bit counter mode AES encryption and decryption.

  • Added AES counter mode support for Intel QuickAssist devices.

    Enabled support for the AES CTR algorithm for Intel QuickAssist devices. Provided support for algorithm-chaining operations.

  • Added KASUMI SW PMD.

    A new Crypto PMD has been added, which provides KASUMI F8 (UEA1) ciphering and KASUMI F9 (UIA1) hashing.

  • Added multi-writer support for RTE Hash with Intel TSX.

    The following features/modifications have been added to rte_hash library:

    • Enabled application developers to use an extra flag for rte_hash creation to specify default behavior (multi-thread safe/unsafe) with the rte_hash_add_key function.
    • Changed the Cuckoo Hash Search algorithm to breadth first search for multi-writer routines and split Cuckoo Hash Search and Move operations in order to reduce transactional code region and improve TSX performance.
    • Added a hash multi-writer test case to the test app.
  • Improved IP Pipeline Application.

    The following features have been added to the ip_pipeline application:

    • Configure the MAC address in the routing pipeline and automatic route updates with change in link state.
    • Enable RSS per network interface through the configuration file.
    • Streamline the CLI code.
  • Added keepalive enhancements.

    Added support for reporting of core states other than “dead” to monitoring applications, enabling the support of broader liveness reporting to external processes.

  • Added packet capture framework.

    • A new library librte_pdump is added to provide a packet capture API.
    • A new app/pdump tool is added to demonstrate capture packets in DPDK.
  • Added floating VEB support for i40e PF driver.

    A “floating VEB” is a special Virtual Ethernet Bridge (VEB) which does not have an upload port, but instead is used for switching traffic between virtual functions (VFs) on a port.

    For information on this feature, please see the “I40E Poll Mode Driver” section of the “Network Interface Controller Drivers” document.

  • Added support for live migration of a VM with SRIOV VF.

    Live migration of a VM with Virtio and VF PMD’s using the bonding PMD.

Resolved Issues

EAL
  • igb_uio: Fixed possible mmap failure for Linux >= 4.5.

    The mmaping of the iomem range of the PCI device fails for kernels that enabled the CONFIG_IO_STRICT_DEVMEM option. The error seen by the user is as similar to the following:

    EAL: pci_map_resource():
    
        cannot mmap(39, 0x7f1c51800000, 0x100000, 0x0):
        Invalid argument (0xffffffffffffffff)
    

    The CONFIG_IO_STRICT_DEVMEM kernel option was introduced in Linux v4.5.

    The issues was resolve by updating igb_uio to stop reserving PCI memory resources. From the kernel point of view the iomem region looks like idle and mmap works again. This matches the uio_pci_generic usage.

Drivers
  • i40e: Fixed vlan stripping from inner header.

    Previously, for tunnel packets, such as VXLAN/NVGRE, the vlan tags of the inner header will be stripped without putting vlan info to descriptor. Now this issue is fixed by disabling vlan stripping from inner header.

  • i40e: Fixed the type issue of a single VLAN type.

    Currently, if a single VLAN header is added in a packet, it’s treated as inner VLAN. But generally, a single VLAN header is treated as the outer VLAN header. This issue is fixed by changing corresponding register for single VLAN.

  • enic: Fixed several issues when stopping then restarting ports and queues.

    Fixed several crashes related to stopping then restarting ports and queues. Fixed possible crash when re-configuring the number of Rx queue descriptors.

  • enic: Fixed Rx data mis-alignment if mbuf data offset modified.

    Fixed possible Rx corruption when mbufs were returned to a pool with data offset other than RTE_PKTMBUF_HEADROOM.

  • enic: Fixed Tx IP/UDP/TCP checksum offload and VLAN insertion.

  • enic: Fixed Rx error and missed counters.

Libraries
  • mbuf: Fixed refcnt update when detaching.

    Fix the rte_pktmbuf_detach() function to decrement the direct mbuf’s reference counter. The previous behavior was not to affect the reference counter. This lead to a memory leak of the direct mbuf.

Examples
Other

Known Issues

API Changes

  • The following counters are removed from the rte_eth_stats structure:
    • ibadcrc
    • ibadlen
    • imcasts
    • fdirmatch
    • fdirmiss
    • tx_pause_xon
    • rx_pause_xon
    • tx_pause_xoff
    • rx_pause_xoff
  • The extended statistics are fetched by ids with rte_eth_xstats_get after a lookup by name rte_eth_xstats_get_names.
  • The function rte_eth_dev_info_get fill the new fields nb_rx_queues and nb_tx_queues in the structure rte_eth_dev_info.
  • The vhost function rte_vring_available_entries is renamed to rte_vhost_avail_entries.
  • All existing vhost APIs and callbacks with virtio_net struct pointer as the parameter have been changed due to the ABI refactoring described below. It is replaced by int vid.
  • The function rte_vhost_enqueue_burst no longer supports concurrent enqueuing packets to the same queue.
  • The function rte_eth_dev_set_mtu adds a new return value -EBUSY, which indicates the operation is forbidden because the port is running.
  • The script dpdk_nic_bind.py is renamed to dpdk-devbind.py. And the script setup.sh is renamed to dpdk-setup.sh.

ABI Changes

  • The rte_port_source_params structure has new fields to support PCAP files. It was already in release 16.04 with RTE_NEXT_ABI flag.
  • The rte_eth_dev_info structure has new fields nb_rx_queues and nb_tx_queues to support the number of queues configured by software.
  • A Vhost ABI refactoring has been made: the virtio_net structure is no longer exported directly to the application. Instead, a handle, vid, has been used to represent this structure internally.

Shared Library Versions

The libraries prepended with a plus sign were incremented in this version.

+ libethdev.so.4
  librte_acl.so.2
  librte_cfgfile.so.2
  librte_cmdline.so.2
  librte_cryptodev.so.1
  librte_distributor.so.1
  librte_eal.so.2
  librte_hash.so.2
  librte_ip_frag.so.1
  librte_ivshmem.so.1
  librte_jobstats.so.1
  librte_kni.so.2
  librte_kvargs.so.1
  librte_lpm.so.2
  librte_mbuf.so.2
+ librte_mempool.so.2
  librte_meter.so.1
  librte_pdump.so.1
  librte_pipeline.so.3
  librte_pmd_bond.so.1
  librte_pmd_ring.so.2
+ librte_port.so.3
  librte_power.so.1
  librte_reorder.so.1
  librte_ring.so.1
  librte_sched.so.1
  librte_table.so.2
  librte_timer.so.1
+ librte_vhost.so.3

Tested Platforms

  1. SuperMicro 1U
    • BIOS: 1.0c
    • Processor: Intel(R) Atom(TM) CPU C2758 @ 2.40GHz
  2. SuperMicro 1U
    • BIOS: 1.0a
    • Processor: Intel(R) Xeon(R) CPU D-1540 @ 2.00GHz
    • Onboard NIC: Intel(R) X552/X557-AT (2x10G)
      • Firmware-version: 0x800001cf
      • Device ID (PF/VF): 8086:15ad /8086:15a8
    • kernel driver version: 4.2.5 (ixgbe)
  3. SuperMicro 2U
    • BIOS: 1.0a
    • Processor: Intel(R) Xeon(R) CPU E5-4667 v3 @ 2.00GHz
  4. Intel(R) Server board S2600GZ
    • BIOS: SE5C600.86B.02.02.0002.122320131210
    • Processor: Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
  5. Intel(R) Server board W2600CR
    • BIOS: SE5C600.86B.02.01.0002.082220131453
    • Processor: Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
  6. Intel(R) Server board S2600CWT
    • BIOS: SE5C610.86B.01.01.0009.060120151350
    • Processor: Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz
  7. Intel(R) Server board S2600WTT
    • BIOS: SE5C610.86B.01.01.0005.101720141054
    • Processor: Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz
  8. Intel(R) Server board S2600WTT
    • BIOS: SE5C610.86B.11.01.0044.090120151156
    • Processor: Intel(R) Xeon(R) CPU E5-2695 v4 @ 2.10GHz

Tested NICs

  1. Intel(R) Ethernet Controller X540-AT2
    • Firmware version: 0x80000389
    • Device id (pf): 8086:1528
    • Driver version: 3.23.2 (ixgbe)
  2. Intel(R) 82599ES 10 Gigabit Ethernet Controller
    • Firmware version: 0x61bf0001
    • Device id (pf/vf): 8086:10fb / 8086:10ed
    • Driver version: 4.0.1-k (ixgbe)
  3. Intel(R) Corporation Ethernet Connection X552/X557-AT 10GBASE-T
    • Firmware version: 0x800001cf
    • Device id (pf/vf): 8086:15ad / 8086:15a8
    • Driver version: 4.2.5 (ixgbe)
  4. Intel(R) Ethernet Converged Network Adapter X710-DA4 (4x10G)
    • Firmware version: 5.04
    • Device id (pf/vf): 8086:1572 / 8086:154c
    • Driver version: 1.4.26 (i40e)
  5. Intel(R) Ethernet Converged Network Adapter X710-DA2 (2x10G)
    • Firmware version: 5.04
    • Device id (pf/vf): 8086:1572 / 8086:154c
    • Driver version: 1.4.25 (i40e)
  6. Intel(R) Ethernet Converged Network Adapter XL710-QDA1 (1x40G)
    • Firmware version: 5.04
    • Device id (pf/vf): 8086:1584 / 8086:154c
    • Driver version: 1.4.25 (i40e)
  7. Intel(R) Ethernet Converged Network Adapter XL710-QDA2 (2X40G)
    • Firmware version: 5.04
    • Device id (pf/vf): 8086:1583 / 8086:154c
    • Driver version: 1.4.25 (i40e)
  8. Intel(R) Corporation I350 Gigabit Network Connection
    • Firmware version: 1.48, 0x800006e7
    • Device id (pf/vf): 8086:1521 / 8086:1520
    • Driver version: 5.2.13-k (igb)
  9. Intel(R) Ethernet Multi-host Controller FM10000
    • Firmware version: N/A
    • Device id (pf/vf): 8086:15d0
    • Driver version: 0.17.0.9 (fm10k)

Tested OSes

  • CentOS 7.0
  • Fedora 23
  • Fedora 24
  • FreeBSD 10.3
  • Red Hat Enterprise Linux 7.2
  • SUSE Enterprise Linux 12
  • Ubuntu 15.10
  • Ubuntu 16.04 LTS
  • Wind River Linux 8

DPDK Release 16.04

New Features

  • Added function to check primary process state.

    A new function rte_eal_primary_proc_alive() has been added to allow the user to detect if a primary process is running. Use cases for this feature include fault detection, and monitoring using secondary processes.

  • Enabled bulk allocation of mbufs.

    A new function rte_pktmbuf_alloc_bulk() has been added to allow the user to bulk allocate mbufs.

  • Added device link speed capabilities.

    The structure rte_eth_dev_info now has a speed_capa bitmap, which allows the application to determine the supported speeds of each device.

  • Added bitmap of link speeds to advertise.

    Added a feature to allow the definition of a set of advertised speeds for auto-negotiation, explicitly disabling link auto-negotiation (single speed) and full auto-negotiation.

  • Added new poll-mode driver for Amazon Elastic Network Adapters (ENA).

    The driver operates for a variety of ENA adapters through feature negotiation with the adapter and upgradable commands set. The ENA driver handles PCI Physical and Virtual ENA functions.

  • Restored vmxnet3 TX data ring.

    TX data ring has been shown to improve small packet forwarding performance on the vSphere environment.

  • Added vmxnet3 TX L4 checksum offload.

    Added support for TCP/UDP checksum offload to vmxnet3.

  • Added vmxnet3 TSO support.

    Added support for TSO to vmxnet3.

  • Added vmxnet3 support for jumbo frames.

    Added support for linking multi-segment buffers together to handle Jumbo packets.

  • Enabled Virtio 1.0 support.

    Enabled Virtio 1.0 support for Virtio pmd driver.

  • Supported Virtio for ARM.

    Enabled Virtio support for ARMv7/v8. Tested for ARM64. Virtio for ARM supports VFIO-noiommu mode only. Virtio can work with other non-x86 architectures as well, like PowerPC.

  • Supported Virtio offload in vhost-user.

    Added the offload and negotiation of checksum and TSO between vhost-user and vanilla Linux Virtio guest.

  • Added vhost-user live migration support.

  • Added vhost driver.

    Added a virtual PMD that wraps librte_vhost.

  • Added multicast promiscuous mode support on VF for ixgbe.

    Added multicast promiscuous mode support for the ixgbe VF driver so all VFs can receive the multicast packets.

    Please note if you want to use this promiscuous mode, you need both PF and VF driver to support it. The reason is that this VF feature is configured in the PF. If you use kernel PF driver and the dpdk VF driver, make sure the kernel PF driver supports VF multicast promiscuous mode. If you use dpdk PF and dpdk VF ensure the PF driver is the same version as the VF.

  • Added support for E-tag on X550.

    E-tag is defined in 802.1BR - Bridge Port Extension.

    This feature is for the VF, but the settings are on the PF. It means the CLIs should be used on the PF, but some of their effects will be shown on the VF. The forwarding of E-tag packets based on GRP and E-CID_base will have an effect on the PF. Theoretically, the E-tag packets can be forwarded to any pool/queue but normally we’d like to forward the packets to the pools/queues belonging to the VFs. And E-tag insertion and stripping will have an effect on VFs. When a VF receives E-tag packets it should strip the E-tag. When the VF transmits packets, it should insert the E-tag. Both actions can be offloaded.

    When we want to use this E-tag support feature, the forwarding should be enabled to forward the packets received by the PF to the indicated VFs. And insertion and stripping should be enabled for VFs to offload the effort to hardware.

    Features added:

    • Support E-tag offloading of insertion and stripping.
    • Support Forwarding E-tag packets to pools based on GRP and E-CID_base.
  • Added support for VxLAN and NVGRE checksum off-load on X550.

    • Added support for VxLAN and NVGRE RX/TX checksum off-load on X550. RX/TX checksum off-load is provided on both inner and outer IP header and TCP header.
    • Added functions to support VxLAN port configuration. The default VxLAN port number is 4789 but this can be updated programmatically.
  • Added support for new X550EM_a devices.

    Added support for new X550EM_a devices and their MAC types, X550EM_a and X550EM_a_vf. Updated the relevant PMD to use the new devices and MAC types.

  • Added x550em_x V2 device support.

    Added support for x550em_x V2 device. Only x550em_x V1 was supported before. A mask for V1 and V2 is defined and used to support both.

  • Supported link speed auto-negotiation on X550EM_X

    Normally the auto-negotiation is supported by firmware and software doesn’t care about it. But on x550em_x, firmware doesn’t support auto-negotiation. As the ports of x550em_x are 10GbE, if we connect the port with a peer which is 1GbE, the link will always be down. We added the support for auto-negotiation by software to avoid this link down issue.

  • Added software-firmware sync on X550EM_a.

    Added support for software-firmware sync for resource sharing. Use the PHY token, shared between software-firmware for PHY access on X550EM_a.

  • Updated the i40e base driver.

    The i40e base driver was updated with changes including the following:

    • Use RX control AQ commands to read/write RX control registers.
    • Add new X722 device IDs, and removed X710 one was never used.
    • Expose registers for HASH/FD input set configuring.
  • Enabled PCI extended tag for i40e.

    Enabled extended tag for i40e by checking and writing corresponding PCI config space bytes, to boost the performance. The legacy method of reading/writing sysfile supported by kernel module igb_uio is now deprecated.

  • Added i40e support for setting mac addresses.

  • Added dump of i40e registers and EEPROM.

  • Supported ether type setting of single and double VLAN for i40e

  • Added VMDQ DCB mode in i40e.

    Added support for DCB in VMDQ mode to i40e driver.

  • Added i40e VEB switching support.

  • Added Flow director enhancements in i40e.

  • Added PF reset event reporting in i40e VF driver.

  • Added fm10k RX interrupt support.

  • Optimized fm10k TX.

    Optimized fm10k TX by freeing multiple mbufs at a time.

  • Handled error flags in fm10k vector RX.

    Parse error flags in RX descriptor and set error bits in mbuf with vector instructions.

  • Added fm10k FTAG based forwarding support.

  • Added mlx5 flow director support.

    Added flow director support (RTE_FDIR_MODE_PERFECT and RTE_FDIR_MODE_PERFECT_MAC_VLAN).

    Only available with Mellanox OFED >= 3.2.

  • Added mlx5 RX VLAN stripping support.

    Added support for RX VLAN stripping.

    Only available with Mellanox OFED >= 3.2.

  • Added mlx5 link up/down callbacks.

    Implemented callbacks to bring link up and down.

  • Added mlx5 support for operation in secondary processes.

    Implemented TX support in secondary processes (like mlx4).

  • Added mlx5 RX CRC stripping configuration.

    Until now, CRC was always stripped. It can now be configured.

    Only available with Mellanox OFED >= 3.2.

  • Added mlx5 optional packet padding by HW.

    Added an option to make PCI bus transactions rounded to a multiple of a cache line size for better alignment.

    Only available with Mellanox OFED >= 3.2.

  • Added mlx5 TX VLAN insertion support.

    Added support for TX VLAN insertion.

    Only available with Mellanox OFED >= 3.2.

  • Changed szedata2 driver type from vdev to pdev.

    Previously szedata2 device had to be added by --vdev option. Now szedata2 PMD recognizes the device automatically during EAL initialization.

  • Added szedata2 functions for setting link up/down.

  • Added szedata2 promiscuous and allmulticast modes.

  • Added af_packet dynamic removal function.

    An af_packet device can now be detached using the API, like other PMD devices.

  • Increased number of next hops for LPM IPv4 to 2^24.

    The next_hop field has been extended from 8 bits to 24 bits for IPv4.

  • Added support of SNOW 3G (UEA2 and UIA2) for Intel Quick Assist devices.

    Enabled support for the SNOW 3G wireless algorithm for Intel Quick Assist devices. Support for cipher-only and hash-only is also provided along with algorithm-chaining operations.

  • Added SNOW3G SW PMD.

    A new Crypto PMD has been added, which provides SNOW 3G UEA2 ciphering and SNOW3G UIA2 hashing.

  • Added AES GCM PMD.

    Added new Crypto PMD to support AES-GCM authenticated encryption and authenticated decryption in software.

  • Added NULL Crypto PMD

    Added new Crypto PMD to support null crypto operations in software.

  • Improved IP Pipeline Application.

    The following features have been added to ip_pipeline application;

    • Added CPU utilization measurement and idle cycle rate computation.
    • Added link identification support through existing port-mask option or by specifying PCI device in every LINK section in the configuration file.
    • Added load balancing support in passthrough pipeline.
  • Added IPsec security gateway example.

    Added a new application implementing an IPsec Security Gateway.

Resolved Issues

Drivers
  • ethdev: Fixed overflow for 100Gbps.

    100Gbps in Mbps (100000) was exceeding the 16-bit max value of link_speed in rte_eth_link.

  • ethdev: Fixed byte order consistency between fdir flow and mask.

    Fixed issue in ethdev library where the structure for setting fdir’s mask and flow entry was not consistent in byte ordering.

  • cxgbe: Fixed crash due to incorrect size allocated for RSS table.

    Fixed a segfault that occurs when accessing part of port 0’s RSS table that gets overwritten by subsequent port 1’s part of the RSS table due to incorrect size allocated for each entry in the table.

  • cxgbe: Fixed setting wrong device MTU.

    Fixed an incorrect device MTU being set due to the Ethernet header and CRC lengths being added twice.

  • ixgbe: Fixed zeroed VF mac address.

    Resolved an issue where the VF MAC address is zeroed out in cases where the VF driver is loaded while the PF interface is down. The solution is to only set it when we get an ACK from the PF.

  • ixgbe: Fixed setting flow director flag twice.

    Resolved an issue where packets were being dropped when switching to perfect filters mode.

  • ixgbe: Set MDIO speed after MAC reset.

    The MDIO clock speed must be reconfigured after the MAC reset. The MDIO clock speed becomes invalid, therefore the driver reads invalid PHY register values. The driver now set the MDIO clock speed prior to initializing PHY ops and again after the MAC reset.

  • ixgbe: Fixed maximum number of available TX queues.

    In IXGBE, the maximum number of TX queues varies depending on the NIC operating mode. This was not being updated in the device information, providing an incorrect number in some cases.

  • i40e: Generated MAC address for each VFs.

    It generates a MAC address for each VFs during PF host initialization, and keeps the VF MAC address the same among different VF launch.

  • i40e: Fixed failure of reading/writing RX control registers.

    Fixed i40e issue of failing to read/write rx control registers when under stress with traffic, which might result in application launch failure.

  • i40e: Enabled vector driver by default.

    Previously, vector driver was disabled by default as it couldn’t fill packet type info for l3fwd to work well. Now there is an option for l3fwd to analyze the packet type so the vector driver is enabled by default.

  • i40e: Fixed link info of VF.

    Previously, the VF’s link speed stayed at 10GbE and status always was up. It did not change even when the physical link’s status changed. Now this issue is fixed to make VF’s link info consistent with physical link.

  • mlx5: Fixed possible crash during initialization.

    A crash could occur when failing to allocate private device context.

  • mlx5: Added port type check.

    Added port type check to prevent port initialization on non-Ethernet link layers and to report an error.

  • mlx5: Applied VLAN filtering to broadcast and IPv6 multicast flows.

    Prevented reception of multicast frames outside of configured VLANs.

  • mlx5: Fixed RX checksum offload in non L3/L4 packets.

    Fixed report of bad checksum for packets of unknown type.

  • aesni_mb: Fixed wrong return value when creating a device.

    The cryptodev_aesni_mb_init() function was returning the device id of the device created, instead of 0 (on success) that rte_eal_vdev_init() expects. This made it impossible to create more than one aesni_mb device from the command line.

  • qat: Fixed AES GCM decryption.

    Allowed AES GCM on the cryptodev API, but in some cases gave invalid results due to incorrect IV setting.

Libraries
  • hash: Fixed CRC32c hash computation for non multiple of 4 bytes sizes.

    Fix crc32c hash functions to return a valid crc32c value for data lengths not a multiple of 4 bytes.

  • hash: Fixed hash library to support multi-process mode.

    Fix hash library to support multi-process mode, using a jump table, instead of storing a function pointer to the key compare function. Multi-process mode only works with the built-in compare functions, however a custom compare function (not in the jump table) can only be used in single-process mode.

  • hash: Fixed return value when allocating an existing hash table.

    Changed the rte_hash*_create() functions to return NULL and set rte_errno to EEXIST when the object name already exists. This is the behavior described in the API documentation in the header file. The previous behavior was to return a pointer to the existing object in that case, preventing the caller from knowing if the object had to be freed or not.

  • lpm: Fixed return value when allocating an existing object.

    Changed the rte_lpm*_create() functions to return NULL and set rte_errno to EEXIST when the object name already exists. This is the behavior described in the API documentation in the header file. The previous behavior was to return a pointer to the existing object in that case, preventing the caller from knowing if the object had to be freed or not.

  • librte_port: Fixed segmentation fault for ring and ethdev writer nodrop.

    Fixed core dump issue on txq and swq when dropless is set to yes.

Examples
  • l3fwd-power: Fixed memory leak for non-IP packet.

    Fixed issue in l3fwd-power where, on receiving packets of types other than IPv4 or IPv6, the mbuf was not released, and caused a memory leak.

  • l3fwd: Fixed using packet type blindly.

    l3fwd makes use of packet type information without querying if devices or PMDs really set it. For those devices that don’t set ptypes, add an option to parse it.

  • examples/vhost: Fixed frequent mbuf allocation failure.

    The vhost-switch often fails to allocate mbuf when dequeue from vring because it wrongly calculates the number of mbufs needed.

API Changes

  • The ethdev statistics counter imissed is considered to be independent of ierrors. All drivers are now counting the missed packets only once, i.e. drivers will not increment ierrors anymore for missed packets.
  • The ethdev structure rte_eth_dev_info was changed to support device speed capabilities.
  • The ethdev structures rte_eth_link and rte_eth_conf were changed to support the new link API.
  • The functions rte_eth_dev_udp_tunnel_add and rte_eth_dev_udp_tunnel_delete have been renamed into rte_eth_dev_udp_tunnel_port_add and rte_eth_dev_udp_tunnel_port_delete.
  • The outer_mac and inner_mac fields in structure rte_eth_tunnel_filter_conf are changed from pointer to struct in order to keep code’s readability.
  • The fields in ethdev structure rte_eth_fdir_masks were changed to be in big endian.
  • A parameter vlan_type has been added to the function rte_eth_dev_set_vlan_ether_type.
  • The af_packet device init function is no longer public. The device should be attached via the API.
  • The LPM next_hop field is extended from 8 bits to 24 bits for IPv4 while keeping ABI compatibility.
  • A new rte_lpm_config structure is used so the LPM library will allocate exactly the amount of memory which is necessary to hold application’s rules. The previous ABI is kept for compatibility.
  • The prototype for the pipeline input port, output port and table action handlers are updated: the pipeline parameter is added, the packets mask parameter has been either removed or made input-only.

ABI Changes

  • The RETA entry size in rte_eth_rss_reta_entry64 has been increased from 8-bit to 16-bit.
  • The ethdev flow director structure rte_eth_fdir_flow structure was changed. New fields were added to extend flow director’s input set.
  • The cmdline buffer size has been increase from 256 to 512.

Shared Library Versions

The libraries prepended with a plus sign were incremented in this version.

+ libethdev.so.3
  librte_acl.so.2
  librte_cfgfile.so.2
+ librte_cmdline.so.2
  librte_distributor.so.1
  librte_eal.so.2
  librte_hash.so.2
  librte_ip_frag.so.1
  librte_ivshmem.so.1
  librte_jobstats.so.1
  librte_kni.so.2
  librte_kvargs.so.1
  librte_lpm.so.2
  librte_mbuf.so.2
  librte_mempool.so.1
  librte_meter.so.1
+ librte_pipeline.so.3
  librte_pmd_bond.so.1
  librte_pmd_ring.so.2
  librte_port.so.2
  librte_power.so.1
  librte_reorder.so.1
  librte_ring.so.1
  librte_sched.so.1
  librte_table.so.2
  librte_timer.so.1
  librte_vhost.so.2

Tested Platforms

  1. SuperMicro 1U
    • BIOS: 1.0c
    • Processor: Intel(R) Atom(TM) CPU C2758 @ 2.40GHz
  2. SuperMicro 1U
    • BIOS: 1.0a
    • Processor: Intel(R) Xeon(R) CPU D-1540 @ 2.00GHz
    • Onboard NIC: Intel(R) X552/X557-AT (2x10G)
      • Firmware-version: 0x800001cf
      • Device ID (PF/VF): 8086:15ad /8086:15a8
    • kernel driver version: 4.2.5 (ixgbe)
  3. SuperMicro 1U
    • BIOS: 1.0a
    • Processor: Intel(R) Xeon(R) CPU E5-4667 v3 @ 2.00GHz
  4. Intel(R) Server board S2600GZ
    • BIOS: SE5C600.86B.02.02.0002.122320131210
    • Processor: Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
  5. Intel(R) Server board W2600CR
    • BIOS: SE5C600.86B.02.01.0002.082220131453
    • Processor: Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
  6. Intel(R) Server board S2600CWT
    • BIOS: SE5C610.86B.01.01.0009.060120151350
    • Processor: Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz
  7. Intel(R) Server board S2600WTT
    • BIOS: SE5C610.86B.01.01.0005.101720141054
    • Processor: Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz
  8. Intel(R) Server board S2600WTT
    • BIOS: SE5C610.86B.11.01.0044.090120151156
    • Processor: Intel(R) Xeon(R) CPU E5-2695 v4 @ 2.10GHz

Tested NICs

  1. Intel(R) Ethernet Controller X540-AT2
    • Firmware version: 0x80000389
    • Device id (pf): 8086:1528
    • Driver version: 3.23.2 (ixgbe)
  2. Intel(R) 82599ES 10 Gigabit Ethernet Controller
    • Firmware version: 0x61bf0001
    • Device id (pf/vf): 8086:10fb / 8086:10ed
    • Driver version: 4.0.1-k (ixgbe)
  3. Intel(R) Corporation Ethernet Connection X552/X557-AT 10GBASE-T
    • Firmware version: 0x800001cf
    • Device id (pf/vf): 8086:15ad / 8086:15a8
    • Driver version: 4.2.5 (ixgbe)
  4. Intel(R) Ethernet Converged Network Adapter X710-DA4 (4x10G)
    • Firmware version: 5.02 0x80002284
    • Device id (pf/vf): 8086:1572 / 8086:154c
    • Driver version: 1.4.26 (i40e)
  5. Intel(R) Ethernet Converged Network Adapter X710-DA2 (2x10G)
    • Firmware version: 5.02 0x80002282
    • Device id (pf/vf): 8086:1572 / 8086:154c
    • Driver version: 1.4.25 (i40e)
  6. Intel(R) Ethernet Converged Network Adapter XL710-QDA1 (1x40G)
    • Firmware version: 5.02 0x80002281
    • Device id (pf/vf): 8086:1584 / 8086:154c
    • Driver version: 1.4.25 (i40e)
  7. Intel(R) Ethernet Converged Network Adapter XL710-QDA2 (2X40G)
    • Firmware version: 5.02 0x80002285
    • Device id (pf/vf): 8086:1583 / 8086:154c
    • Driver version: 1.4.25 (i40e)
  8. Intel(R) 82576EB Gigabit Ethernet Controller
    • Firmware version: 1.2.1
    • Device id (pf): 8086:1526
    • Driver version: 5.2.13-k (igb)
  9. Intel(R) Ethernet Controller I210
    • Firmware version: 3.16, 0x80000500, 1.304.0
    • Device id (pf): 8086:1533
    • Driver version: 5.2.13-k (igb)
  10. Intel(R) Corporation I350 Gigabit Network Connection
    • Firmware version: 1.48, 0x800006e7
    • Device id (pf/vf): 8086:1521 / 8086:1520
    • Driver version: 5.2.13-k (igb)
  11. Intel(R) Ethernet Multi-host Controller FM10000
    • Firmware version: N/A
    • Device id (pf/vf): 8086:15d0
    • Driver version: 0.17.0.9 (fm10k)

DPDK Release 2.2

New Features

  • Introduce ARMv7 and ARMv8 architectures.

    • It is now possible to build DPDK for the ARMv7 and ARMv8 platforms.
    • ARMv7 can be tested with virtual PMD drivers.
    • ARMv8 can be tested with virtual and physical PMD drivers.
  • Enabled freeing of ring.

    A new function rte_ring_free() has been added to allow the user to free a ring if it was created with rte_ring_create().

  • Added keepalive support to EAL and example application.

  • Added experimental cryptodev API

    The cryptographic processing of packets is provided as a preview with two drivers for:

    • Intel QuickAssist devices
    • Intel AES-NI multi-buffer library

    Due to its experimental state, the API may change without prior notice.

  • Added ethdev APIs for additional IEEE1588 support.

    Added functions to read, write and adjust system time in the NIC. Added client slave sample application to demonstrate the IEEE1588 functionality.

  • Extended Statistics.

    Defined an extended statistics naming scheme to store metadata in the name string of each statistic. Refer to the Extended Statistics section of the Programmers Guide for more details.

    Implemented the extended statistics API for the following PMDs:

    • igb
    • igbvf
    • i40e
    • i40evf
    • fm10k
    • virtio
  • Added API in ethdev to retrieve RX/TX queue information.

    • Added the ability for the upper layer to query RX/TX queue information.
    • Added new fields in rte_eth_dev_info to represent information about RX/TX descriptors min/max/align numbers, per queue, for the device.
  • Added RSS dynamic configuration to bonding.

  • Updated the e1000 base driver.

    The e1000 base driver was updated with several features including the following:

    • Added new i218 devices
    • Allowed both ULP and EEE in Sx state
    • Initialized 88E1543 (Marvell 1543) PHY
    • Added flags to set EEE advertisement modes
    • Supported inverted format ETrackId
    • Added bit to disable packetbuffer read
    • Added defaults for i210 RX/TX PBSIZE
    • Check more errors for ESB2 init and reset
    • Check more NVM read errors
    • Return code after setting receive address register
    • Removed all NAHUM6LP_HW tags
  • Added e1000 RX interrupt support.

  • Added igb TSO support for both PF and VF.

  • Added RSS enhancements to Intel x550 NIC.

    • Added support for 512 entry RSS redirection table.
    • Added support for per VF RSS redirection table.
  • Added Flow director enhancements on Intel x550 NIC.

    • Added 2 new flow director modes on x550. One is MAC VLAN mode, the other is tunnel mode.
  • Updated the i40e base driver.

    The i40e base driver was updated with several changes including the following:

    • Added promiscuous on VLAN support
    • Added a workaround to drop all flow control frames
    • Added VF capabilities to virtual channel interface
    • Added TX Scheduling related AQ commands
    • Added additional PCTYPES supported for FortPark RSS
    • Added parsing for CEE DCBX TLVs
    • Added FortPark specific registers
    • Added AQ functions to handle RSS Key and LUT programming
    • Increased PF reset max loop limit
  • Added i40e vector RX/TX.

  • Added i40e RX interrupt support.

  • Added i40e flow control support.

  • Added DCB support to i40e PF driver.

  • Added RSS/FD input set granularity on Intel X710/XL710.

  • Added different GRE key length for input set on Intel X710/XL710.

  • Added flow director support in i40e VF.

  • Added i40e support of early X722 series.

    Added early X722 support, for evaluation only, as the hardware is alpha.

  • Added fm10k vector RX/TX.

  • Added fm10k TSO support for both PF and VF.

  • Added fm10k VMDQ support.

  • New NIC Boulder Rapid support.

    Added support for the Boulder Rapid variant of Intel’s fm10k NIC family.

  • Enhanced support for the Chelsio CXGBE driver.

    • Added support for Jumbo Frames.
    • Optimized forwarding performance for Chelsio T5 40GbE cards.
  • Improved enic TX packet rate.

    Reduced frequency of TX tail pointer updates to the NIC.

  • Added support for link status interrupts in mlx4.

  • Added partial support (TX only) for secondary processes in mlx4.

  • Added support for Mellanox ConnectX-4 adapters (mlx5).

    The mlx5 poll-mode driver implements support for Mellanox ConnectX-4 EN and Mellanox ConnectX-4 Lx EN families of 10/25/40/50/100 Gb/s adapters.

    Like mlx4, this PMD is only available for Linux and is disabled by default due to external dependencies (libibverbs and libmlx5).

  • Added driver for Netronome nfp-6xxx card.

    Support for using Netronome nfp-6xxx with PCI VFs.

  • Added virtual szedata2 driver for COMBO cards.

    Added virtual PMD for COMBO-100G and COMBO-80G cards. PMD is disabled in default configuration.

  • Enhanced support for virtio driver.

    • Virtio ring layout optimization (fixed avail ring)
    • Vector RX
    • Simple TX
  • Added vhost-user multiple queue support.

  • Added port hotplug support to vmxnet3.

  • Added port hotplug support to xenvirt.

  • Added ethtool shim and sample application.

  • Added experimental performance thread example application.

    The new sample application demonstrates L3 forwarding with different threading models: pthreads, cgroups, or lightweight threads. The example includes a simple cooperative scheduler.

    Due to its experimental state this application may change without notice. The application is supported only for Linux x86_64.

  • Enhancements to the IP pipeline application.

    The following features have been added to the ip_pipeline application;

    • Added Multiple Producers/Multiple Consumers (MPSC) and fragmentation/reassembly support to software rings.
    • Added a dynamic pipeline reconfiguration feature that allows binding a pipeline to other threads at runtime using CLI commands.
    • Added enable/disable of promisc mode from ip_pipeline configuration file.
    • Added check on RX queues and TX queues of each link whether they are used correctly in the ip_pipeline configuration file.
    • Added flow id parameters to the flow-classification table entries.
    • Added more functions to the routing pipeline: ARP table enable/disable, Q-in-Q and MPLS encapsulation, add color (traffic-class for QoS) to the MPLS tag.
    • Added flow-actions pipeline for traffic metering/marking (for e.g. Two Rate Three Color Marker (trTCM)), policer etc.
    • Modified the pass-through pipeline’s actions-handler to implement a generic approach to extract fields from the packet’s header and copy them to packet metadata.

Resolved Issues

EAL
  • eal/linux: Fixed epoll timeout.

    Fixed issue where the rte_epoll_wait() function didn’t return when the underlying call to epoll_wait() timed out.

Drivers
  • e1000/base: Synchronize PHY interface on non-ME systems.

    On power up, the MAC - PHY interface needs to be set to PCIe, even if the cable is disconnected. In ME systems, the ME handles this on exit from the Sx (Sticky mode) state. In non-ME, the driver handles it. Added a check for non-ME system to the driver code that handles it.

  • e1000/base: Increased timeout of reset check.

    Previously, in check_reset_block RSPCIPHY was polled for 100 ms before determining that the ME veto was set. This was not enough and it was increased to 300 ms.

  • e1000/base: Disabled IPv6 extension header parsing on 82575.

    Disabled IPv6 options as per hardware limitation.

  • e1000/base: Prevent ULP flow if cable connected.

    Enabling ULP on link down when the cable is connected caused an infinite loop of link up/down indications in the NDIS driver. The driver now enables ULP only when the cable is disconnected.

  • e1000/base: Support different EEARBC for i210.

    EEARBC has changed on i210. It means EEARBC has a different address on i210 than on other NICs. So, add a new entity named EEARBC_I210 to the register list and make sure the right one is being used on i210.

  • e1000/base: Fix K1 configuration.

    Added fix for the following updates to the K1 configurations: TX idle period for entering K1 should be 128 ns. Minimum TX idle period in K1 should be 256 ns.

  • e1000/base: Fix link detect flow.

    Fix link detect flow in case where auto-negotiate is not enabled, by calling e1000_setup_copper_link_generic instead of e1000_phy_setup_autoneg.

  • e1000/base: Fix link check for i354 M88E1112 PHY.

    The e1000_check_for_link_media_swap() function is supposed to check PHY page 0 for copper and PHY page 1 for “other” (fiber) links. The driver switched back from page 1 to page 0 too soon, before e1000_check_for_link_82575() is executed and was never finding the link on the fiber (other).

    If the link is copper, as the M88E1112 page address is set to 1, it should be set back to 0 before checking this link.

  • e1000/base: Fix beacon duration for i217.

    Fix for I217 Packet Loss issue - The Management Engine sets the FEXTNVM4 Beacon Duration incorrectly. This fix ensures that the correct value will always be set. Correct value for this field is 8 usec.

  • e1000/base: Fix TIPG for non 10 half duplex mode.

    TIPG value is increased when setting speed to 10 half duplex to prevent packet loss. However, it was never decreased again when speed changed. This caused performance issues in the NDIS driver. Fix this to restore TIPG to default value on non 10 half duplex.

  • e1000/base: Fix reset of DH89XXCC SGMII.

    For DH89XXCC_SGMII, a write flush leaves registers of this device trashed (0xFFFFFFFF). Add check for this device.

    Also, after both Port SW Reset and Device Reset case, the platform should wait at least 3ms before reading any registers. Remove this condition since waiting is conditionally executed only for Device Reset.

  • e1000/base: Fix redundant PHY power down for i210.

    Bit 11 of PHYREG 0 is used to power down PHY. The use of PHYREG 16 is no longer necessary.

  • e1000/base: fix jumbo frame CRC failures.

    Change the value of register 776.20[11:2] for jumbo mode from 0x1A to 0x1F. This is to enlarge the gap between read and write pointers in the TX FIFO.

  • e1000/base: Fix link flap on 82579.

    Several customers have reported a link flap issue on 82579. The symptoms are random and intermittent link losses when 82579 is connected to specific switches. the Issue was root caused as an inter-operability problem between the NIC and at least some Broadcom PHYs in the Energy Efficient Ethernet wake mechanism.

    To fix the issue, we are disabling the Phase Locked Loop shutdown in 100M Low Power Idle. This solution will cause an increase of power in 100M EEE link. It may cost an additional 28mW in this specific mode.

  • igb: Fixed IEEE1588 frame identification in I210.

    Fixed issue where the flag PKT_RX_IEEE1588_PTP was not being set in the Intel I210 NIC, as the EtherType in RX descriptor is in bits 8:10 of Packet Type and not in the default bits 0:2.

  • igb: Fixed VF start with PF stopped.

    VF needs the PF interrupt support initialized even if not started.

  • igb: Fixed VF MAC address when using with DPDK PF.

    Assign a random MAC address in VF when not assigned by PF.

  • igb: Removed CRC bytes from byte counter statistics.

  • ixgbe: Fixed issue with X550 DCB.

    Fixed a DCB issue with x550 where for 8 TCs (Traffic Classes), if a packet with user priority 6 or 7 was injected to the NIC, then the NIC would only put 3 packets into the queue. There was also a similar issue for 4 TCs.

  • ixgbe: Removed burst size restriction of vector RX.

    Fixed issue where a burst size less than 32 didn’t receive anything.

  • ixgbe: Fixed VF start with PF stopped.

    VF needs the PF interrupt support initialized even if not started.

  • ixgbe: Fixed TX hang when RS distance exceeds HW limit.

    Fixed an issue where the TX queue can hang when a lot of highly fragmented packets have to be sent. As part of that fix, tx_rs_thresh for ixgbe PMD is not allowed to be greater then to 32 to comply with HW restrictions.

  • ixgbe: Fixed rx error statistic counter.

    Fixed an issue that the rx error counter of ixgbe was not accurate. The mac short packet discard count (mspdc) was added to the counter. Mac local faults and mac remote faults are removed as they do not count packets but errors, and jabber errors were removed as they are already accounted for by the CRC error counter. Finally the XEC (l3 / l4 checksum error) counter was removed due to errata, see commit 256ff05a9cae for details.

  • ixgbe: Removed CRC bytes from byte counter statistics.

  • i40e: Fixed base driver allocation when not using first numa node.

    Fixed i40e issue that occurred when a DPDK application didn’t initialize ports if memory wasn’t available on socket 0.

  • i40e: Fixed maximum of 64 queues per port.

    Fixed an issue in i40e where it would not support more than 64 queues per port, even though the hardware actually supports it. The real number of queues may vary, as long as the total number of queues used in PF, VFs, VMDq and FD does not exceeds the hardware maximum.

  • i40e: Fixed statistics of packets.

    Added discarding packets on VSI to the stats and rectify the old statistics.

  • i40e: Fixed issue of not freeing memzone.

    Fixed an issue of not freeing a memzone in the call to free the memory for adminq DMA.

  • i40e: Removed CRC bytes from byte counter statistics.

  • mlx: Fixed driver loading.

    The mlx drivers were unable to load when built as a shared library, due to a missing symbol in the mempool library.

  • mlx4: Performance improvements.

    Fixed bugs in TX and RX flows that improves mlx4 performance.

  • mlx4: Fixed TX loss after initialization.

  • mlx4: Fixed scattered TX with too many segments.

  • mlx4: Fixed memory registration for indirect mbuf data.

  • vhost: Fixed Qemu shutdown.

    Fixed issue with libvirt virsh destroy not killing the VM.

  • virtio: Fixed crash after changing link state.

    Fixed IO permission in the interrupt handler.

  • virtio: Fixed crash when releasing queue.

    Fixed issue when releasing null control queue.

Libraries
  • hash: Fixed memory allocation of Cuckoo Hash key table.

    Fixed issue where an incorrect Cuckoo Hash key table size could be calculated limiting the size to 4GB.

  • hash: Fixed incorrect lookup if key is all zero.

    Fixed issue in hash library that occurred if an all zero key was not added to the table and the key was looked up, resulting in an incorrect hit.

  • hash: Fixed thread scaling by reducing contention.

    Fixed issue in the hash library where, using multiple cores with hardware transactional memory support, thread scaling did not work, due to the global ring that is shared by all cores.

Examples
  • l3fwd: Fixed crash with IPv6.
  • vhost_xen: Fixed compile error.
Other
  • This release drops compatibility with Linux kernel 2.6.33. The minimum kernel requirement is now 2.6.34.

Known Issues

  • Some drivers do not fill in the packet type when receiving. As the l3fwd example application requires this info, the i40e vector driver must be disabled to benefit of the packet type with i40e.
  • Some (possibly all) VF drivers (e.g. i40evf) do not handle any PF reset events/requests in the VF driver. This means that the VF driver may not work after a PF reset in the host side. The workaround is to avoid triggering any PF reset events/requests on the host side.
  • 100G link report support is missing.
  • Mellanox PMDs (mlx4 & mlx5):
    • PMDs do not support CONFIG_RTE_BUILD_COMBINE_LIBS and CONFIG_RTE_BUILD_SHARED_LIB simultaneously.
    • There is performance degradation for small packets when the PMD is compiled with SGE_WR_N = 4 compared to the performance when SGE_WR_N = 1. If scattered packets are not used it is recommended to compile the PMD with SGE_WR_N = 1.
    • When a Multicast or Broadcast packet is sent to the SR-IOV mlx4 VF, it is returned back to the port.
    • PMDs report “bad” L4 checksum when IP packet is received.
    • mlx5 PMD reports “bad” checksum although the packet has “good” checksum. Will be fixed in upcoming MLNX_OFED release.

API Changes

  • The deprecated flow director API is removed. It was replaced by rte_eth_dev_filter_ctrl().
  • The dcb_queue is renamed to dcb_tc in following dcb configuration structures: rte_eth_dcb_rx_conf, rte_eth_dcb_tx_conf, rte_eth_vmdq_dcb_conf, rte_eth_vmdq_dcb_tx_conf.
  • The rte_eth_rx_queue_count() function now returns “int” instead of “uint32_t” to allow the use of negative values as error codes on return.
  • The function rte_eal_pci_close_one() is removed. It was replaced by rte_eal_pci_detach().
  • The deprecated ACL API ipv4vlan is removed.
  • The deprecated hash function rte_jhash2() is removed. It was replaced by rte_jhash_32b().
  • The deprecated KNI functions are removed: rte_kni_create(), rte_kni_get_port_id() and rte_kni_info_get().
  • The deprecated ring PMD functions are removed: rte_eth_ring_pair_create() and rte_eth_ring_pair_attach().
  • The devargs union field virtual is renamed to virt for C++ compatibility.

ABI Changes

  • The EAL and ethdev structures rte_intr_handle and rte_eth_conf were changed to support RX interrupt. This was already included in 2.1 under the CONFIG_RTE_NEXT_ABI #define.
  • The ethdev flow director entries for SCTP were changed. This was already included in 2.1 under the CONFIG_RTE_NEXT_ABI #define.
  • The ethdev flow director structure rte_eth_fdir_flow_ext structure was changed. New fields were added to support flow director filtering in VF.
  • The size of the ethdev structure rte_eth_hash_filter_info is changed by adding a new element rte_eth_input_set_conf in a union.
  • New fields rx_desc_lim and tx_desc_lim are added into rte_eth_dev_info structure.
  • For debug builds, the functions rte_eth_rx_burst(), rte_eth_tx_burst() rte_eth_rx_descriptor_done() and rte_eth_rx_queue_count() will no longer be separate functions in the DPDK libraries. Instead, they will only be present in the rte_ethdev.h header file.
  • The maximum number of queues per port CONFIG_RTE_MAX_QUEUES_PER_PORT is increased to 1024.
  • The mbuf structure was changed to support the unified packet type. This was already included in 2.1 under the CONFIG_RTE_NEXT_ABI #define.
  • The dummy malloc library is removed. The content was moved into EAL in 2.1.
  • The LPM structure is changed. The deprecated field mem_location is removed.
  • librte_table LPM: A new parameter to hold the table name will be added to the LPM table parameter structure.
  • librte_table hash: The key mask parameter is added to the hash table parameter structure for 8-byte key and 16-byte key extendable bucket and LRU tables.
  • librte_port: Macros to access the packet meta-data stored within the packet buffer has been adjusted to cover the packet mbuf structure.
  • librte_cfgfile: Allow longer names and values by increasing the constants CFG_NAME_LEN and CFG_VALUE_LEN to 64 and 256 respectively.
  • vhost: a new field enabled is added to the vhost_virtqueue structure.
  • vhost: a new field virt_qp_nb is added to virtio_net structure, and the virtqueue field is moved to the end of virtio_net structure.
  • vhost: a new operation vring_state_changed is added to virtio_net_device_ops structure.
  • vhost: a few spaces are reserved both at vhost_virtqueue and virtio_net structure for future extension.

Shared Library Versions

The libraries prepended with a plus sign were incremented in this version.

+ libethdev.so.2
+ librte_acl.so.2
+ librte_cfgfile.so.2
  librte_cmdline.so.1
  librte_distributor.so.1
+ librte_eal.so.2
+ librte_hash.so.2
  librte_ip_frag.so.1
  librte_ivshmem.so.1
  librte_jobstats.so.1
+ librte_kni.so.2
  librte_kvargs.so.1
+ librte_lpm.so.2
+ librte_mbuf.so.2
  librte_mempool.so.1
  librte_meter.so.1
+ librte_pipeline.so.2
  librte_pmd_bond.so.1
+ librte_pmd_ring.so.2
+ librte_port.so.2
  librte_power.so.1
  librte_reorder.so.1
  librte_ring.so.1
  librte_sched.so.1
+ librte_table.so.2
  librte_timer.so.1
+ librte_vhost.so.2

DPDK Release 2.1

New Features

  • Enabled cloning of indirect mbufs.

    This feature removes a limitation of rte_pktmbuf_attach() which generated the warning: “mbuf we’re attaching to must be direct”.

    Now, when attaching to an indirect mbuf it is possible to:

    • Copy all relevant fields (address, length, offload, ...) as before.
    • Get the pointer to the mbuf that embeds the data buffer (direct mbuf), and increase the reference counter.

    When detaching the mbuf, we can now retrieve this direct mbuf as the pointer is determined from the buffer address.

  • Extended packet type support.

    In previous releases mbuf packet types were indicated by 6 bits in the ol_flags. This was not enough for some supported NICs. For example i40e hardware can recognize more than 150 packet types. Not being able to identify these additional packet types limits access to hardware offload capabilities

    So an extended “unified” packet type was added to support all possible PMDs. The 16 bit packet_type in the mbuf structure was changed to 32 bits and used for this purpose.

    To avoid breaking ABI compatibility, the code changes for this feature are enclosed in a RTE_NEXT_ABI ifdef. This is enabled by default but can be turned off for ABI compatibility with DPDK R2.0.

  • Reworked memzone to be allocated by malloc and also support freeing.

    In the memory hierarchy, memsegs are groups of physically contiguous hugepages, memzones are slices of memsegs, and malloc slices memzones into smaller memory chunks.

    This feature modifies malloc() so it partitions memsegs instead of memzones. Now memzones allocate their memory from the malloc heap.

    Backward compatibility with API and ABI are maintained.

    This allow memzones, and any other structure based on memzones, for example mempools, to be freed. Currently only the API from freeing memzones is supported.

  • Interrupt mode PMD.

    This feature introduces a low-latency one-shot RX interrupt into DPDK. It also adds a polling and interrupt mode switch control example.

    DPDK userspace interrupt notification and handling mechanism is based on UIO/VFIO with the following limitations:

    • Per queue RX interrupt events are only allowed in VFIO which supports multiple MSI-X vectors.
    • In UIO, the RX interrupt shares the same vector with other interrupts. When the RX interrupt and LSC interrupt are both enabled, only the former is available.
    • RX interrupt is only implemented for the linuxapp target.
    • The feature is only currently enabled for tow PMDs: ixgbe and igb.
  • Packet Framework enhancements.

    Several enhancements were made to the Packet Framework:

    • A new configuration file syntax has been introduced for IP pipeline applications. Parsing of the configuration file is changed.
    • Implementation of the IP pipeline application is modified to make it more structured and user friendly.
    • Implementation of the command line interface (CLI) for each pipeline type has been moved to the separate compilation unit. Syntax of pipeline CLI commands has been changed.
    • Initialization of IP pipeline is modified to match the new parameters structure.
    • New implementation of pass-through pipeline, firewall pipeline, routing pipeline, and flow classification has been added.
    • Master pipeline with CLI interface has been added.
    • Added extended documentation of the IP Pipeline.
  • Added API for IEEE1588 timestamping.

    This feature adds an ethdev API to enable, disable and read IEEE1588/802.1AS PTP timestamps from devices that support it. The following functions were added:

    • rte_eth_timesync_enable()
    • rte_eth_timesync_disable()
    • rte_eth_timesync_read_rx_timestamp()
    • rte_eth_timesync_read_tx_timestamp()

    The “ieee1588” forwarding mode in testpmd was also refactored to demonstrate the new API.

  • Added multicast address filtering.

    Added multicast address filtering via a new ethdev function set_mc_addr_list().

    This overcomes a limitation in previous releases where the receipt of multicast packets on a given port could only be enabled by invoking the rte_eth_allmulticast_enable() function. This method did not work for VFs in SR-IOV architectures when the host PF driver does not allow these operation on VFs. In such cases, joined multicast addresses had to be added individually to the set of multicast addresses that are filtered by the [VF] port.

  • Added Flow Director extensions.

    Several Flow Director extensions were added such as:

    • Support for RSS and Flow Director hashes in vector RX.
    • Added Flow Director for L2 payload.
  • Added RSS hash key size query per port.

    This feature supports querying the RSS hash key size of each port. A new field hash_key_size has been added in the rte_eth_dev_info struct for storing hash key size in bytes.

  • Added userspace ethtool support.

    Added userspace ethtool support to provide a familiar interface for applications that manage devices via kernel-space ethtool_op and net_device_op.

    The initial implementation focuses on operations that can be implemented through existing netdev APIs. More operations will be supported in later releases.

  • Updated the ixgbe base driver.

    The ixgbe base driver was updated with several changes including the following:

    • Added a new 82599 device id.
    • Added new X550 PHY ids.
    • Added SFP+ dual-speed support.
    • Added wait helper for X550 IOSF accesses.
    • Added X550em features.
    • Added X557 PHY LEDs support.
    • Commands for flow director.
    • Issue firmware command when resetting X550em.

    See the git log for full details of the ixgbe/base changes.

  • Added additional hotplug support.

    Port hotplug support was added to the following PMDs:

    • e1000/igb.
    • ixgbe.
    • i40e.
    • fm10k.
    • ring.
    • bonding.
    • virtio.

    Port hotplug support was added to BSD.

  • Added ixgbe LRO support.

    Added LRO support for x540 and 82599 devices.

  • Added extended statistics for ixgbe.

    Implemented xstats_get() and xstats_reset() in dev_ops for ixgbe to expose detailed error statistics to DPDK applications.

    These will be implemented for other PMDs in later releases.

  • Added proc_info application.

    Created a new proc_info application, by refactoring the existing dump_cfg application, to demonstrate the usage of retrieving statistics, and the new extended statistics (see above), for DPDK interfaces.

  • Updated the i40e base driver.

    The i40e base driver was updated with several changes including the following:

    • Support for building both PF and VF driver together.
    • Support for CEE DCBX on recent firmware versions.
    • Replacement of i40e_debug_read_register().
    • Rework of i40e_hmc_get_object_va.
    • Update of shadow RAM read/write functions.
    • Enhancement of polling NVM semaphore.
    • Enhancements on adminq init and sending asq command.
    • Update of get/set LED functions.
    • Addition of AOC phy types to case statement in get_media_type.
    • Support for iSCSI capability.
    • Setting of FLAG_RD when sending driver version to FW.

    See the git log for full details of the i40e/base changes.

  • Added support for port mirroring in i40e.

    Enabled mirror functionality in the i40e driver.

  • Added support for i40e double VLAN, QinQ, stripping and insertion.

    Added support to the i40e driver for offloading double VLAN (QinQ) tags to the mbuf header, and inserting double vlan tags by hardware to the packets to be transmitted. Added a new field vlan_tci_outer in the rte_mbuf struct, and new flags in ol_flags to support this feature.

  • Added fm10k promiscuous mode support.

    Added support for promiscuous/allmulticast enable and disable in the fm10k PF function. VF is not supported yet.

  • Added fm10k jumbo frame support.

    Added support for jumbo frame less than 15K in both VF and PF functions in the fm10k pmd.

  • Added fm10k mac vlan filtering support.

    Added support for the fm10k MAC filter, only available in PF. Updated the VLAN filter to add/delete one static entry in the MAC table for each combination of VLAN and MAC address.

  • Added support for the Broadcom bnx2x driver.

    Added support for the Broadcom NetXtreme II bnx2x driver. It is supported only on Linux 64-bit and disabled by default.

  • Added support for the Chelsio CXGBE driver.

    Added support for the CXGBE Poll Mode Driver for the Chelsio Terminator 5 series of 10G/40G adapters.

  • Enhanced support for Mellanox ConnectX-3 driver (mlx4).

    • Support Mellanox OFED 3.0.
    • Improved performance for both RX and TX operations.
    • Better link status information.
    • Outer L3/L4 checksum offload support.
    • Inner L3/L4 checksum offload support for VXLAN.
  • Enabled VMXNET3 vlan filtering.

    Added support for the VLAN filter functionality of the VMXNET3 interface.

  • Added support for vhost live migration.

    Added support to allow live migration of vhost. Without this feature, qemu will report the following error: “migrate: Migration disabled: vhost lacks VHOST_F_LOG_ALL feature”.

  • Added support for pcap jumbo frames.

    Extended the PCAP PMD to support jumbo frames for RX and TX.

  • Added support for the TILE-Gx architecture.

    Added support for the EZchip TILE-Gx family of SoCs.

  • Added hardware memory transactions/lock elision for x86.

    Added the use of hardware memory transactions (HTM) on fast-path for rwlock and spinlock (a.k.a. lock elision). The methods are implemented for x86 using Restricted Transactional Memory instructions (Intel(r) Transactional Synchronization Extensions). The implementation fall-backs to the normal rwlock if HTM is not available or memory transactions fail. This is not a replacement for all rwlock usages since not all critical sections protected by locks are friendly to HTM. For example, an attempt to perform a HW I/O operation inside a hardware memory transaction always aborts the transaction since the CPU is not able to roll-back should the transaction fail. Therefore, hardware transactional locks are not advised to be used around rte_eth_rx_burst() and rte_eth_tx_burst() calls.

  • Updated Jenkins Hash function

    Updated the version of the Jenkins Hash (jhash) function used in DPDK from the 1996 version to the 2006 version. This gives up to 35% better performance, compared to the original one.

    Note, the hashes generated by the updated version differ from the hashes generated by the previous version.

  • Added software implementation of the Toeplitz RSS hash

    Added a software implementation of the Toeplitz hash function used by RSS. It can be used either for packet distribution on a single queue NIC or for simulating RSS computation on a specific NIC (for example after GRE header de-encapsulation).

  • Replaced the existing hash library with a Cuckoo hash implementation.

    Replaced the existing hash library with another approach, using the Cuckoo Hash method to resolve collisions (open addressing). This method pushes items from a full bucket when a new entry must be added to it, storing the evicted entry in an alternative location, using a secondary hash function.

    This gives the user the ability to store more entries when a bucket is full, in comparison with the previous implementation.

    The API has not been changed, although new fields have been added in the rte_hash structure, which has been changed to internal use only.

    The main change when creating a new table is that the number of entries per bucket is now fixed, so its parameter is ignored now (it is still there to maintain the same parameters structure).

    Also, the maximum burst size in lookup_burst function hash been increased to 64, to improve performance.

  • Optimized KNI RX burst size computation.

    Optimized KNI RX burst size computation by avoiding checking how many entries are in kni->rx_q prior to actually pulling them from the fifo.

  • Added KNI multicast.

    Enabled adding multicast addresses to KNI interfaces by adding an empty callback for set_rx_mode (typically used for setting up hardware) so that the ioctl succeeds. This is the same thing as the Linux tap interface does.

  • Added cmdline polling mode.

    Added the ability to process console input in the same thread as packet processing by using the poll() function.

  • Added VXLAN Tunnel End point sample application.

    Added a Tunnel End point (TEP) sample application that simulates a VXLAN Tunnel Endpoint (VTEP) termination in DPDK. It is used to demonstrate the offload and filtering capabilities of Intel XL710 10/40 GbE NICsfor VXLAN packets.

  • Enabled combining of the ``-m`` and ``–no-huge`` EAL options.

    Added option to allow combining of the -m and --no-huge EAL command line options.

    This allows user application to run as non-root but with higher memory allocations, and removes a constraint on --no-huge mode being limited to 64M.

Resolved Issues

  • acl: Fix ambiguity between test rules.

    Some test rules had equal priority for the same category. That could cause an ambiguity in building the trie and test results.

  • acl: Fix invalid rule wildness calculation for bitmask field type.

  • acl: Fix matching rule.

  • acl: Fix unneeded trie splitting for subset of rules.

    When rebuilding a trie for limited rule-set, don’t try to split the rule-set even further.

  • app/testpmd: Fix crash when port id out of bound.

    Fixed issues in testpmd where using a port greater than 32 would cause a seg fault.

    Fixes: edab33b1c01d (“app/testpmd: support port hotplug”)

  • app/testpmd: Fix reply to a multicast ICMP request.

    Set the IP source and destination addresses in the IP header of the ICMP reply.

  • app/testpmd: fix MAC address in ARP reply.

    Fixed issue where in the icmpecho forwarding mode, ARP replies from testpmd contain invalid zero-filled MAC addresses.

    Fixes: 31db4d38de72 (“net: change arp header struct declaration”)

  • app/testpmd: fix default flow control values.

    Fixes: 422a20a4e62d (“app/testpmd: fix uninitialized flow control variables”)

  • bonding: Fix crash when stopping inactive slave.

  • bonding: Fix device initialization error handling.

  • bonding: Fix initial link status of slave.

    On Fortville NIC, link status change interrupt callback was not executed when slave in bonding was (re-)started.

  • bonding: Fix socket id for LACP slave.

    Fixes: 46fb43683679 (“bond: add mode 4”)

  • bonding: Fix device initialization error handling.

  • cmdline: Fix small memory leak.

    A function in cmdline.c had a return that did not free the buf properly.

  • config: Enable same drivers options for Linux and BSD.

    Enabled vector ixgbe and i40e bulk alloc for BSD as it is already done for Linux.

    Fixes: 304caba12643 (“config: fix bsd options”) Fixes: 0ff3324da2eb (“ixgbe: rework vector pmd following mbuf changes”)

  • devargs: Fix crash on failure.

    This problem occurred when passing an invalid PCI id to the blacklist API in devargs.

  • e1000/i40e: Fix descriptor done flag with odd address.

  • e1000/igb: fix ieee1588 timestamping initialization.

    Fixed issue with e1000 ieee1588 timestamp initialization. On initialization the IEEE1588 functions read the system time to set their timestamp. However, on some 1G NICs, for example, i350, system time is disabled by default and the IEEE1588 timestamp was always 0.

  • eal/bsd: Fix inappropriate header guards.

  • eal/bsd: Fix virtio on FreeBSD.

    Closing the /dev/io fd caused a SIGBUS in inb/outb instructions as the process lost the IOPL privileges once the fd is closed.

    Fixes: 8a312224bcde (“eal/bsd: fix fd leak”)

  • eal/linux: Fix comments on vfio MSI.

  • eal/linux: Fix irq handling with igb_uio.

    Fixed an issue where the the introduction of uio_pci_generic broke interrupt handling with igb_uio.

    Fixes: c112df6875a5 (“eal/linux: toggle interrupt for uio_pci_generic”)

  • eal/linux: Fix numa node detection.

  • eal/linux: Fix socket value for undetermined numa node.

    Sets zero as the default value of pci device numa_node if the socket could not be determined. This provides the same default value as FreeBSD which has no NUMA support, and makes the return value of rte_eth_dev_socket_id() be consistent with the API description.

  • eal/ppc: Fix cpu cycle count for little endian.

    On IBM POWER8 PPC64 little endian architecture, the definition of tsc union will be different. This fix enables the right output from rte_rdtsc().

  • ethdev: Fix check of threshold for TX freeing.

    Fixed issue where the parameter to tx_free_thresh was not consistent between the drivers.

  • ethdev: Fix crash if malloc of user callback fails.

    If rte_zmalloc() failed in rte_eth_dev_callback_register then the NULL pointer would be dereferenced.

  • ethdev: Fix illegal port access.

    To obtain a detachable flag, pci_drv is accessed in rte_eth_dev_is_detachable(). However pci_drv is only valid if port is enabled. Fixed by checking rte_eth_dev_is_valid_port() first.

  • ethdev: Make tables const.

  • ethdev: Rename and extend the mirror type.

  • examples/distributor: Fix debug macro.

    The macro to turn on additional debug output when the app was compiled with -DDEBUG was broken.

    Fixes: 07db4a975094 (“examples/distributor: new sample app”)

  • examples/kni: Fix crash on exit.

  • examples/vhost: Fix build with debug enabled.

    Fixes: 72ec8d77ac68 (“examples/vhost: rework duplicated code”)

  • fm10k: Fix RETA table initialization.

    The fm10k driver has 128 RETA entries in 32 registers, but it only initialized the first 32 when doing multiple RX queue configurations. This fix initializes all 128 entries.

  • fm10k: Fix RX buffer size.

  • fm10k: Fix TX multi-segment frame.

  • fm10k: Fix TX queue cleaning after start error.

  • fm10k: Fix Tx queue cleaning after start error.

  • fm10k: Fix default mac/vlan in switch.

  • fm10k: Fix interrupt fault handling.

  • fm10k: Fix jumbo frame issue.

  • fm10k: Fix mac/vlan filtering.

  • fm10k: Fix maximum VF number.

  • fm10k: Fix maximum queue number for VF.

    Both PF and VF shared code in function fm10k_stats_get(). The function worked with PF, but had problems with VF since it has less queues than PF.

    Fixes: a6061d9e7075 (“fm10k: register PF driver”)

  • fm10k: Fix queue disabling.

  • fm10k: Fix switch synchronization.

  • i40e/base: Fix error handling of NVM state update.

  • i40e/base: Fix hardware port number for pass-through.

  • i40e/base: Rework virtual address retrieval for lan queue.

  • i40e/base: Update LED blinking.

  • i40e/base: Workaround for PHY type with firmware < 4.4.

  • i40e: Disable setting of PHY configuration.

  • i40e: Fix SCTP flow director.

  • i40e: Fix check of descriptor done flag.

    Fixes: 4861cde46116 (“i40e: new poll mode driver”) Fixes: 05999aab4ca6 (“i40e: add or delete flow director”)

  • i40e: Fix condition to get VMDQ info.

  • i40e: Fix registers access from big endian CPU.

  • i40evf: Clear command when error occurs.

  • i40evf: Fix RSS with less RX queues than TX queues.

  • i40evf: Fix crash when setup TX queues.

  • i40evf: Fix jumbo frame support.

  • i40evf: Fix offload capability flags.

    Added checksum offload capability flags which have already been supported for a long time.

  • ivshmem: Fix crash in corner case.

    Fixed issues where depending on the configured segments it was possible to hit a segmentation fault as a result of decrementing an unsigned index with value 0.

    Fixes: 40b966a211ab (“ivshmem: library changes for mmaping using ivshmem”)

  • ixgbe/base: Fix SFP probing.

  • ixgbe/base: Fix TX pending clearing.

  • ixgbe/base: Fix X550 CS4227 address.

  • ixgbe/base: Fix X550 PCIe master disabling.

  • ixgbe/base: Fix X550 check.

  • ixgbe/base: Fix X550 init early return.

  • ixgbe/base: Fix X550 link speed.

  • ixgbe/base: Fix X550em CS4227 speed mode.

  • ixgbe/base: Fix X550em SFP+ link stability.

  • ixgbe/base: Fix X550em UniPHY link configuration.

  • ixgbe/base: Fix X550em flow control for KR backplane.

  • ixgbe/base: Fix X550em flow control to be KR only.

  • ixgbe/base: Fix X550em link setup without SFP.

  • ixgbe/base: Fix X550em mux after MAC reset.

    Fixes: d2e72774e58c (“ixgbe/base: support X550”)

  • ixgbe/base: Fix bus type overwrite.

  • ixgbe/base: Fix init handling of X550em link down.

  • ixgbe/base: Fix lan id before first i2c access.

  • ixgbe/base: Fix mac type checks.

  • ixgbe/base: Fix tunneled UDP and TCP frames in flow director.

  • ixgbe: Check mbuf refcnt when clearing a ring.

    The function to clear the TX ring when a port was being closed, e.g. on exit in testpmd, was not checking the mbuf refcnt before freeing it. Since the function in the vector driver to clear the ring after TX does not setting the pointer to NULL post-free, this caused crashes if mbuf debugging was turned on.

  • ixgbe: Fix RX with buffer address not word aligned.

    Niantic HW expects the Header Buffer Address in the RXD must be word aligned.

  • ixgbe: Fix RX with buffer address not word aligned.

  • ixgbe: Fix Rx queue reset.

    Fix to reset vector related RX queue fields to their initial values.

    Fixes: c95584dc2b18 (“ixgbe: new vectorized functions for Rx/Tx”)

  • ixgbe: Fix TSO in IPv6.

    When TSO was used with IPv6, the generated frames were incorrect. The L4 frame was OK, but the length field of IPv6 header was not populated correctly.

  • ixgbe: Fix X550 flow director check.

  • ixgbe: Fix check for split packets.

    The check for split packets to be reassembled in the vector ixgbe PMD was incorrectly only checking the first 16 elements of the array instead of all 32.

    Fixes: cf4b4708a88a (“ixgbe: improve slow-path perf with vector scattered Rx”)

  • ixgbe: Fix data access on big endian cpu.

  • ixgbe: Fix flow director flexbytes offset.

    Fixes: d54a9888267c (“ixgbe: support flexpayload configuration of flow director”)

  • ixgbe: Fix number of segments with vector scattered Rx.

    Fixes: cf4b4708a88a (ixgbe: improve slow-path perf with vector scattered Rx)

  • ixgbe: Fix offload config option name.

    The RX_OLFLAGS option was renamed from DISABLE to ENABLE in the driver code and Linux config. It is now renamed also in the BSD config and documentation.

    Fixes: 359f106a69a9 (“ixgbe: prefer enabling olflags rather than not disabling”)

  • ixgbe: Fix release queue mbufs.

    The calculations of what mbufs were valid in the RX and TX queues were incorrect when freeing the mbufs for the vector PMD. This led to crashes due to invalid reference counts when mbuf debugging was turned on, and possibly other more subtle problems (such as mbufs being freed when in use) in other cases.

    Fixes: c95584dc2b18 (“ixgbe: new vectorized functions for Rx/Tx”)

  • ixgbe: Move PMD specific fields out of base driver.

    Move rx_bulk_alloc_allowed and rx_vec_allowed from ixgbe_hw to ixgbe_adapter.

    Fixes: 01fa1d6215fa (“ixgbe: unify Rx setup”)

  • ixgbe: Rename TX queue release function.

  • ixgbevf: Fix RX function selection.

    The logic to select ixgbe the VF RX function is different than the PF.

  • ixgbevf: Fix link status for PF up/down events.

  • kni: Fix RX loop limit.

    Loop processing packets dequeued from rx_q was using the number of packets requested, not how many it actually received.

  • kni: Fix ioctl in containers, like Docker.

  • kni: Fix multicast ioctl handling.

  • log: Fix crash after log_history dump.

  • lpm: Fix big endian support.

  • lpm: Fix depth small entry add.

  • mbuf: Fix cloning with private mbuf data.

    Added a new priv_size field in mbuf structure that should be initialized at mbuf pool creation. This field contains the size of the application private data in mbufs.

    Introduced new static inline functions rte_mbuf_from_indirect() and rte_mbuf_to_baddr() to replace the existing macros, which take the private size into account when attaching and detaching mbufs.

  • mbuf: Fix data room size calculation in pool init.

    Deduct the mbuf data room size from mempool->elt_size and priv_size, instead of using an hardcoded value that is not related to the real buffer size.

    To use rte_pktmbuf_pool_init(), the user can either:

    • Give a NULL parameter to rte_pktmbuf_pool_init(): in this case, the private size is assumed to be 0, and the room size is mp->elt_size - sizeof(struct rte_mbuf).
    • Give the rte_pktmbuf_pool_private filled with appropriate data_room_size and priv_size values.
  • mbuf: Fix init when private size is not zero.

    Allow the user to use the default rte_pktmbuf_init() function even if the mbuf private size is not 0.

  • mempool: Add structure for object headers.

    Each object stored in mempools are prefixed by a header, allowing for instance to retrieve the mempool pointer from the object. When debug is enabled, a cookie is also added in this header that helps to detect corruptions and double-frees.

    Introduced a structure that materializes the content of this header, and will simplify future patches adding things in this header.

  • mempool: Fix pages computation to determine number of objects.

  • mempool: Fix returned value after counting objects.

    Fixes: 148f963fb532 (“xen: core library changes”)

  • mlx4: Avoid requesting TX completion events to improve performance.

    Instead of requesting a completion event for each TX burst, request it on a fixed schedule once every MLX4_PMD_TX_PER_COMP_REQ (currently 64) packets to improve performance.

  • mlx4: Fix compilation as a shared library and on 32 bit platforms.

  • mlx4: Fix possible crash on scattered mbuf allocation failure.

    Fixes issue where failing to allocate a segment, mlx4_rx_burst_sp() could call rte_pktmbuf_free() on an incomplete scattered mbuf whose next pointer in the last segment is not set.

  • mlx4: Fix support for multiple vlan filters.

    This fixes the “Multiple RX VLAN filters can be configured, but only the first one works” bug.

  • pcap: Fix storage of name and type in queues.

    pcap_rx_queue/pcap_tx_queue should store it’s own copy of name/type values, not the pointer to temporary allocated space.

  • pci: Fix memory leaks and needless increment of map address.

  • pci: Fix uio mapping differences between linux and bsd.

  • port: Fix unaligned access to metadata.

    Fix RTE_MBUF_METADATA macros to allow for unaligned accesses to meta-data fields.

  • ring: Fix return of new port id on creation.

  • timer: Fix race condition.

    Eliminate problematic race condition in rte_timer_manage() that can lead to corruption of per-lcore pending-lists (implemented as skip-lists).

  • vfio: Fix overflow of BAR region offset and size.

    Fixes: 90a1633b2347 (“eal/Linux: allow to map BARs with MSI-X tables”)

  • vhost: Fix enqueue/dequeue to handle chained vring descriptors.

  • vhost: Fix race for connection fd.

  • vhost: Fix virtio freeze due to missed interrupt.

  • virtio: Fix crash if CQ is not negotiated.

    Fix NULL dereference if virtio control queue is not negotiated.

  • virtio: Fix ring size negotiation.

    Negotiate the virtio ring size. The host may allow for very large rings but application may only want a smaller ring. Conversely, if the number of descriptors requested exceeds the virtio host queue size, then just silently use the smaller host size.

    This fixes issues with virtio in non-QEMU environments. For example Google Compute Engine allows up to 16K elements in ring.

  • vmxnet3: Fix link state handling.

Known Issues

  • When running the vmdq sample or vhost sample applications with the Intel(R) XL710 (i40e) NIC, the configuration option CONFIG_RTE_MAX_QUEUES_PER_PORT should be increased from 256 to 1024.
  • VM power manager may not work on systems with more than 64 cores.

API Changes

  • The order that user supplied RX and TX callbacks are called in has been changed to the order that they were added (fifo) in line with end-user expectations. The previous calling order was the reverse of this (lifo) and was counter intuitive for users. The actual API is unchanged.

ABI Changes

  • The rte_hash structure has been changed to internal use only.

DPDK Release 2.0

New Features

  • Poll-mode driver support for an early release of the PCIE host interface of the Intel(R) Ethernet Switch FM10000.
    • Basic Rx/Tx functions for PF/VF
    • Interrupt handling support for PF/VF
    • Per queue start/stop functions for PF/VF
    • Support Mailbox handling between PF/VF and PF/Switch Manager
    • Receive Side Scaling (RSS) for PF/VF
    • Scatter receive function for PF/VF
    • Reta update/query for PF/VF
    • VLAN filter set for PF
    • Link status query for PF/VF

Note

The software is intended to run on pre-release hardware and may contain unknown or unresolved defects or issues related to functionality and performance. The poll mode driver is also pre-release and will be updated to a released version post hardware and base driver release. Should the official hardware release be made between DPDK releases an updated poll-mode driver will be made available.

  • Link Bonding

    • Support for adaptive load balancing (mode 6) to the link bonding library.
    • Support for registration of link status change callbacks with link bonding devices.
    • Support for slaves devices which do not support link status change interrupts in the link bonding library via a link status polling mechanism.
  • PCI Hotplug with NULL PMD sample application

  • ABI versioning

  • x32 ABI

  • Non-EAL Thread Support

  • Multi-pthread Support

  • Re-order Library

  • ACL for AVX2

  • Architecture Independent CRC Hash

  • uio_pci_generic Support

  • KNI Optimizations

  • Vhost-user support

  • Virtio (link, vlan, mac, port IO, perf)

  • IXGBE-VF RSS

  • RX/TX Callbacks

  • Unified Flow Types

  • Indirect Attached MBUF Flag

  • Use default port configuration in TestPMD

  • Tunnel offloading in TestPMD

  • Poll Mode Driver - 40 GbE Controllers (librte_pmd_i40e)

    • Support for Flow Director
    • Support for ethertype filter
    • Support RSS in VF
    • Support configuring redirection table with different size from 1GbE and 10 GbE
    • 128/512 entries of 40GbE PF
    • 64 entries of 40GbE VF
    • Support configuring hash functions
    • Support for VXLAN packet on Intel® 40GbE Controllers
  • Poll Mode Driver for Mellanox ConnectX-3 EN adapters (mlx4)

Note

This PMD is only available for Linux and is disabled by default due to external dependencies (libibverbs and libmlx4). Please refer to the NIC drivers guide for more information.

  • Packet Distributor Sample Application
  • Job Stats library and Sample Application.
  • Enhanced Jenkins hash (jhash) library

Note

The hash values returned by the new jhash library are different from the ones returned by the previous library.

DPDK Release 1.8

New Features

  • Link Bonding

    • Support for 802.3ad link aggregation (mode 4) and transmit load balancing (mode 5) to the link bonding library.
    • Support for registration of link status change callbacks with link bonding devices.
    • Support for slaves devices which do not support link status change interrupts in the link bonding library via a link status polling mechanism.
  • Poll Mode Driver - 40 GbE Controllers (librte_pmd_i40e)

    • Support for Flow Director
    • Support for ethertype filter
    • Support RSS in VF
    • Support configuring redirection table with different size from 1GbE and 10 GbE
    • 128/512 entries of 40GbE PF
    • 64 entries of 40GbE VF
    • Support configuring hash functions
    • Support for VXLAN packet on Intel 40GbE Controllers
  • Packet Distributor Sample Application

Known Issues and Limitations in Legacy Releases

This section describes known issues with the DPDK software that aren’t covered in the version specific release notes sections.

Pause Frame Forwarding does not work properly on igb

Description:
For igb devices rte_eth_flow_ctrl_set does not work as expected. Pause frames are always forwarded on igb, regardless of the RFCE, MPMCF and DPF registers.
Implication:
Pause frames will never be rejected by the host on 1G NICs and they will always be forwarded.
Resolution/Workaround:
There is no workaround available.
Affected Environment/Platform:
All.
Driver/Module:
Poll Mode Driver (PMD).

In packets provided by the PMD, some flags are missing

Description:
In packets provided by the PMD, some flags are missing. The application does not have access to information provided by the hardware (packet is broadcast, packet is multicast, packet is IPv4 and so on).
Implication:
The ol_flags field in the rte_mbuf structure is not correct and should not be used.
Resolution/Workaround:
The application has to parse the Ethernet header itself to get the information, which is slower.
Affected Environment/Platform:
All.
Driver/Module:
Poll Mode Driver (PMD).

The rte_malloc library is not fully implemented

Description:
The rte_malloc library is not fully implemented.
Implication:
All debugging features of rte_malloc library described in architecture documentation are not yet implemented.
Resolution/Workaround:
No workaround available.
Affected Environment/Platform:
All.
Driver/Module:
rte_malloc.

HPET reading is slow

Description:
Reading the HPET chip is slow.
Implication:
An application that calls rte_get_hpet_cycles() or rte_timer_manage() runs slower.
Resolution/Workaround:
The application should not call these functions too often in the main loop. An alternative is to use the TSC register through rte_rdtsc() which is faster, but specific to an lcore and is a cycle reference, not a time reference.
Affected Environment/Platform:
All.
Driver/Module:
Environment Abstraction Layer (EAL).

HPET timers do not work on the Osage customer reference platform

Description:
HPET timers do not work on the Osage customer reference platform which includes an Intel® Xeon® processor 5500 series processor) using the released BIOS from Intel.
Implication:
On Osage boards, the implementation of the rte_delay_us() function must be changed to not use the HPET timer.
Resolution/Workaround:
This can be addressed by building the system with the CONFIG_RTE_LIBEAL_USE_HPET=n configuration option or by using the --no-hpet EAL option.
Affected Environment/Platform:
The Osage customer reference platform. Other vendor platforms with Intel® Xeon® processor 5500 series processors should work correctly, provided the BIOS supports HPET.
Driver/Module:
lib/librte_eal/common/include/rte_cycles.h

Not all variants of supported NIC types have been used in testing

Description:

The supported network interface cards can come in a number of variants with different device ID’s. Not all of these variants have been tested with the DPDK.

The NIC device identifiers used during testing:

  • Intel® Ethernet Controller XL710 for 40GbE QSFP+ [8086:1584]
  • Intel® Ethernet Controller XL710 for 40GbE QSFP+ [8086:1583]
  • Intel® Ethernet Controller X710 for 10GbE SFP+ [8086:1572]
  • Intel® 82576 Gigabit Ethernet Controller [8086:10c9]
  • Intel® 82576 Quad Copper Gigabit Ethernet Controller [8086:10e8]
  • Intel® 82580 Dual Copper Gigabit Ethernet Controller [8086:150e]
  • Intel® I350 Quad Copper Gigabit Ethernet Controller [8086:1521]
  • Intel® 82599 Dual Fibre 10 Gigabit Ethernet Controller [8086:10fb]
  • Intel® Ethernet Server Adapter X520-T2 [8086: 151c]
  • Intel® Ethernet Controller X540-T2 [8086:1528]
  • Intel® 82574L Gigabit Network Connection [8086:10d3]
  • Emulated Intel® 82540EM Gigabit Ethernet Controller [8086:100e]
  • Emulated Intel® 82545EM Gigabit Ethernet Controller [8086:100f]
  • Intel® Ethernet Server Adapter X520-4 [8086:154a]
  • Intel® Ethernet Controller I210 [8086:1533]
Implication:
Risk of issues with untested variants.
Resolution/Workaround:
Use tested NIC variants. For those supported Ethernet controllers, additional device IDs may be added to the software if required.
Affected Environment/Platform:
All.
Driver/Module:
Poll-mode drivers

Multi-process sample app requires exact memory mapping

Description:
The multi-process example application assumes that it is possible to map the hugepage memory to the same virtual addresses in client and server applications. Occasionally, very rarely with 64-bit, this does not occur and a client application will fail on startup. The Linux “address-space layout randomization” security feature can sometimes cause this to occur.
Implication:
A multi-process client application fails to initialize.
Resolution/Workaround:
See the “Multi-process Limitations” section in the DPDK Programmer’s Guide for more information.
Affected Environment/Platform:
All.
Driver/Module:
Multi-process example application

Packets are not sent by the 1 GbE/10 GbE SR-IOV driver when the source MAC is not the MAC assigned to the VF NIC

Description:
The 1 GbE/10 GbE SR-IOV driver can only send packets when the Ethernet header’s source MAC address is the same as that of the VF NIC. The reason for this is that the Linux ixgbe driver module in the host OS has its anti-spoofing feature enabled.
Implication:
Packets sent using the 1 GbE/10 GbE SR-IOV driver must have the source MAC address correctly set to that of the VF NIC. Packets with other source address values are dropped by the NIC if the application attempts to transmit them.
Resolution/Workaround:
Configure the Ethernet source address in each packet to match that of the VF NIC.
Affected Environment/Platform:
All.
Driver/Module:
1 GbE/10 GbE VF Poll Mode Driver (PMD).

SR-IOV drivers do not fully implement the rte_ethdev API

Description:

The SR-IOV drivers only supports the following rte_ethdev API functions:

  • rte_eth_dev_configure()
  • rte_eth_tx_queue_setup()
  • rte_eth_rx_queue_setup()
  • rte_eth_dev_info_get()
  • rte_eth_dev_start()
  • rte_eth_tx_burst()
  • rte_eth_rx_burst()
  • rte_eth_dev_stop()
  • rte_eth_stats_get()
  • rte_eth_stats_reset()
  • rte_eth_link_get()
  • rte_eth_link_get_no_wait()
Implication:
Calling an unsupported function will result in an application error.
Resolution/Workaround:
Do not use other rte_ethdev API functions in applications that use the SR-IOV drivers.
Affected Environment/Platform:
All.
Driver/Module:
VF Poll Mode Driver (PMD).

PMD does not work with –no-huge EAL command line parameter

Description:
Currently, the DPDK does not store any information about memory allocated by malloc()` (for example, NUMA node, physical address), hence PMD drivers do not work when the ``--no-huge command line parameter is supplied to EAL.
Implication:
Sending and receiving data with PMD will not work.
Resolution/Workaround:
Use huge page memory or use VFIO to map devices.
Affected Environment/Platform:
Systems running the DPDK on Linux
Driver/Module:
Poll Mode Driver (PMD).

Some hardware off-load functions are not supported by the VF Driver

Description:

Currently, configuration of the following items is not supported by the VF driver:

  • IP/UDP/TCP checksum offload
  • Jumbo Frame Receipt
  • HW Strip CRC
Implication:
Any configuration for these items in the VF register will be ignored. The behavior is dependent on the current PF setting.
Resolution/Workaround:
For the PF (Physical Function) status on which the VF driver depends, there is an option item under PMD in the config file. For others, the VF will keep the same behavior as PF setting.
Affected Environment/Platform:
All.
Driver/Module:
VF (SR-IOV) Poll Mode Driver (PMD).

Kernel crash on IGB port unbinding

Description:
Kernel crash may occur when unbinding 1G ports from the igb_uio driver, on 2.6.3x kernels such as shipped with Fedora 14.
Implication:
Kernel crash occurs.
Resolution/Workaround:
Use newer kernels or do not unbind ports.
Affected Environment/Platform:
2.6.3x kernels such as shipped with Fedora 14
Driver/Module:
IGB Poll Mode Driver (PMD).

Discrepancies between statistics reported by different NICs

Description:
Gigabit Ethernet devices from Intel include CRC bytes when calculating packet reception statistics regardless of hardware CRC stripping state, while 10-Gigabit Ethernet devices from Intel do so only when hardware CRC stripping is disabled.
Implication:
There may be a discrepancy in how different NICs display packet reception statistics.
Resolution/Workaround:
None
Affected Environment/Platform:
All.
Driver/Module:
Poll Mode Driver (PMD).

Error reported opening files on DPDK initialization

Description:
On DPDK application startup, errors may be reported when opening files as part of the initialization process. This occurs if a large number, for example, 500 or more, or if hugepages are used, due to the per-process limit on the number of open files.
Implication:
The DPDK application may fail to run.
Resolution/Workaround:
If using 2 MB hugepages, consider switching to a fewer number of 1 GB pages. Alternatively, use the ulimit command to increase the number of files which can be opened by a process.
Affected Environment/Platform:
All.
Driver/Module:
Environment Abstraction Layer (EAL).

Intel® QuickAssist Technology sample application does not work on a 32-bit OS on Shumway

Description:
The Intel® Communications Chipset 89xx Series device does not fully support NUMA on a 32-bit OS. Consequently, the sample application cannot work properly on Shumway, since it requires NUMA on both nodes.
Implication:
The sample application cannot work in 32-bit mode with emulated NUMA, on multi-socket boards.
Resolution/Workaround:
There is no workaround available.
Affected Environment/Platform:
Shumway
Driver/Module:
All.

Differences in how different Intel NICs handle maximum packet length for jumbo frame

Description:
10 Gigabit Ethernet devices from Intel do not take VLAN tags into account when calculating packet size while Gigabit Ethernet devices do so for jumbo frames.
Implication:
When receiving packets with VLAN tags, the actual maximum size of useful payload that Intel Gigabit Ethernet devices are able to receive is 4 bytes (or 8 bytes in the case of packets with extended VLAN tags) less than that of Intel 10 Gigabit Ethernet devices.
Resolution/Workaround:
Increase the configured maximum packet size when using Intel Gigabit Ethernet devices.
Affected Environment/Platform:
All.
Driver/Module:
Poll Mode Driver (PMD).

Binding PCI devices to igb_uio fails on Linux kernel 3.9 when more than one device is used

Description:
A known bug in the uio driver included in Linux kernel version 3.9 prevents more than one PCI device to be bound to the igb_uio driver.
Implication:
The Poll Mode Driver (PMD) will crash on initialization.
Resolution/Workaround:
Use earlier or later kernel versions, or apply the following patch.
Affected Environment/Platform:
Linux systems with kernel version 3.9
Driver/Module:
igb_uio module

GCC might generate Intel® AVX instructions for processors without Intel® AVX support

Description:
When compiling DPDK (and any DPDK app), gcc may generate Intel® AVX instructions, even when the processor does not support Intel® AVX.
Implication:
Any DPDK app might crash while starting up.
Resolution/Workaround:
Either compile using icc or set EXTRA_CFLAGS='-O3' prior to compilation.
Affected Environment/Platform:
Platforms which processor does not support Intel® AVX.
Driver/Module:
Environment Abstraction Layer (EAL).

Ethertype filter could receive other packets (non-assigned) in Niantic

Description:

On Intel® Ethernet Controller 82599EB When Ethertype filter (priority enable) was set, unmatched packets also could be received on the assigned queue, such as ARP packets without 802.1q tags or with the user priority not equal to set value. Launch the testpmd by disabling RSS and with multiply queues, then add the ethertype filter like the following and then start forwarding:

add_ethertype_filter 0 ethertype 0x0806 priority enable 3 queue 2 index 1

When sending ARP packets without 802.1q tag and with user priority as non-3 by tester, all the ARP packets can be received on the assigned queue.

Implication:
The user priority comparing in Ethertype filter cannot work probably. It is a NIC’s issue due to the following: “In fact, ETQF.UP is not functional, and the information will be added in errata of 82599 and X540.”
Resolution/Workaround:
None
Affected Environment/Platform:
All.
Driver/Module:
Poll Mode Driver (PMD).

Devices bound to igb_uio with VT-d enabled do not work on Linux kernel 3.15-3.17

Description:
When VT-d is enabled (iommu=pt intel_iommu=on), devices are 1:1 mapped. In the Linux kernel unbinding devices from drivers removes that mapping which result in IOMMU errors. Introduced in Linux kernel 3.15 commit, solved in Linux kernel 3.18 commit.
Implication:

Devices will not be allowed to access memory, resulting in following kernel errors:

dmar: DRHD: handling fault status reg 2
dmar: DMAR:[DMA Read] Request device [02:00.0] fault addr a0c58000
DMAR:[fault reason 02] Present bit in context entry is clear
Resolution/Workaround:
Use earlier or later kernel versions, or avoid driver binding on boot by blacklisting the driver modules. I.e., in the case of ixgbe, we can pass the kernel command line option: modprobe.blacklist=ixgbe. This way we do not need to unbind the device to bind it to igb_uio.
Affected Environment/Platform:
Linux systems with kernel versions 3.15 to 3.17.
Driver/Module:
igb_uio module.

VM power manager may not work on systems with more than 64 cores

Description:
When using VM power manager on a system with more than 64 cores, VM(s) should not use cores 64 or higher.
Implication:
VM power manager should not be used with VM(s) that are using cores 64 or above.
Resolution/Workaround:
Do not use cores 64 or above.
Affected Environment/Platform:
Platforms with more than 64 cores.
Driver/Module:
VM power manager application.

DPDK may not build on some Intel CPUs using clang < 3.7.0

Description:
When compiling DPDK with an earlier version than 3.7.0 of clang, CPU flags are not detected on some Intel platforms such as Intel Broadwell/Skylake (and possibly future CPUs), and therefore compilation fails due to missing intrinsics.
Implication:
DPDK will not build when using a clang version < 3.7.0.
Resolution/Workaround:
Use clang 3.7.0 or higher, or gcc.
Affected Environment/Platform:
Platforms with Intel Broadwell/Skylake using an old clang version.
Driver/Module:
Environment Abstraction Layer (EAL).

The last EAL argument is replaced by the program name in argv[]

Description:
The last EAL argument is replaced by program name in argv[] after eal_parse_args is called. This is the intended behavior but it causes the pointer to the last EAL argument to be lost.
Implication:
If the last EAL argument in argv[] is generated by a malloc function, changing it will cause memory issues when freeing the argument.
Resolution/Workaround:
An application should not consider the value in argv[] as unchanged.
Affected Environment/Platform:
ALL.
Driver/Module:
Environment Abstraction Layer (EAL).

I40e VF may not receive packets in the promiscuous mode

Description:
Promiscuous mode is not supported by the DPDK i40e VF driver when using the i40e Linux kernel driver as host driver.
Implication:
The i40e VF does not receive packets when the destination MAC address is unknown.
Resolution/Workaround:
Use a explicit destination MAC address that matches the VF.
Affected Environment/Platform:
All.
Driver/Module:
Poll Mode Driver (PMD).

uio pci generic module bind failed in X710/XL710/XXV710

Description:
The uio_pci_generic module is not supported by XL710, since the errata of XL710 states that the Interrupt Status bit is not implemented. The errata is the item #71 from the xl710 controller spec. The hw limitation is the same as other X710/XXV710 NICs.
Implication:
When use --bind=uio_pci_generic, the uio_pci_generic module probes device and check the Interrupt Status bit. Since it is not supported by X710/XL710/XXV710, it return a failed value. The statement that these products don’t support INTx masking, is indicated in the related linux kernel commit.
Resolution/Workaround:
Do not bind the uio_pci_generic module in X710/XL710/XXV710 NICs.
Affected Environment/Platform:
All.
Driver/Module:
Poll Mode Driver (PMD).

virtio tx_burst() function cannot do TSO on shared packets

Description:

The standard TX function of virtio driver does not manage shared packets properly when doing TSO. These packets should be read-only but the driver modifies them.

When doing TSO, the virtio standard expects that the L4 checksum is set to the pseudo header checksum in the packet data, which is different than the DPDK API. The driver patches the L4 checksum to conform to the virtio standard, but this solution is invalid when dealing with shared packets (clones), because the packet data should not be modified.

Implication:
In this situation, the shared data will be modified by the driver, potentially causing race conditions with the other users of the mbuf data.
Resolution/Workaround:
The workaround in the application is to ensure that the network headers in the packet data are not shared.
Affected Environment/Platform:
Virtual machines running a virtio driver.
Driver/Module:
Poll Mode Driver (PMD).

igb uio legacy mode can not be used in X710/XL710/XXV710

Description:
X710/XL710/XXV710 NICs lack support for indicating INTx is asserted via the interrupt bit in the PCI status register. Linux delected them from INTx support table. The related commit.
Implication:
When insmod igb_uio with intr_mode=legacy and test link status interrupt. Since INTx interrupt is not supported by X710/XL710/XXV710, it will cause Input/Output error when reading file descriptor.
Resolution/Workaround:
Do not bind igb_uio with legacy mode in X710/XL710/XXV710 NICs, or do not use kernel version >4.7 when you bind igb_uio with legacy mode.
Affected Environment/Platform:
ALL.
Driver/Module:
Poll Mode Driver (PMD).

igb_uio can not be used when running l3fwd-power

Description:
Link Status Change(LSC) interrupt and packet receiving interrupt are all enabled in l3fwd-power APP. Because of UIO only support one interrupt, so these two kinds of interrupt need to share one, and the receiving interrupt have the higher priority, so can’t get the right link status.
Implication:
When insmod igb_uio and running l3fwd-power APP, link status getting doesn’t work properly.
Resolution/Workaround:
Use vfio-pci when LSC and packet receiving interrupt enabled.
Affected Environment/Platform:
ALL.
Driver/Module:
igb_uio module.

ABI and API Deprecation

See the guidelines document for details of the ABI policy. API and ABI deprecation notices are to be posted here.

Deprecation Notices

  • eal: several API and ABI changes are planned for rte_devargs in v18.02. The format of device command line parameters will change. The bus will need to be explicitly stated in the device declaration. The enum rte_devtype was used to identify a bus and will disappear. The structure rte_devargs will change. The rte_devargs_list will be made private. The following functions are deprecated starting from 17.08 and will either be modified or removed in 18.02:
    • rte_eal_devargs_add
    • rte_eal_devargs_type_count
    • rte_eal_parse_devargs_str, replaced by rte_eal_devargs_parse
  • pci: Several exposed functions are misnamed. The following functions are deprecated starting from v17.11 and are replaced:
    • eal_parse_pci_BDF replaced by rte_pci_addr_parse
    • eal_parse_pci_DomBDF replaced by rte_pci_addr_parse
    • rte_eal_compare_pci_addr replaced by rte_pci_addr_cmp
  • ethdev: a new Tx and Rx offload API was introduced on 17.11. In the new API, offloads are divided into per-port and per-queue offloads. Offloads are disabled by default and enabled per application request. The old offloads API is target to be deprecated on 18.05. This includes:
    • removal of ETH_TXQ_FLAGS_NO* flags.
    • removal of txq_flags field from rte_eth_txconf struct.
    • removal of the offloads bit-field from rte_eth_rxmode struct.
  • ethdev: the legacy filter API, including rte_eth_dev_filter_supported(), rte_eth_dev_filter_ctrl() as well as filter types MACVLAN, ETHERTYPE, FLEXIBLE, SYN, NTUPLE, TUNNEL, FDIR, HASH and L2_TUNNEL, is superseded by the generic flow API (rte_flow) in PMDs that implement the latter. Target release for removal of the legacy API will be defined once most PMDs have switched to rte_flow.
  • i40e: The default flexible payload configuration which extracts the first 16 bytes of the payload for RSS will be deprecated starting from 18.02. If required the previous behavior can be configured using existing flow director APIs. There is no ABI/API break. This change will just remove a global configuration setting and require explicit configuration.
  • librte_meter: The API will change to accommodate configuration profiles. Most of the API functions will have an additional opaque parameter.

FAQ

This document contains some Frequently Asked Questions that arise when working with DPDK.

What does “EAL: map_all_hugepages(): open failed: Permission denied Cannot init memory” mean?

This is most likely due to the test application not being run with sudo to promote the user to a superuser. Alternatively, applications can also be run as regular user. For more information, please refer to DPDK Getting Started Guide.

If I want to change the number of hugepages allocated, how do I remove the original pages allocated?

The number of pages allocated can be seen by executing the following command:

grep Huge /proc/meminfo

Once all the pages are mmapped by an application, they stay that way. If you start a test application with less than the maximum, then you have free pages. When you stop and restart the test application, it looks to see if the pages are available in the /dev/huge directory and mmaps them. If you look in the directory, you will see n number of 2M pages files. If you specified 1024, you will see 1024 page files. These are then placed in memory segments to get contiguous memory.

If you need to change the number of pages, it is easier to first remove the pages. The usertools/dpdk-setup.sh script provides an option to do this. See the “Quick Start Setup Script” section in the DPDK Getting Started Guide for more information.

If I execute “l2fwd -l 0-3 -m 64 -n 3 – -p 3”, I get the following output, indicating that there are no socket 0 hugepages to allocate the mbuf and ring structures to?

I have set up a total of 1024 Hugepages (that is, allocated 512 2M pages to each NUMA node).

The -m command line parameter does not guarantee that huge pages will be reserved on specific sockets. Therefore, allocated huge pages may not be on socket 0. To request memory to be reserved on a specific socket, please use the –socket-mem command-line parameter instead of -m.

I am running a 32-bit DPDK application on a NUMA system, and sometimes the application initializes fine but cannot allocate memory. Why is that happening?

32-bit applications have limitations in terms of how much virtual memory is available, hence the number of hugepages they are able to allocate is also limited (1 GB size). If your system has a lot (>1 GB size) of hugepage memory, not all of it will be allocated. Due to hugepages typically being allocated on a local NUMA node, the hugepages allocation the application gets during the initialization depends on which NUMA node it is running on (the EAL does not affinitize cores until much later in the initialization process). Sometimes, the Linux OS runs the DPDK application on a core that is located on a different NUMA node from DPDK master core and therefore all the hugepages are allocated on the wrong socket.

To avoid this scenario, either lower the amount of hugepage memory available to 1 GB size (or less), or run the application with taskset affinitizing the application to a would-be master core.

For example, if your EAL coremask is 0xff0, the master core will usually be the first core in the coremask (0x10); this is what you have to supply to taskset:

taskset 0x10 ./l2fwd -l 4-11 -n 2

In this way, the hugepages have a greater chance of being allocated to the correct socket. Additionally, a --socket-mem option could be used to ensure the availability of memory for each socket, so that if hugepages were allocated on the wrong socket, the application simply will not start.

On application startup, there is a lot of EAL information printed. Is there any way to reduce this?

Yes, the option --log-level= accepts one of these numbers:

#define RTE_LOG_EMERG 1U    /* System is unusable. */
#define RTE_LOG_ALERT 2U    /* Action must be taken immediately. */
#define RTE_LOG_CRIT 3U     /* Critical conditions. */
#define RTE_LOG_ERR 4U      /* Error conditions. */
#define RTE_LOG_WARNING 5U  /* Warning conditions. */
#define RTE_LOG_NOTICE 6U   /* Normal but significant condition. */
#define RTE_LOG_INFO 7U     /* Informational. */
#define RTE_LOG_DEBUG 8U    /* Debug-level messages. */

It is also possible to change the default level at compile time with CONFIG_RTE_LOG_LEVEL.

How can I tune my network application to achieve lower latency?

Traditionally, there is a trade-off between throughput and latency. An application can be tuned to achieve a high throughput, but the end-to-end latency of an average packet typically increases as a result. Similarly, the application can be tuned to have, on average, a low end-to-end latency at the cost of lower throughput.

To achieve higher throughput, the DPDK attempts to aggregate the cost of processing each packet individually by processing packets in bursts. Using the testpmd application as an example, the “burst” size can be set on the command line to a value of 32 (also the default value). This allows the application to request 32 packets at a time from the PMD. The testpmd application then immediately attempts to transmit all the packets that were received, in this case, all 32 packets. The packets are not transmitted until the tail pointer is updated on the corresponding TX queue of the network port. This behavior is desirable when tuning for high throughput because the cost of tail pointer updates to both the RX and TX queues can be spread across 32 packets, effectively hiding the relatively slow MMIO cost of writing to the PCIe* device.

However, this is not very desirable when tuning for low latency, because the first packet that was received must also wait for the other 31 packets to be received. It cannot be transmitted until the other 31 packets have also been processed because the NIC will not know to transmit the packets until the TX tail pointer has been updated, which is not done until all 32 packets have been processed for transmission.

To consistently achieve low latency even under heavy system load, the application developer should avoid processing packets in bunches. The testpmd application can be configured from the command line to use a burst value of 1. This allows a single packet to be processed at a time, providing lower latency, but with the added cost of lower throughput.

Without NUMA enabled, my network throughput is low, why?

I have a dual Intel® Xeon® E5645 processors 2.40 GHz with four Intel® 82599 10 Gigabit Ethernet NICs. Using eight logical cores on each processor with RSS set to distribute network load from two 10 GbE interfaces to the cores on each processor.

Without NUMA enabled, memory is allocated from both sockets, since memory is interleaved. Therefore, each 64B chunk is interleaved across both memory domains.

The first 64B chunk is mapped to node 0, the second 64B chunk is mapped to node 1, the third to node 0, the fourth to node 1. If you allocated 256B, you would get memory that looks like this:

256B buffer
Offset 0x00 - Node 0
Offset 0x40 - Node 1
Offset 0x80 - Node 0
Offset 0xc0 - Node 1

Therefore, packet buffers and descriptor rings are allocated from both memory domains, thus incurring QPI bandwidth accessing the other memory and much higher latency. For best performance with NUMA disabled, only one socket should be populated.

I am getting errors about not being able to open files. Why?

As the DPDK operates, it opens a lot of files, which can result in reaching the open files limits, which is set using the ulimit command or in the limits.conf file. This is especially true when using a large number (>512) of 2 MB huge pages. Please increase the open file limit if your application is not able to open files. This can be done either by issuing a ulimit command or editing the limits.conf file. Please consult Linux manpages for usage information.

VF driver for IXGBE devices cannot be initialized

Some versions of Linux IXGBE driver do not assign a random MAC address to VF devices at initialization. In this case, this has to be done manually on the VM host, using the following command:

ip link set <interface> vf <VF function> mac <MAC address>

where <interface> being the interface providing the virtual functions for example, eth0, <VF function> being the virtual function number, for example 0, and <MAC address> being the desired MAC address.

Is it safe to add an entry to the hash table while running?

Currently the table implementation is not a thread safe implementation and assumes that locking between threads and processes is handled by the user’s application. This is likely to be supported in future releases.

What is the purpose of setting iommu=pt?

DPDK uses a 1:1 mapping and does not support IOMMU. IOMMU allows for simpler VM physical address translation. The second role of IOMMU is to allow protection from unwanted memory access by an unsafe device that has DMA privileges. Unfortunately, the protection comes with an extremely high performance cost for high speed NICs.

Setting iommu=pt disables IOMMU support for the hypervisor.

When trying to send packets from an application to itself, meaning smac==dmac, using Intel(R) 82599 VF packets are lost.

Check on register LLE(PFVMTXSSW[n]), which allows an individual pool to send traffic and have it looped back to itself.

Can I split packet RX to use DPDK and have an application’s higher order functions continue using Linux pthread?

The DPDK’s lcore threads are Linux pthreads bound onto specific cores. Configure the DPDK to do work on the same cores and run the application’s other work on other cores using the DPDK’s “coremask” setting to specify which cores it should launch itself on.

Is it possible to exchange data between DPDK processes and regular userspace processes via some shared memory or IPC mechanism?

Yes - DPDK processes are regular Linux/BSD processes, and can use all OS provided IPC mechanisms.

Can the multiple queues in Intel(R) I350 be used with DPDK?

I350 has RSS support and 8 queue pairs can be used in RSS mode. It should work with multi-queue DPDK applications using RSS.

How can hugepage-backed memory be shared among multiple processes?

See the Primary and Secondary examples in the multi-process sample application.

Why can’t my application receive packets on my system with UEFI Secure Boot enabled?

If UEFI secure boot is enabled, the Linux kernel may disallow the use of UIO on the system. Therefore, devices for use by DPDK should be bound to the vfio-pci kernel module rather than igb_uio or uio_pci_generic.