dperf is a DPDK based 100Gbps network performance and load testing software.
Alternatives To Dperf
Project NameStarsDownloadsRepos Using ThisPackages Using ThisMost Recent CommitTotal ReleasesLatest ReleaseOpen IssuesLicenseLanguage
F Stack3,428
11 days ago212otherC
F-Stack is an user space network development kit with high performance based on DPDK, FreeBSD TCP/IP stack and coroutine API.
2 days ago126otherC
DPVS is a high performance Layer-4 load balancer based on DPDK.
14 days ago1apache-2.0C
dperf is a DPDK based 100Gbps network performance and load testing software.
5 years ago1gpl-3.0C
A high performance authority-only dns server implemented with DPDK
4 years agoapache-2.0C++
dns server with high performance, based on dpdk.
2 years agoShell
NFVi Performance Lab using Red Hat OpenStack
5 years ago
Advanced extra high speed deep packet inspect library with private AFDL language supported
4 years agoC
Bolt [NSDI'19] software network function performance analysis tool
Dpdk Experiment5
4 years ago1Makefile
Collection of DPDK Experiments
F Stack5
6 years agoC
F-Stack is an open source high performance network framework based on DPDK.
Alternatives To Dperf
Select To Compare

Alternative Project Comparisons

dperf Apache V2 License

English |

dperf is a DPDK based 100Gbps network performance and load testing software.


  • High performance
    • Based on DPDK, dperf can generate huge traffic with a single x86 server: tens of millions of HTTP CPShundreds of Gbps throughput and billions of concurrent connections.
  • Detailed statistics
    • Provides detailed statistics and identify every packet loss.
  • Support multiple scenarios
    • Load testing and stability testing for Layer 4 Load Balancer and other Layer 4 gateways
    • Network performance testing for servers on cloud
    • Performance test of network package processing capability for NIC and CPU
    • Can be used as a high performance HTTP server or client for load testing


HTTP Connections per Second

Client Cores Server Cores HTTP CPS
1 1 2,101,044
2 2 4,000,423
4 4 7,010,743
6 6 10,027,172

HTTP Throughput per Second

Client Cores Server Cores RX(Gbps) TX(Gbps) Client CPU Usage(%) Server CPU Usage(%)
1 1 18 18 60 59
2 2 35 35 60 59
4 4 46 46 43 43

HTTP Current Connections

Client Cores Server Cores Current Connections Client CPU Usage(%) Server CPU Usage(%)
1 1 100,000,000 34 39
2 2 200,000,000 36 39
4 4 400,000,000 40 41


Client Cores TX MPPS Client CPU Usage(%)
1 15.96 95
2 29.95 95
4 34.92 67
6 35.92 54
8 37.12 22

Note: this test is under one 25Gbps Mellanox CX4

Client & Server Configuration

Above performance is obtained with below configurations:

  • MEM: 512GB(hugepage 100GB)
  • NIC: Mellanox MT27710 25Gbps * 2
  • Kernel: 4.19.90


dperf outputs various statistics every second

  • TPS, CPS, various PPS
  • Errors of TCP/Socket/HTTP
  • Packets loss/drop
  • Retransmissions of TCP Flags
seconds 22                 cpuUsage 52
pktRx   3,001,058          pktTx    3,001,025          bitsRx   2,272,799,040      bitsTx  1,920,657,600      dropTx  0
arpRx   0                  arpTx    0                  icmpRx   0                  icmpTx  0                  otherRx 0          badRx 0
synRx   1,000,345          synTx    1,000,330          finRx    1,000,350          finTx   1,000,350          rstRx   0          rstTx 0
synRt   0                  finRt    0                  ackRt    0                  pushRt  0                  tcpDrop 0
skOpen  1,000,330          skClose  1,000,363          skCon    230                skErr   0
httpGet 1,000,345          http2XX  1,000,350          httpErr  0
ierrors 0                  oerrors  0                  imissed  0

Getting Started

Set hugepages

#edit '/boot/grub2/grub.cfg' like this, and reboot the OS
linux16 /vmlinuz-... nopku transparent_hugepage=never default_hugepagesz=1G hugepagesz=1G hugepages=8

Build DPDK

#edit 'config/common_base' to enable PMDs
#Mellanox CX4/CX5 requires 'CONFIG_RTE_LIBRTE_MLX5_PMD=y'

TARGET=x86_64-native-linuxapp-gcc #or arm64-armv8a-linuxapp-gcc

cd /root/dpdk/dpdk-stable-19.11.10
make install T=$TARGET -j16

Build dperf

cd dperf
make -j8 RTE_SDK=/root/dpdk/dpdk-stable-19.11.10 RTE_TARGET=$TARGET

Bind interface

#Mellanox NIC skip this step!
#Suppose your PCI number is 0000:1b:00.0

modprobe uio
modprobe uio_pci_generic
/root/dpdk/dpdk-stable-19.11.10/usertools/dpdk-devbind.py -b uio_pci_generic 0000:1b:00.0

Start dperf server

#dperf server bind at,  gateway is
./build/dperf -c test/http/server-cps.conf

Send request from a client

#the client IP must be in the range of 'client' in the configuration file

Running the tests

Below example will start a HTTP CPS stress test.
#run dperf server ./build/dperf -c test/http/server-cps.conf

#from another host, run dperf client
./build/dperf -c test/http/client-cps.conf



  • dperf requires that the HTTP message is in one packet, which is not suitable for the test of layer 7 load balancer;
  • dperf requires exclusive use of the network interfaces.
  • dperf does not have routing capability. It is recommended to build a test environment with a switch.

Related Articles


dperf welcomes your contribution. See the CONTRIBUTING file for details. Welcome to the development team, please contact Jianzhang Peng to add you to the developer mailing list.


  • Jianzhang Peng - Initial work


dperf is distributed under the Apache License, Version 2.0.

Popular Performance Projects
Popular Dpdk Projects
Popular Software Performance Categories
Related Searches

Get A Weekly Email With Trending Projects For These Categories
No Spam. Unsubscribe easily at any time.