summaryrefslogtreecommitdiff
path: root/doc/guides/nics/nfb.rst
blob: 8df76c077e40e997efef2c58bd56910790415128 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
..  SPDX-License-Identifier: BSD-3-Clause
    Copyright 2019 Cesnet
    Copyright 2019 Netcope Technologies

NFB poll mode driver library
=================================

The NFB poll mode driver library implements support for the Netcope
FPGA Boards (**NFB-***), FPGA-based programmable NICs.
The NFB PMD uses interface provided by the libnfb library to communicate
with the NFB cards over the nfb layer.

More information about the
`NFB cards <http://www.netcope.com/en/products/fpga-boards>`_
and used technology
(`Netcope Development Kit <http://www.netcope.com/en/products/fpga-development-kit>`_)
can be found on the `Netcope Technologies website <http://www.netcope.com/>`_.

.. note::

   This driver has external dependencies.
   Therefore it is disabled in default configuration files.
   It can be enabled by setting ``CONFIG_RTE_LIBRTE_NFB_PMD=y``
   and recompiling.

.. note::

   Currently the driver is supported only on x86_64 architectures.
   Only x86_64 versions of the external libraries are provided.

Prerequisites
-------------

This PMD requires kernel modules which are responsible for initialization and
allocation of resources needed for nfb layer function.
Communication between PMD and kernel modules is mediated by libnfb library.
These kernel modules and library are not part of DPDK and must be installed
separately:

*  **libnfb library**

   The library provides API for initialization of nfb transfers, receiving and
   transmitting data segments.

*  **Kernel modules**

   * nfb

   Kernel modules manage initialization of hardware, allocation and
   sharing of resources for user space applications.

Dependencies can be found here:
`Netcope common <https://www.netcope.com/en/company/community-support/dpdk-libsze2#NFB>`_.

Versions of the packages
~~~~~~~~~~~~~~~~~~~~~~~~

The minimum version of the provided packages:

* for DPDK from 19.05

Configuration
-------------

These configuration options can be modified before compilation in the
``.config`` file:

*  ``CONFIG_RTE_LIBRTE_NFB_PMD`` default value: **n**

   Value **y** enables compilation of nfb PMD.

Using the NFB PMD
----------------------

Kernel modules have to be loaded before running the DPDK application.

NFB card architecture
---------------------

The NFB cards are multi-port multi-queue cards, where (generally) data from any
Ethernet port may be sent to any queue.
They are represented in DPDK as a single port.

NFB-200G2QL card employs an add-on cable which allows to connect it to two
physical PCI-E slots at the same time (see the diagram below).
This is done to allow 200 Gbps of traffic to be transferred through the PCI-E
bus (note that a single PCI-E 3.0 x16 slot provides only 125 Gbps theoretical
throughput).

Although each slot may be connected to a different CPU and therefore to a different
NUMA node, the card is represented as a single port in DPDK. To work with data
from the individual queues on the right NUMA node, connection of NUMA nodes on
first and last queue (each NUMA node has half of the queues) need to be checked.

.. figure:: img/szedata2_nfb200g_architecture.*
    :align: center

    NFB-200G2QL high-level diagram

Limitations
-----------

Driver is usable only on Linux architecture, namely on CentOS.

Since a card is always represented as a single port, but can be connected to two
NUMA nodes, there is need for manual check where master/slave is connected.

Example of usage
----------------

Read packets from 0. and 1. receive queue and write them to 0. and 1.
transmit queue:

.. code-block:: console

   $RTE_TARGET/app/testpmd -l 0-3 -n 2 \
   -- --port-topology=chained --rxq=2 --txq=2 --nb-cores=2 -i -a

Example output:

.. code-block:: console

   [...]
   EAL: PCI device 0000:06:00.0 on NUMA socket -1
   EAL:   probe driver: 1b26:c1c1 net_nfb
   PMD: Initializing NFB device (0000:06:00.0)
   PMD: Available DMA queues RX: 8 TX: 8
   PMD: NFB device (0000:06:00.0) successfully initialized
   Interactive-mode selected
   Auto-start selected
   Configuring Port 0 (socket 0)
   Port 0: 00:11:17:00:00:00
   Checking link statuses...
   Port 0 Link Up - speed 10000 Mbps - full-duplex
   Done
   Start automatic packet forwarding
     io packet forwarding - CRC stripping disabled - packets/burst=32
     nb forwarding cores=2 - nb forwarding ports=1
     RX queues=2 - RX desc=128 - RX free threshold=0
     RX threshold registers: pthresh=0 hthresh=0 wthresh=0
     TX queues=2 - TX desc=512 - TX free threshold=0
     TX threshold registers: pthresh=0 hthresh=0 wthresh=0
     TX RS bit threshold=0 - TXQ flags=0x0
   testpmd>