The options are described in detail in the rtpengine(1) man page. If you're reading this on Github, you can view the current master's man page here.
In normal userspace-only operation, the overhead involved in processing each individual RTP or media packet is quite significant. This comes from the fact that each time a packet is received on a network interface, the packet must first traverse the stack of the kernel's network protocols, down to locating a process's file descriptor. At this point the linked user process (the daemon) has to be signalled that a new packet is available to be read, the process has to be scheduled to run, once running the process must read the packet, which means it must be copied from kernel space to user space, involving an expensive context switch. Once the packet has been processed by the daemon, it must be sent out again, reversing the whole process.
All this wouldn't be a big deal if it wasn't for the fact that RTP traffic generally consists of many small packets being transferred at high rates. Since the forwarding overhead is incurred on a per-packet basis, the ratio of useful data processed to overhead drops dramatically.
For these reasons, rtpengine provides a kernel module to offload the bulk of the packet forwarding duties from user space to kernel space. Using this technique, a large percentage of the overhead can be eliminated, CPU usage greatly reduced and the number of concurrent calls possible to be handled increased.
In-kernel packet forwarding is implemented as an nftables module
(or more precisely, an x_tables module). As such, it requires two parts
for proper operation. One part is the actual kernel module called
nft_rtpengine. The second part is a rule in the local nftables chains that
gets hit by UDP packets so that they can be processed by the kernel module.
In short, the prerequisites for in-kernel packet forwarding are:
nft_rtpengine kernel module must be loaded. The module supports
auto-loading when correctly installed.RTPENGINE target. This rule should be limited to UDP packets, but otherwise there
are no restrictions. The rtpengine daemon manages creation and deletion of this rule.rtpengine daemon must be running.The sequence of events for a newly established media stream is then:
rtpengine daemon allocates local UDP ports and sets up preliminary forward rules
based on the info received
from the SIP proxy. Only userspace forwarding is set up, nothing is pushed to the kernel module yet.The kernel module supports multiple forwarding tables (not to be confused with the tables managed by nftables), which are identified through their ID number. By default, up to 64 forwarding tables can be created and used, giving them the ID numbers 0 through 63.
Each forwarding table can be thought of a separate proxy instance. Each running instance of the rtpengine daemon controls one such table, and each table can only be controlled by one running instance of the daemon at any given time. In the most common setup, there will be only a single instance of the daemon running and there will be only a single forwarding table in use, with ID zero.
The kernel module can be loaded with the command modprobe nft_rtpengine. It
isn't normally necessary to do this explicitly or manually as the module is
automatically loaded when the appropriate nftables rule is created (see
below). The module supports a few options (see modinfo -p nft_rtpengine) which
can either be set at the modprobe command line, and/or through an entry in
/etc/modprobe.d/ (by default /etc/modprobe.d/rtpengine.conf) for options to
be set when auto-loading the module. Options can be inspected during runtime
via /sys/module/nft_rtpengine/parameters/.
With the module loaded, a new
directory will appear in /proc/, namely /proc/rtpengine/. After loading, the directory will contain
only two pseudo-files, control and list. The control file is write-only and is used to create and
delete forwarding tables, while the list file is read-only and will produce a list of currently
active forwarding tables. With no tables active, it will produce an empty output.
The control pseudo-file supports two commands, add and del, each followed by the forwarding table
ID number. To manually create a forwarding table with ID 42, the following command can be used:
echo 'add 42' > /proc/rtpengine/control
After this, the list pseudo-file will produce the single line 42 as output. This will also create a
directory called 42 in /proc/rtpengine/, which contains additional pseudo-files to control this
particular forwarding table.
To delete this forwarding table, the command del 42 can be issued like above. This will only work
if no rtpengine daemon is currently running and controlling this table.
Each subdirectory /proc/rtpengine/$ID/ corresponding to each forwarding table contains the pseudo-files
blist, control, list and status. The control file is write-only while the others are read-only.
The control file will be kept open by the rtpengine daemon while it's running to issue updates
to the forwarding rules during runtime. The daemon also reads the blist file on a regular basis, which
produces a list of currently active forwarding rules together with their stats and other details
within that table in a binary format. The same output,
but in human-readable format, can be obtained by reading the list file. Lastly, the status file produces
a short stats output for the forwarding table.
Manual creation of forwarding tables is normally not required as the daemon will do so itself, however deletion of tables may be required after shutdown of the daemon or before a restart to ensure that the daemon can create the table it wants to use.
The kernel module can be unloaded through rmmod nft_rtpengine, however this only works if no forwarding
table currently exists and no nftables rule currently exists.
In order for the kernel module to be able to actually forward packets, an
nftables rule must be set up to send packets into the module. Previously
these rules were managed via the iptables and ip6tables CLI tools (in
combination with a plugin module, e.g. via a command such as iptables -I INPUT -p udp -j RTPENGINE --id 42), but with iptables being phased out and
replaced by nftables this is no longer necessary.
The rtpengine daemon manages creation and deletion of the required nftables
rules directly, creating them during startup and deleting them during
shutdown. The location of the rules and the names of the chains can be
configured via the options nftables-chain and nftables-base-chain (see
manual page).
NOTE: As the nft command-line tool isn't able to understand the non-standard
RTPENGINE target rules, it may emit a warning when displaying the rules via
nft list ruleset or similar.
The rules can also be managed explicitly and directly from the command line via
the \-\-nftables-start and \-\-nftables-stop options (see manual page).
This is useful to manage these rules from a script if so desired, or to
re-create them during runtime if they had gotten lost.
A typical output from nft list ruleset with a rule loaded may look like this
(including the aforementioned warning message):
# Warning: table ip filter is managed by iptables-nft, do not touch!
table ip filter {
chain INPUT {
type filter hook input priority filter; policy accept;
ip protocol udp counter packets 1 bytes 201 jump rtpengine
}
chain rtpengine {
XT target RTPENGINE not found
counter packets 1 bytes 201
}
}
In a typical setup no extra steps need to be taken to facilitate in-kernel forwarding, provided the kernel module is compiled and installed for the currently running kernel (which should be handled by dkms on compatible systems).
A cheat sheet with various related commands is below:
# Load module: this only needs to be done once after system
# (re-) boot, but can be omitted if auto-load is working correctly
modprobe nft_rtpengine
# Load module with some options set
# (usually handled via /etc/modprobe.d/rtpengine.conf)
modprobe nft_rtpengine proc_uid=1234 proc_gid=5678
# Create nftables rules: This normally happens automatically during
# startup. Kernel table ID and nftables chain names are taken from
# the config file
rtpengine --nftables-start
# Create nftables rules overriding the default chain names, or the
# names from the config file
rtpengine --nftables-start --table=42 --nftables-chain=foobar --nftables-base-chain=
# Delete a leftover table. Might be needed if the kernel module
# can't be unloaded (see /proc/rtpengine/list)
echo 'del 0' > /proc/rtpengine/control
# Start daemon: this creates the nftables rules for table 0
/usr/bin/rtpengine --table=0 --interface=10.64.73.31 --interface=2001:db8::4f3:3d \
--listen-ng=127.0.0.1:2223 --tos=184 --pidfile=/run/rtpengine.pid --no-fallback
# Start daemon without implicitly creating any nftables rules (empty chain name)
/usr/bin/rtpengine --table=0 --interface=10.64.73.31 --interface=2001:db8::4f3:3d \
--listen-ng=127.0.0.1:2223 --tos=184 --pidfile=/run/rtpengine.pid --no-fallback \
--nftables-chain=
In some cases it may be desired to run multiple instances of rtpengine on the same machine, for example if the host is multi-homed and has multiple usable network interfaces with different addresses and different port ranges. This is supported by running multiple instances of the daemon using different command-line options (different local addresses and different listening ports) or different config files, together with multiple different kernel forwarding tables.
For example, if one local network interface has address 10.64.73.31 and another has address 192.168.65.73, then the start-up sequence might look like this:
/usr/bin/rtpengine --config-file=none --table=0 --interface=10.64.73.31 \
--listen-ng=127.0.0.1:2223 --tos=184 --pidfile=/run/rtpengine-10.pid --no-fallback
/usr/bin/rtpengine --config-file=none --table=1 --interface=192.168.65.73 \
--listen-ng=127.0.0.1:2224 --tos=184 --pidfile=/run/rtpengine-192.pid --no-fallback
With this setup, the SIP proxy can choose which instance of rtpengine to talk to and thus which local interface to use by sending its control messages to either port 2223 or port 2224.
For ease of use it's recommended to use multiple config files
(\-\-config-file=...) or config file sections (\-\-config-section=...)
instead. When running under systemd a new service unit can be created via
systemctl edit ....