|LCM Reference Manual|
UDP Multicast Setup
UDP Multicast Setup — Getting maximum performance on your LAN or local host
A quirk of multicast, at least on Linux, is that the loopback interface 127.0.0.1 cannot be used for multicast traffic. This means that to use LCM, even for inter-application communication on a single host, you must have a network interface. If this network interface is already connected to the Internet, LCM will generally "just work" because the default route will allow LCM to find the correct network interface for multicast traffic.
If the network interface is not connected to anything, you must explicitly enable it for multicast traffic. First, bring the interface up using a command like ifconfig eth0 up. It need not be assigned an IP address. Then, you must add a route for multicast traffic. For LCM's default multicast group, the command would be:
route add -net 184.108.40.206/16 dev eth0
Remember, you must always do this to use LCM locally if your machine is not connected to any external network.
LCM by default uses a time-to-live (TTL) value of 0. This will prevent any LCM packets from being transmitted on the wire. Only local applications will see them. Choose a value of 1 for the entire subnet to see the traffic. Even larger values will enable the packets to pass through routers. However, these routers must be set up with multicast routing tables in order to successfully relay multicast traffic.
There are two ways to set the TTL. First, LCM constructors in all
languages provide a way to set the TTL (e.g. see
lcm_create), so you can set this in your
code. Second, if your program does not explicitly set its provider URL,
then you can use the
variable. For exaple, in the bash shell:
When the multicast protocol was designed, it was intended that on local subnets it would act like broadcast traffic. However, for high-bandwidth applications, this may not be desirable since it will degrade the bandwidth available for hosts not participating in the multicast group.
Switch manufacturers have come up with a solution to this problem called "IGMP snooping". IGMP snooping consists of two parts. First, the switch must monitor network traffic across its ports and decode any IGMP packets that deal with multicast group subscriptions. The switch can then send multicast traffic for a particular group only to those ports where a subscription request has been observed. The second part is IGMP querying. Hosts will only send out subscription requests upon first subscribing or when queried by a router. In order for the switch to maintain its state, these queries must be sent every few minutes. If there is no router on your subnet sending these queries, you must configure the switch to send out "fake queries". Without this, the hosts will eventually stop sending IGMP subscription requests, and the switch will "fail open" causing all multicast traffic to be broadcast again.
If you have a subnet with devices of different speeds, such as 10Mbps links mixed with 100Mbps links, be extra careful using multicast. LCM traffic will be sent to all devices (unless you enable IGMP snooping, above). If two 100Mbps devices are exchanging more than 10Mbps of traffic, this traffic will be relayed to the slow device by the switch, even if that device does not run LCM. This can overwhelm the bandwidth of that link, and depending on the switch, may even cause back pressure on the 100Mbps devices. That means, not only will the 10Mbps link be swamped, but the 100Mbps devices will be unable to transmit faster than 10Mbps because their send queues will start filling up and cause blocking writes in the LCM API.
The easiest solution to this problem is to enable IGMP snooping, as described above. If that is not available, sometimes it is sufficient to upgrade from consumer-grade switches to managed switches, which often have better buffering strategies. As a last resort, separating the slow devices from high-speed devices using two levels of switches can also improve the situation.
Much of this information is taken from http://www.29west.com/docs/THPM/udp-buffer-sizing.html
When used on a properly shielded local area network, the most common source of dropped and lost packets is not electrical disturbances. Instead, it will usually be packet buffers that have reached capacity and cannot store any more packets.
Operating system kernels typically allocate both a
send buffer and receive buffer
for a network socket. The send buffer stores packets that the operating
system has agreed to transmit (i.e. a call to
was successful) but that it hasn't actually yet transmitted over
the network. The receive buffer stores packets that the operating system
has received from the network, but that the application hasn't yet
retrieved. In both cases, the buffers have a maximum capacity, and no new
packets can be sent or received if that capacity is reached. In the
send case, a call to
send() blocks until there is
space in the buffer (or fails with
non-blocking sockets). In the receive case, incoming packets are simply
When LCM is used in a high-bandwidth application, it may become necessary to increase both the default and the maximum kernel receive buffer size to avoid dropped packets.
Configuring the kernel can be done without rebooting, but requires superuser privileges. Table 1, “Setting Kernel UDP Receive Buffer Parameters” demonstrates how to do this and set a 2MB maximum buffer size.
Table 1. Setting Kernel UDP Receive Buffer Parameters
FreeBSD, Darwin (OS X)
Most GNU/Linux distributions also allow setting these permanently
using the file
/etc/sysctl.conf. To do this, add
the following lines into the file somewhere:
An LCM application can also explicitly request a larger receive
buffer, which is useful on Solaris and FreeBSD/OS X. This is done by
recv_buf_size option with the
udpm provider in
lcm_create. In our initial
projects, which involved typical data rates of 10 MB/s, a 2 MB receive
buffer was generally sufficient.