Study Plan
2110 Topo
This lesson explains the Ethernet and the Data Link layer and how it fits into the SMPTE ST 2110 stack.
• 2110 Key Changes and Emphases in the OSI Stack
• Ethernet Frames
• Jumbo Frames
• Ethernet Mechanics
• Frame Reception Process
• Bridges and Hubs
• Switches
• Spanning Tree Protocol
There used to be many schemes or protocols for passing data at the Data Link layer (OSI Layer 2). Today, one has won out: Ethernet. Xerox created Ethernet in 1973 at the Palo Alto Research Center (PARC). It was meant to connect copier subsystems in big machines. Its potential to connect computers gained early support from DEC and Intel, in addition to Xerox PARC. The three companies worked together to create the DIX standard. This stands for Digital-Intel-Xerox. It was published in September 1980 as the "Ethernet Blue Book" (Version 1.0). It used coaxial cable and operated at 10 Mbps. With some modifications, DIX evolved into IEEE 802.3 (finalized in 1983–1985). Some purists claim that "Ethernet" refers only to the original Xerox/DIX specification. They see IEEE 802.3 as the current standard. Still, most people use these terms interchangeably now. Novell's early adoption of Ethernet in its networking products significantly boosted its popularity.
The Data Link layer corresponds to point-to-point connections, as was traditionally found in baseband video and audio systems. Today, these point-to-point connections are typically between network switching devices. This is different from the myriad of boxes used before 2110.
How is the Data Link layer
differentiated today from that layer found in traditional OSI stacks? It supports higher bandwidth options of 10, 25, or 100G. It also needs switches that have low jitter, are non-blocking, support multicast and that can also ensure high Quality of Service (QoS). The switches must also support redundancy. When it comes to creating a 2110 compatible data link layer; "it's all about the switches."
Layer 2 (the Data Link layer) is the lowest OSI
layer where data is formally framed, often described as being placed inside a “wrapper.” Conceptually, this wrapper can be thought of as an envelope around the payload. When a frame arrives at a network interface, the Ethernet (Layer 2) header is processed first. The information in this header—such as source and destination MAC addresses, VLAN tags, and frame type—tells the receiving hardware where the frame came from, who it is intended for, and how the payload should be handled.
Not to get too far ahead of ourselves, but to start off with a good understanding of what comprises the Ethernet packet payload, let's look ahead a bit at what the Ethernet layer is transporting in a typical 2110 scanerio.
Ethernet frames have
a maximum payload (MTU) of 1500 bytes for standard configurations, but the full frame size is up to 1518 bytes when including the header and trailer (excluding the 7-byte preamble, 1-byte Start Frame Delimiter, and 12-byte interframe gap, which are handled at the physical layer and not counted in the MTU). The Ethernet header is always 14 bytes (6-byte destination MAC + 6-byte source MAC + 2-byte EtherType/Length), and the trailer is always 4 bytes (Frame Check Sequence/FCS), totaling 18 bytes of fixed overhead per frame. This structure doesn't change—it's standard per IEEE 802.3.
For carrying HD video (e.g., in SMPTE ST 2110 scenarios, which uses RTP over UDP/IP/Ethernet for uncompressed or lightly compressed streams like 1080p60), RTP is typically paired with UDP (not TCP, due to UDP's low-latency, connectionless nature). Here's a breakdown of a typical packet:
In real ST 2110 HD video deployments:
| Protocol/Layer | Bytes | Purpose in HD Video Scenario |
|---|---|---|
| Ethernet Header + Trailer | 18 (14 header + 4 FCS) | Frames the entire packet for LAN transmission; MAC addressing and error checking. Fixed for all Ethernet frames. |
| IP (IPv4) | 20 | Routing and multicast addressing (e.g., for distributing video streams across switches). |
| UDP | 8 | Lightweight transport; multiplexes RTP streams via ports, with optional checksum for integrity. |
| RTP | 12 | Real-time handling: timestamps for sync (critical for HD video timing with PTP), sequence numbers for reordering/loss detection, and payload type (e.g., video format). |
| Video Payload | Up to 1460 | Actual HD video data (e.g., uncompressed YCbCr samples from ST 2110-20); the "useful" part after all wrappers. |
| Protocol | OSI Layer | TCP/IP Model Layer | Purpose |
|---|---|---|---|
| Ethernet | 2 | Link | Framing, MAC addressing |
| IP | 3 | Internet | Routing, addressing |
| UDP | 4 | Transport | Connectionless transport |
| RTP | 7 | Application | Real-time media timing & sequencing |
| RTCP | 7 | Application | Control & monitoring for RTP |
A major difference
between a typical business website and a professional SMPTE ST 2110 media facility is optional use of "jumbo" frames. But these frames are not required by the ST 2110 standards. We will see that there are pros and cons to their use.
Why are jumbo frames used?
Jumbo frames are essential for high-bandwidth applications like SMPTE ST 2110, where HD/4K/8K uncompressed streams would otherwise fragment excessively or overwhelm the network stack. In practice, almost every professional 2110 deployment uses jumbo frames (commonly 9000 bytes) network-wide.
Why 1500-byte MTU would be preferred:
How Jumbo Frames Affect the Data-Link Layer (Layer 2)
Bottom line: Jumbo frames are a Layer 2 configuration change that dramatically improves efficiency for high-throughput, low-latency media like 2110, but they require consistent end-to-end support across the entire switched network. Misconfiguration is one of the most common reasons 2110 streams fail silently or fragment.
Let's now look
at what Ethernet is and how it works. It is a four-step process. Ethernet uses a process called CSMA/CD (Carrier Sense, Multiple Access/Collision Detection).
Broken down this means:
CSMA/CD is simple but collision-prone, so real-world Ethernet networks rarely achieve full advertised speeds (especially in shared/half-duplex environments). Modern switched full-duplex Ethernet eliminates collisions.
A Collision Domain
(CD) is a network segment where all nodes receive traffic from any other node in that segment (e.g., everything on a hub or shared coax). In early large collision domains (before widespread switches), the 5-4-3 Rule limited network size to prevent undetected collisions due to propagation delay:
No more than 5 segments connected by 4 repeaters. Only 3 segments can be populated (have nodes attached); the other 2 are unpopulated link segments between repeaters.
This rule applied mainly to 10 Mbps Ethernet (10BASE5 thicknet and 10BASE2 thinnet) in tree/bus topologies. It is now obsolete in switched networks, where each switch port creates a separate collision domain.
A NIC sends Ethernet frames (based on IEEE 802.3) consisting of 26 bytes of overhead, surround the payload data consisting of the following:

The frame "envelops" upper-layer data (e.g., IP packets) without knowing the content.

Frame Reception Process:
A MAC address
consists of six bytes (48 bits).
A Data-Link broadcast Layer 2 broadcast is indicated when all bits in the destination address are high, or FF:FF:FF:FF:FF:FF. This means that this packet is intended for all nodes.
What happens when a node receives an Ethernet frame?
Often a network consists ofmultiple local LANs. The simplest way to do this historically was with a Bridge. They were first used to break up collision domains (CDs). Remember that a CD usually spans all nodes connected to a hub, or in early networks, all nodes connected together on the same length of tapped coax with AUIs.
Bridges and Hubs
Historically,
multiple local LANs were connected via bridges (early devices to separate collision domains). Bridges learn MAC addresses, filter traffic, forward broadcasts, and use Spanning Tree Protocol (STP) to prevent loops (bridging loops cause broadcast storms). Bridges do not modify frames.
Hubs repeat every frame to all ports (one large CD). Switches can replace hubs directly, providing dedicated bandwidth (only source/destination in same CD).
If a bridge is designed to connect different protocols, such as Ethernet to Token Ring, those are referred to as Translational bridges. You won’t find many of these anymore. Bridges always forward broadcast traffic, which is traffic intended for all nodes. Network devices use broadcast addresses to discover information about a network, such as when looking for a resource on the network.
Bridges can’t handle multiple routes between two different nodes. They get confused as to which collision domains the two nodes are actually in. This is known as a bridging loop. Most bridges use a technique called the Spanning Tree Algorithm (more on this in a bit) to detect bridging loops and automatically disable bridge paths that form loops.
Switches are multiport hardware-accelerated bridges (ASIC-based). They break collision domains (each port is its own CD), learn MAC addresses, filter traffic, and prevent loops via STP. Switches are faster than routers (Layer 2 vs. Layer 3). Modern networks are flat (single broadcast domain via switches), with collapsed backbones replaced by distributed switching.

Spanning Tree Protocol (STP)
As we have mentioned,
you can confuse bridges and switches if there are multiple paths between network segments. To combat this situation upon wakeup, bridges and switches invoke Spanning Tree Protocol (STP) where each port on either device either allows traffic forwarding to other segments, or the port is set to block traffic between two segments.
To put this in a simpler light, the switches get together and elect a 'Root Bridge' — basically the boss switch. The one with the lowest Bridge ID wins (that's usually the default priority of 32768 plus its MAC address — lower MAC breaks ties).
Then, every other switch figures out: 'What's my best, cleanest, fastest path to reach this Root Bridge?' That best path gets to be the Root Port on each non-root switch.Next, on every network segment (link between switches), they pick one switch to be in charge of that segment — the one with the best path to the Root gets to be the Designated Port for that link. Now, here's the important part: any ports that would create a loop? STP just says 'nope' and blocks them. They sit there quietly, ready to jump in if needed, but not forwarding traffic right now.
Each port settles into either a forwarding or a blocking state, this process is known as convergence. When all ports on bridges and switches have transitioned to either forwarding or blocking modes, the LAN is said to be converged. The goal is to leave only one path between network segments. During the convergence phase, no data will be forwarded. Convergence determines that all ports are either in the forwarding or blocking mode. STP is a data link protocol. From here on out we will concentrate on switches.
When convergence is complete, every port on every switch will be left in one of three states. It will be a root port, a designated port, or a blocked port. A root port is the port looking at a network segment that provides the shortest path back to the designated root switch. A designated port is a port “designated” to communicate with a network segment. If a switch is surrounded by four network segments, then hree ports will be “designated ports” (DP), and one will be the “root port” (RP), connected to the network segment leading back to the root switch. A “blocked port” (BP) will not communicate with the network segment it is connected to, as another port on a different switch is handling that task. This is how multiple paths between nodes are prevented. Multiple paths between nodes can lead to “broadcast storms” where multiple switch ports are all trying to push the same data to a single node. This can quickly avalanche as responses from the destination are multiplied on the return path.


If a port fails on a switch, all the switches will re-enter the convergence mode and a port that was blocked from talking to a segment will now be the “designated” port for that segment. Often, this causes a ripple effect where a number of ports will change between the three states.

To prepare a group of network segments to be tied together by switches, the IT administrator designates Bridge IDs (BIDs). Switches still use the BID term for each switch. Whatever switch they want to designate as the root, where all other switches point to this switch, is assigned the lowest BID value. This BID value is combined with the switch port’s MAC. The BID default value is 32768, and can only be changed in increments of 4096. The switch assigned the lowest BID will be the root switch when convergence is completed. BPDU packets contain BID fields.
If a switch receives a data packet and it can't determine the output port to use, it broadcasts the frame out of all non-blocked ports except the port on which the frame was received. Once a switch determines which port to use for a particular destination, it puts that information into a table. The table is expanded as new destinations are encountered. This is called transparent bridging. When convergence occurs, those tables are erased and the process of discovery starts anew. Temporarily slow network traffic might be a symptom that a network failure or change has forced your network to re-converge.
If a node on a switch cannot communicate with another one connected to the same switch, the problem is most likely that either STP shut down one of the two ports, or that VLAN membership for the port is configured incorrectly. We will look at VLANS shortly.
So switches are either forwarding, flooding out all ports, or dropping received frames. No notification to the sender is done regardless of the switch’s decision. As we will see later, lost data packets are handled at a higher layer. So to recap; the switch has three functions—Address learning, filtering, and loop avoidance. You can set switch ports to authenticate authorized MAC addresses by keeping a list generated by the administrator that is not eliminated during a re-convergence.
It should be stressed that while switches recognize frames and pay attention to source and destination MAC addresses, as well as to the ports that Ethernet frames are received and sent on, hubs do not. Hubs dutifully relay every frame received to every other Hub port. A network switch can usually simply replace a hub with no other changes. The most cost-effective way to start implementing switches is to keep the hubs, but add a central switch that all the hubs connect to. The other advantage is that you can connect central servers to the switch so that they only receive traffic intended for them. Each node connected to its own dedicated switch port means that only two nodes are in the same collision domain. This means more bandwidth and thus throughput.
While we stated earlier that a switch connects one port to only one other port at a time, there is an exception to this. Port mirroring is used on a switch to send a copy of packets seen on one port to another port. This is also called Switch Port Analyzer (SPAN) or Port Spanning. It can cause a lot of overhead on the switch.

Ports also end up as access, trunk, and hybrid on switches. These concepts are key in VLAN networking, but they vary a bit depending on the vendor. While we will look at that next lets talk briefly about them now.
Access ports belongs to exactly one VLAN. Perfect for connecting computers, printers, and servers. While a Trunk port carries traffic for multiple VLANs over one link (classic inter-switch connection). A Hybrid Port is currently a single vender mode from Huawei. You can configure multiple VLANs to exit. Great for mixed scenarios, or weird legacy setups.
To add some confusion to switches, while they are generally found at the layer 2/Data-Link level, there are devices also called switches found at higher OSI layers. Some switches have ambitions of working at the next layer up, level 3/Network, and provide some rudimentary router capabilities. There are also switches that operate on layers 4–7. These provide content switching that allows load balancing. This is useful for HTTP, HTTPs, VPN traffic, especially when converging on servers and other resources.
Now we'll look at VLANs