Optimizing Ethernet For Speed, Power, Reach, and Latency

Anritsu discusses Ethernet usage trends in data center networks. They also explore the technologies helping operators meet growing bandwidth demands and verify network speed, power, latency, and performance.
Aug. 10, 2022
2 min read

In a new white paper, Anritsu discusses Ethernet usage trends in data center networks. They also explore the technologies helping operators to meet growing bandwidth demands and verify network speed, power, latency, and performance.

Get the full report

“Growing demand for information has created an explosion in data center traffic,” according to a new white paper from Anritsu. They say this demand is increasing the need for data center architectures to support ever higher Ethernet transfer rates. As operators seek to “optimize Ethernet media types for speed, power, reach, and latency,” they’re being forced to reevaluate some long-held assumptions in these areas, according to the paper.

The authors explain that the need to reduce latency is increasingly important as data centers transform into edge computing networks. They say, “as computing resources move closer to the edge, the latency key performance indicator (KPI) tightens. This KPI is application-service dependent. Latency affects the user experience for applications and must be considered when deploying Ethernet connects.”

As data center network operators move to 400 Gigabit Ethernet and beyond, they will face new challenges such as signal integrity, network interoperability, and maintaining service level agreements (SLAs) for different applications. – Anritsu, “Ethernet in Data Center Networks

To address concerns around power and speed, operators are turning to optical transceivers and high speed breakout cables but, according to the paper, these technologies are not without their challenges. The authors note that “not all 400G Ethernet optics are created equal and their performance on forward error correction (FEC) KPI thresholds varies.” Likewise, high speed breakout cables are less expensive, but have performance and distance issues.

The paper goes on to explain how networking equipment manufacturers are turning to testing solutions to measure the signal integrity of new high speed optical interfaces.

Anritsu also explores how “with multi-access edge computing and network virtualization, data center providers can maintain different SLAs for different applications.”

Download the full report for more information on technologies that can verify network performance at high speeds.

About the Author

Kathy Hitchens

Kathy Hitchens

Kathy Hitchens has been writing professionally for more than 30 years. She focuses on the renewable energy, electric vehicle, utility, data center, and financial services sectors. Kathy has a BFA from the University of Arizona and a MBA from the University of Denver.  

Sign up for our eNewsletters
Get the latest news and updates
Tirachard Kumtanom/Shutterstock.com
Source: Tirachard Kumtanom/Shutterstock.com
Sponsored
Philip Tappe of Modius explores how modern DCIM solutions can optimize data center efficiency and PUE with real-time data, AI-driven analytics, and seamless integration. He shares...
Image courtesy of Schneider Electric
Source: Image courtesy of Schneider Electric
Sponsored
Alexander Rakow, Director of Sustainability, Secure Power and Data Centers for Schneider Electric, explains the importance of collaboration between solutions providers, data center...