Latency : What is it? Why it matters…
Latency…Huh?
When we’re talking tech, latency is the time that elapses between a user request and the completion of that request. Even processes that seem instantaneous have some measurable delay. The goal is to have the shortest latency possible!
Why does it matter?
Nearly all applications and functions that we rely on today…especially today, require very low latency in order to perform as intended. Think quicker and improved user experience and customer satisfaction. Things like cloud hosting, virtual meeting solutions like zoom and teams, or mission-critical computation applications.
Why does it happen?
When a trigger happens like a user, app, or solution requests information from another system, the info request gets processed locally, then sent over the network to a server or system. There, it is processed again, and a response is formed, starting the reply transmission process for the return trip.
Along the way, and in each direction, are network components known as switches, routers, protocol changes, translators, and changes between the network cabling, fiber, and wireless transmission. At each step, tiny delays are introduced, which can add up to discernible wait times for the user.
Think of driving on a highway, what’s it like during rush hour, high travel times? As all network traffic continues to grow, inactivity is increased for all users as the line to complete transmissions backs up and micro-latencies add up. This creates high latency, a frustrating delay before the webpage loading begins. You wont likely notice a 5 second delay on the internet, but you will notice it on a voice call. This is why it’s incredibly important to set up a voice network which is separate from the network that all of your other IT solutions run off of.
The geographical distance that data must travel can also have a significant effect. This is why edge computing, the practice of locating data and applications closer to users, is a well-known strategy for reducing latency. In some cases (see below), reducing this distance is a smart, effective way to lower network latency.
A latency network? WHAT?
An LL network is one that has been designed and optimized to reduce lag as much as possible. However, a low latency network can’t improve latency caused by factors outside the network.
What is jitter?
Jitters when it deviates unpredictably from an average; in other words, when it is low at one moment, high at the next. For some applications, this unpredictability is more problematic than high latency.
What is ultra-low latency?
A ULL is measured in nanoseconds, while low latency is measured in milliseconds. Therefore, ultra-low latency delivers a response much faster, with fewer delays than low latency.
How is low latency achieved?
For new deployments, lag is improved through the use of a next-generation programmable network platform built on software-defined hardware, programmable network switches, smart network interface cards, and FPGA-based software applications.
To reduce latency in an existing network, follow the steps below:
- Identify the network problems and impacts
- Confirm that the problems and impacts are caused by high latency
- Identify the IT infrastructure that is contributing to the high-latency problem
- Evaluate which network switches and network interface cards can be replaced to create a low latency environment
- Evaluate which network functions can be offloaded to a field-programmable gate array (FPGA)-programmable switch or smart network interface cards (SmartNICs) to reduce latency to milliseconds or nanosecondsl