The State of Edge Compute and Delivery 2024
From aggregation to smaller, regional data centers, edge continues to influence streaming approaches. Looking back over articles from the past few Streaming Media Sourcebooks, as well as about a dozen surveys we’ve done in conjunction with StreamingMedia.com on a variety of streaming topics, the term “edge compute” surfaces frequently in conversations and responses.
In the 2022 Sourcebook, I mentioned that edge compute was the buzzword topic. While it’s been outpaced somewhat by AI in the past year—which is neither artificial nor intelligent, despite the rising sophistication of machine learning algorithms—interest in edge compute remains high.
In this year’s State of Edge Compute and Delivery, I’m going to explore the three edges that seem to matter most for streaming: aggregation, on-prem, and regional.
Aggregation
To illustrate the edge computing concept of aggregation, consider the so-called Internet of Things label that we’re bombarded with in our private home lives. The term’s a bit of a misnomer, as most of these “things” require a connection to a central server solution that’s proprietary between a given thing and the manufacturer of that particular type of thing. Think of refrigerators talking to their manufacturer’s platform, while the connected robot vacuum on the floor next to the fridge communicates with its manufacturer’s completely different platform. Both use the same transit path out of the home (typically Wi-Fi connections via a fiber or cable modem), but the things can’t communicate with one another unless all of them are manufactured and serviced by the same company, in what most refer to as “ecosystems” these days instead of the “walled garden” terminology of the past.
The lack of edge communication is often the same with your mobile device ecosystem. Apple products connect easily with other Apple products, thanks to the tie-in of iCloud aggregation between devices, but they often can’t talk directly to one another while sitting in the same room, thereby requiring costly and slow uploading and downloading of content between the two devices. The same is true for Android products, but exacerbated even more since the ecosystem limitations are narrowed down to manufacturer-specific devices.
Edge computing offers a potential way to cluster access to those discrete services, in the form of distinct microservices at the edge for each ecosystem, while at the same time lowering overall latency—by maintaining discrete authentications for each “thing” to a physically closer set of servers—as well as lowering the risk of a centralized data breach impacting the entire customer set.
One subset of this aggregation, which is playing out in the European Union with regard to app stores, is the increasing number of biometric authentications being required to access disparate devices. Given the fact that a number of countries have unique privacy and content-sharing stipulations, both for biometric and other user-specific data, and that these stipulations apply to any content that sits within the country’s geographic jurisdictional boundaries, regardless of where an end user resides, there’s an even more compelling reason to consider edge compute for authentication purposes.
The secure access service edge (SASE) premise has a direct impact on streaming authentication, potentially eliminating the authentication Achilles’ heel that’s plagued high-profile pay-per-view live streams, even when the stream itself worked fine. Now this approach appears to be gaining traction.
While a 2021 projection by Gartner showed that the SASE market would grow from 2021 to 2025 at a compound annual growth rate (CAGR) of 36% to almost $15 billion by 2025, this was a scaled-back projection from a year earlier, when Gartner had predicted a CAGR of 42%, reaching almost $11 billion by 2024. Other recent predictions have also lowered expectations, with one recent report projecting SASE would see a CAGR of about 28% from 2022 through 2027, but end with a value of only $5.3 billion.
Regardless of the growth rate, two benefits come into play when SASE solutions are deployed. First, there’s the breadth of potential applications. “Edge computing continues to play a key role in an increasing number of biometric applications involving various modalities,” according to Chris Burt, managing editor of Biometric Update.
Second, edge computing has the power to lower overall power consumption in certain parts of the streaming media workflow. “Advanced processors running on low power and algorithms with small processing footprints have unlocked a range of possible use cases and features for biometrics that were previously impossible,” says Burt.
Content Privacy Benefits
As the promise of edge computing grows, authentication engines will potentially allow more constrained storage of sensitive content at the edge rather than at a massive, generic cloud computing data center. Security remains a concern for cloud-based delivery of sensitive or restricted content. In addition, as mentioned earlier, a number of countries have unique privacy and content-sharing stipulations that apply to any content that sits within their geographic jurisdictional boundaries.
That’s where edge authentication might just shine, and I expect we’ll see a ramping up of per-asset authentication solutions to address both enterprise (with its still historically high percentage of employees working from outside the traditional enterprise office space) and social user-generated media content.
In addition, edge-based aggregation can take on additional forms, such as more finely tuned recommendation engines in a specific geographic region. Analysts are constantly trying to pinpoint consumer behavior trends in smaller markets, as tastes vary widely even across little segments of a larger city (think Chinatown versus Haight-Ashbury versus SoMa in San Francisco).
Deriving too small a sample size, however, raises privacy concerns. Therefore, a model going forward would be to generate recommendations based on like-kind pockets across a variety of cities (think Chinatown in San Francisco as well as Flushing, New York), where the anonymized consumer behavior from each edge location is compared to edge data from other locations as a way to further optimize recommendations—and caching decisions, as a cost-savings measure—based on similar communities rather than delivering all the content from a central location.
On-Prem
The trend toward more granular data also has a corollary for streaming: The closer the server, the lower the latency. This has given rise to the concept of micro data centers for educational and enterprise use cases.
“Micro data center deployments are expected to be commonly deployed to power smart buildings or campuses for institutions or enterprises,” says Burt. “The actual server racks could be found contained inside a closet, the room that has traditionally held on-premise servers, or even somewhere in plain sight, as they are neither noisy nor visually obvious.”
One of the research charities I’m involved with, Help Me Reconnect, has been exploring the concept of using lateral filing cabinets—the typical tan or white two-to-five-drawer filing cabinet that’s been a staple of legal or medical offices for decades—as a way to embed micro data centers into existing office environments. To Burt’s point, these cabinets just blend into the background and provide a way to increase edge compute capabilities without requiring additional floor space. At Help Me Reconnect, we’ve merged donated lateral cabinets and donated desktop mainboards to create an environmentally friendly reuse of old technology to meet applications as varied as electronic medical records and educational campus media servers at almost inaudible noise levels.
From an enterprise standpoint, edge compute will enhance, rather than replace, enterprise content delivery networks (eCDNs) due to advancements in caching optimization. In fact, caching via an eCDN will become more important in terms of bandwidth reductions within corporate facilities, as more and more of this on-demand content is stored outside the traditional corporate firewall. As edge caching is further optimized in an on-prem model, additional resources could be focused on ramping up the use of multicast networks within small segments of the corporate firewall via software-defined networking (SDN), virtual local area networks (VLANs), or other similar approaches.
Lowering Latency to Near-Zero Thresholds
Along those same lines, the rise of videoconferencing means that there will be greater adoption of WebRTC and low-latency video delivery within the corporate network. To best accomplish scalability for low-latency, bidirectional video calls that traverse the corporate firewall—in which an equal number of participants in a videoconference could be in a single corporate location, communicating with colleagues who are working from home—enterprise video teams may end up rebroadcasting the WebRTC-based session to a number of internal corporate participants as a viewer-only stream.
Since most of the live content for a videoconferencing call occurs outside of the corporate network, there’s less reliance on the virtual private network (VPN) tunneling that most workers are familiar with when remotely logging on to the corporate network for email, internal communications presentations, or accessing a shared folder. However, for on-demand content, an eCDN still makes sense. After all, if there’s a training video to watch, and the user is logged in remotely over the VPN, the utilization of video caching means that VPN users retrieve the requested video from the cache, over VPN, rather than having to stream the content from the internet through the VPN. Essentially, caching halves the data requirements for streaming content over the VPN.
Small, Regional Data Centers
Pulling back slightly from on-prem edge scenarios, is there common ground between fully local and fully centralized content aggregation and delivery? To answer this question, first let’s look at the centralized model. The edge compute that most services often pitch sits, without irony, in a large, central data center. That’s typically true for one of three reasons:
- A cloud service provider’s ability to spin up more resources for its overall customer base, as well as load-balancing for peak traffic models
- A desire for cloud-based services to aggregate user data at a central location (I’ve covered the downfalls of this earlier in the article.)
- A lack of smaller, regional data centers or at least the lack of space in existing, smaller data centers
While the first two are convenience plays on the part of the cloud service provider or its customers, the last one is a tougher challenge to solve. In many ways, while the streaming industry acknowledges that latencies could be lowered by moving toward edge computing—along with bringing benefits like aggregation and authentication—few of the established players in the market have compelling reasons to move from a centralized data center approach to a more regional approach.
Fortunately, two trends are at work to change that. The first, as pointed out by State of Streaming survey respondents, is the continuing need to drive down latencies, especially as live-streaming trends displace classic streaming on-demand delivery bias.
Physical distance is the primary factor in latency, so it almost goes without saying that servers located closer to the end user inherently lower the transit time. The lower the transit distance, the lower the likelihood of network congestion, which further enhances the low-latency edge compute premise that’s directly correlative to smaller, regional data centers.
The second trend is synchronization. For all of the talk of latency in live streaming, a primary shortfall of live-streaming events on best-effort networks (i.e., the classic internet) is a lack of synchronization of content between two devices in the same physical location. A number of companies have developed concepts to address latency, but a much smaller group of solutions addresses synchronization.
Interestingly, it’s likely that synchronization is most critical within a small geographic footprint. While there are scenarios in which your friend in another city is watching the same game as you, receiving it at a latency a few seconds faster than you do and communicating the same to you via text or a messaging app, the majority of complaints we’ve heard across more than a dozen surveys in the past 2 years center on synchronization within hearing distance. It’s as basic as hearing your neighbor yell in delight when their team scores, while you’re waiting for a few seconds to see the same score on your home device. This is a scenario that’s easily solved via edge compute streaming synchronization.
On the work front, there’s an even more compelling reason for edge synchronization: bring your own device (BYOD) meets multiscreen streaming needs. The BYOD approach of a decade ago is far different from today’s mobile computing scenario, as the major mobile operating systems now fully embrace sandboxing of each individual application. In the corporate world, this is critical, as the ability to sandbox a corporate application means that IT can rest easier when employees want to access sensitive corporate data from their own devices.
But also consider synchronization in the mix. If you’re involved in any work-related corporate meetings, online seminars/trade shows, or even just a videoconferencing call that you’ve joined from home, there’s often a limited amount of screen real estate on which to do the three things any good multitasker would do during the workday: join the meeting, follow the meeting, and prepare for your next meeting. In the old days of corporate teleconferences, you could multitask fairly easily, but these days with videoconferencing, you either need to fully engage your screen to be a part of the meeting, turn off your camera and minimize the meeting to be able to get work done, or use two devices to be part of the meeting and also get your own work done.
The third approach has led to a rise in meeting participants wanting to use a second device (such as a smartphone or tablet) to join the meeting, while leaving their primary work computer free to be used for actual work. And that has led to a need to absolutely, positively make sure the content is synchronized between the two devices.
A final note on smaller, regional data centers. Starting in 2022, we’ve asked respondents to the biannual State of Streaming survey to tell us about their future plans. More than half of respondents indicate that their plans include rolling out future business in smaller-footprint data centers. That percentage has grown with each subsequent State of Streaming survey, and this is where I see the practical uptake in edge computing occurring over the next 5 years.
Conclusion
There’s one other area to consider when it comes to edge compute, and that’s the continued use of private clouds. Private clouds, which often offer more scalability compared to on-prem infrastructure, due to the fact that they can be scaled up or down as needed without requiring additional hardware, seem to offer a potential sweet spot in which edge computing can thrive. This cloud-first approach could just as easily be instantiated in localized markets using a combination of smaller, regional data centers, with on-prem equipment as the fallback mode.
Alternatively, the private cloud could be used in an on-prem-first approach, reserving the cloud portion for scaling up when demand requires additional throughput and storage at a lower capital expense.
Related Articles
Newest version of twice-annual report expands on "pain points" and growth patterns, available for download now.
28 Nov 2023
Rising fundamental costs morph into rising platform costs for content owners and consumers
26 Oct 2023
CDN77's Juraj Kacaba sits down with Tim Siglin to discuss low-latency streaming and the edge in this interview from Streaming Media East 2023.
16 Jun 2023
New 2023 Spring edition of State of Streaming survey launches, shows shifting focus
25 May 2023
From road warriors to remote-only companies, enterprise video is shifting. It is, to put it mildly, in a state of flux. At the same time, though, so is the state of business. Given those two facts, the question I'll attempt to answer in this year's State of Enterprise Video is whether these coexistent movements will occur in lockstep or as polar opposites.
12 Apr 2023
With so much of today's internet traffic flowing through large data centers—some estimates say that 20% of all of the world's data traffic is concentrated in a few counties in northern Virginia, with 80% of that being IP video traffic—how are we going to move from the centralized delivery model of data centers that are a million or more square feet in a single-floor cavernous facility to the edge, where processing can occur closer to the homes of even the remotest rural areas of the U.S.?
14 Oct 2022