AI adoption is exploding – and the fastest way to keep pace is to push compute closer to the data at the edge. This trend is driving massive investment in edge data centers, with Research and Markets estimating that the CAGR will increase about 17%, and the global market will increase from about $15 billion in 2024 to nearly $40 billion in 2030.
IT spent the last two decades consolidating compute in the cloud. The next decade will be spent bringing compute back to the data.
Edge data centers are, by definition, located closer to customers, partners, and devices that access an organization’s services. They can range from a few servers in a retail closet to racks of IT equipment in a colocation center (“colo”), and everything in between. In contrast to IoT devices, which make use of microcontrollers and real-time software, edge data centers offer full-stack application processing capabilities, only in a smaller footprint.
The reasons for the transition to edge processing include transaction responsiveness, service scalability, and data sovereignty. Let’s explore these in a bit more detail.
For many organizations, primary data centers are running into constraints. Those include power, cooling, space, and system architectural restrictions. Edge data processing offers an end run around most of those impediments by distributing computation from main data centers and closer to IT service consumers.
Here are some reasons why edge processing makes sense for today’s IT organizations:
These are just some examples on why it may be beneficial to move data processing to the edge. Depending on your organization’s needs, you may find other benefits.
As mentioned above, anything that can relieve system constraints, reduce response times, and eliminate unneeded data transmission and processing is worth considering doing at the edge. One good example is AI inferencing.
Previously we discussed centralized data center’s power and cooling constraints. Generally speaking, for most organizations, AI training and inferencing infrastructure is the number one cause for approaching such limits. While organizations need large, centralized clusters of GPU-accelerated servers to do most AI model training, the infrastructure hurdles associated with deploying AI inference solutions can be much lower. By moving inferencing out to the edge, organizations can reduce their core data center power and cooling requirements to something much more sustainable. Moreover, scaling AI infrastructure to handle inferencing processing loads is much easier to accomplish if it can be done at multiple edge data centers rather than all processed at the core.
Furthermore, moving AI inferencing closer to where services are consumed can help to reduce inferencing response times. Two very different inference profiles live at the edge—each with distinct storage needs:
Why it matters for Solidigm: pairing the right SSD to each profile maximizes $/performance. The QLC-based Solidigm™ D5-P5336 drives shine for massive LLM repositories, while Solidigm TLC-based D7-Series drives cover vision workloads that punish NAND with ≥3 DWPD write rates—providing longevity without over-provisioning.
Beyond performance, data reduction is another advantage of work done at the edge. AI inferencing is increasingly a principal means by which systems distinguish between good data versus bad or redundant data, which would allow elimination of bad data without impact to system capabilities.
Similarly, regulations may require personal data processing to be done in-country or at an institution. Doing inferencing at the edge can help meet border requirements.
AI inferencing is not the only IT workload that can help organizations meet their system goals by deploying it at the edge. But it is a great illustration of work done at the edge that showcases many of the reasons enterprises use to justify deploying edge processing.
Distributed edge processing makes a lot of sense for organizations that have users far from core data centers. Responsiveness is often a main driver for introducing edge processing, but it’s not the only one.
By using edge processing to reduce and eliminate extraneous data at the source, data transmission, processing, and storage costs can all be reduced considerably. Moreover, scaling systems often run into software and hardware limits at central data centers, but by deploying edge processing, organizations can mitigate these constraints.
AI inferencing is a good paradigm where edge processing can pay extensive dividends. As AI applications become ever more ubiquitous in enterprise IT, inferencing done at the edge can be a make-or-break decision across several dimensions.
Making smart data infrastructure choices is a key part of that process. For more information on the Solidigm product portfolio and our recommendations for edge AI, see www.solidigm.com/ai.
Ace Stryker is the director of market development at Solidigm, where he focuses on emerging applications for the company’s portfolio of data center storage solutions.
Nothing herein is intended to create any express or implied warranty, including without limitation, the implied warranties of merchantability, fitness for a particular purpose, and non-infringement, or any warranty arising from course of performance, course of dealing, or usage in trade.
The products described in this document may contain design defects or errors known as “errata,” which may cause the product to deviate from published specifications. Current characterized errata are available on request.
Solidigm does not control or audit third-party data. You should consult other sources to evaluate accuracy.
Contact your Solidigm representative or your distributor to obtain the latest specifications before placing your product order.
SOLIDIGM and the Solidigm “S” logo are trademarks of SK hynix NAND Product Solutions Corp. (d/b/a Solidigm), registered in the United States, People’s Republic of China, Japan, Singapore, the European Union, the United Kingdom, Mexico, and other countries.