TAIPEI, Taiwan--(BUSINESS WIRE)--How much computing power is needed at the edge? How much memory and storage are enough for AI at the edge? Minimum requirements are growing as AI opens the door to innovative applications that need more and faster processing, storage, and memory. How can today’s memory and storage technologies meet the stringent requirements of these challenging new edge applications?
What do we mean by “the edge”?
Edge includes any distributed application where specific processing occurs away from the server, even if the data is eventually sent to a data center. The big idea is to avoid sending all the data over the internet for processing on a server and instead allow data to be processed closer to where it’s collected, avoiding latency issues with long data roundtrips, and enabling near real-time response on site.
Data latency and bandwidth at the industrial edge
In industrial applications, edge computers are typically designed to take inputs from sensors or other devices and act on the inputs accordingly. For example, preventative maintenance takes vibration, temperature, or pressure sensor readings and analyzes them to identify anomalies that indicate slight faults in machines. Machines can be taken offline immediately or when needed to enable maintenance to occur ahead of catastrophic failure. Reaction times must be quick, but data quantity is low. However, AI is putting a strain on these edge systems. AI places a different kind of load on computer systems. AI workloads require faster processors, more memory, and powerful GPUs. AOI, for example, has seen widespread adoption for PCB inspection, using video input from high-speed cameras to identify missing components and quality defects.
Bringing a sliver of data center power to the edge
Essentially, to tackle AI tasks at the edge, we’re bridging the gap between the edge and the data center. Servers tucked away in temperature-controlled data centers have terabytes of memory and vast amounts of storage on hand to handle specific high-capacity loads and keep systems working fast. But when it comes to inference happening far from the data center, it’s a different story. Edge computers don’t enjoy such idyllic settings and must be built to withstand harsh environments. The edge needs hardware that strives for maximum performance while accounting for less-than-ideal conditions.
Hardware for the edge
Adding AI at the industrial edge requires hardware suited to the task. Three things are needed for vision systems, the most prolific AI application to date, memory to support efficient AI inference, storage for the incoming data, and PoE to support the addition of cameras.
Getting more memory in a smaller space can be accomplished with the latest DDR5. It provides more memory capacity at the edge with higher speeds.
Extending capacity is needed for edge applications, as the data must go to the server or stay at the edge for some time, so SSDs are needed for interim storage. The shift from SATA to NVMe has opened the doors to greater speeds and performance and the NVMe PCIe G4X4 SSD, available soon, is the latest SSD in Cervoz’s pipeline, providing the industrial performance for these applications.
Vision systems need cameras. PoE+ is the simplest and most efficient way to add high-speed cameras to the system, providing both power and data transmission through a single cable. Cervoz’s PoE Ethernet Modular PCIe Expansion Card adds this functionality through a small add-on for power.
Cervoz Technology supplies embedded components for the industrial PC market. The company has nearly two decades of experience designing and developing high-performing memory and storage solutions for industrial applications.