Page 16 - EE|Times Europe Magazine - December 2020
P. 16

16 EE|Times EUROPE — The Memory Market



         SPECIAL REPORT: MEMORY TECHNOLOGY
        Memory Technologies Confront Edge AI’s


        Diverse Challenges


        By Sally Ward-Foxton

                  ith the rise of artificial intelligence at the edge comes a
                  whole host of new requirements for memory systems.
                  Can today’s memory technologies live up to the stringent
        Wdemands of this challenging new application, and what do
        emerging memory technologies promise for edge AI in the long term?
          The first thing to realize is that there is no standard “edge AI” appli-
        cation; the edge in its broadest interpretation covers all AI-enabled
        electronic systems outside the cloud. That might include “near edge,”   An Intel Optane 200 Series module. Intel says Optane is already
        which generally covers enterprise data centers and on-premise servers.  being used to power AI applications. (Source: Intel)
          Further out are applications like computer vision for autonomous
        driving. Gateway equipment for manufacturing performs AI inference
        to check for flaws in products on the production line. 5G “edge boxes”   LRDIMM, and highly available persistent memory like NVDIMM.”
        on utility poles analyze video streams for smart-city applications such   Gupta sees Intel Optane, the company’s 3D-Xpoint nonvolatile mem-
        as traffic management. And 5G infrastructure uses AI at the edge for   ory whose properties are between DRAM and flash, as a good solution
        complex but efficient beamforming algorithms.         for server AI applications.
          At the “far edge,” AI is supported in devices such as mobile phones   “Both Optane DIMMs and NVDIMMs are being used as AI acceler-
        — think Snapchat filters — voice control of appliances, and IoT sensor   ators,” he said. “NVDIMMs provide very low-latency tiering, caching,
        nodes in factories performing sensor fusion before sending the results   write buffering, and metadata storage capabilities for AI application
        to another gateway device.                            acceleration. Optane data center DIMMs are used for in-memory
          The role of memory in edge AI systems — to store neural network   database acceleration, where
        weights, model code, input data, and intermediate activations — is   hundreds of gigabytes to terabytes
        the same for most AI applications. Workloads must be accelerated   of persistent memory are used in
        to maximize AI computing capacity in order to remain efficient, so   combination with DRAM. Although
        demands on capacity and bandwidth are generally high. However,   these are both persistent memory
        application-specific demands are many and varied and may include   solutions for AI/ML acceleration
        size, power consumption, low-voltage operation, reliability, thermal/  applications, they have different
        cooling considerations, and cost.                     and separate use cases.”
                                                                Kristie Mann, Intel’s director
        EDGE DATA CENTERS                                     of product marketing for Optane,
        Edge data centers are a key edge market. The use cases include medical   told EE Times that Optane is
        imaging, research, and complex financial algorithms, in which privacy   gaining applications in the server
        prevents uploading to the cloud. Another is self-driving vehicles, in   AI segment.
        which latency prevents it.                              “Our customers are already using   Intel’s Kristie Mann
          These systems use the same memories found in servers in other   Optane persistent memory to power
        applications.                                         their AI applications today,” she said. “They are powering e-commerce,
          “It is important to use low-latency DRAM for fast, byte-level main   video recommendation engines, and real-time financial analysis usages
        memory in applications where AI algorithms are being developed   successfully. We are seeing a shift to in-memory applications because
                                 and trained,” said Pekon Gupta,   of the increased capacity available.”
                                 solutions architect at Smart Mod-  DRAM’s high prices make Optane an attractive alternative. A server
                                 ular Technologies, a designer and   with two Intel Xeon Scalable processors plus Optane persistent mem-
                                 developer of memory products.   ory can hold up to 6 TB of memory for data-hungry applications.
                                 “High-capacity RDIMMs [registered   “DRAM is still the most popular, but it has its limitations from a
                                 dual-in-line memory modules] or   cost and capacity perspective,” said Mann. “New memory and stor-
                                 LRDIMMs [load-reduced DIMMs] are   age technologies like Optane persistent memory and Optane SSD are
                                 needed for large datasets. NVDIMMs   [emerging] as an alternative to DRAM due to their cost, capacity, and
                                 [nonvolatile DIMMs]  are needed for   performance advantage. Optane SSDs are particularly powerful, caching
                                 system acceleration — we use them   HDD and NAND SSD data to continuously feed AI applications data.”
                                 for write caching and checkpointing   Optane also compares favorably with other emerging memories that
                                 instead of slower SSDs.”     are not fully mature or scalable today, she added.
                                   Locating computing nodes close
        Smart Modular            to end users is the approach taken   GPU ACCELERATION
        Technologies’ Pekon Gupta  by telecommunications carriers.  For high-end edge data center and edge server applications, AI com-
                                   “We’re seeing a trend to make   pute accelerators such as GPUs are gaining traction. As well as DRAM,
        these [telco] edge servers more capable of running complex algo-  the memory choices here include graphics double data rate (GDDR),
        rithms,” Gupta said. Hence, “service providers are adding more memory   a special DDR SDRAM designed to feed high-bandwidth GPUs, and
        and processing power to these edge servers using devices like RDIMM,   high-bandwidth memory (HBM), a relatively new die-stacking

        DECEMBER 2020 | www.eetimes.eu
   11   12   13   14   15   16   17   18   19   20   21