Page 44 - EE Times Europe Magazine | April2019
P. 44

44 EE|Times EUROPE — Boards & Solutions Insert



         SPECIAL REPORT: AI AT THE EDGE
        Seeing the AI Inference Market with


        2020 Vision: Five Predictions

        By Geoff Tate
        A                                                                       to get more throughput for the same power
                 new class of artificial-intelligence chips is coming to market in 2020,
                                                                                and price as the solutions they are using today.
                 and the entries will be optimized for inference — not for graphics,
                                                                                The only way to get higher prediction accuracy
                 training, or digital signal processing. In fact, according to Nvidia in
                                                                                will be to get higher throughput per dollar and
                 its recent quarterly earnings call, the AI inference market exceeded AI
        training spending in the data center in 2019.                           higher throughput per watt.
                                                                                4. SELECTING THE RIGHT SOLUTION FOR
          What is surprising here is that very few startups introduced new products this   AUTOMOTIVE
        past year, considering the number of startups funded and the number of compa-  The automotive market for inference will be
                                                                                in the millions of dollars, but deployment
        nies that have been in business for five or more years. Flex Logix expects that to
                                                                                takes time. Car companies and their suppli-
        change this year. Below are our top five predictions based on what we see in the   ers have already selected solutions for the
        market and the many conversations we are having with customers.         2024–2025 model years and will evaluate
                                                                                alternatives in 2020 for the 2026–2027 model
                                                                                year. A year ago, all of the car companies/
        1. THROUGHPUT OVER MEANINGLESS      ily correlate with higher throughput; what   suppliers planned their own chips, but almost
        BENCHMARKS                          customers really want is high throughput per   all have dropped those plans and will use
        A year ago, customers were asking about   dollar. As an example, if you compare the Flex   merchant market solutions instead.
        tera-operations per second (TOPS) and    Logix InferX X1 device with a market-leading
        ResNet-50 at various batch sizes. Today,   graphics processor, the GPU may offer 3× to 4×   5. ADVANTAGE FOR SOLUTIONS WITH
        leading customers have developed models   the throughput, with 10× the TOPS, but it also   BF16
        that work for their applications, and how a   uses 8× the number of DRAMs. The InferX X1   While 8-bit integer (INT8) offers the highest
        solution runs their model is what’s import-  architecture is a lot more resource-efficient.  throughput per dollar and throughput per
        ant to them. The thing that matters most is                             watt, winning solutions will need to have
        throughput on megapixel images (not mean-  3. PREDICTION ACCURACY REQUIRES   a BFloat16 (BF16) option because it allows
        ingless benchmarks), and more companies are   HIGHER THROUGHPUT PER DOLLAR AND   customers to quickly ramp production. For
        going to figure this out in 2020.   PER WATT                            many customers, the cost and complexity of
                                            Applications are being starved for inference   quantizing may never be economical.
        2. INFERENCE THROUGHPUT AT LOWEST   throughput and have been getting by with   It’s going to be an exciting year in the AI/
        COST                                down-sampling their native megapixel images   inference market. There is no doubt that the
        In the server market, some customers will   and processing only a fraction of the frames   availability of high-throughput inference
        want to get more inference per PCIe slot, but   per second. In the market for systems that   capabilities will change the way people live,
        the path to expanding the market will be to   need 5- to 30-W chips/modules with heat sinks   work, and play. ■
        deliver inference throughput at lower price   and no fans (now primarily served by Nvidia
        points. Having more TOPS doesn’t necessar-  Xavier AGX and now NX), customers will want   Geoff Tate is the CEO of Flex Logix.


















                                                                                                                   IMAGE: SHUTTERSTOCK










        APRIL 2020 | www.eetimes.eu
   39   40   41   42   43   44   45   46   47   48   49