Page 39 - EE Times Europe Magazine – June 2024
P. 39

EE|Times EUROPE   39

                                                                    Nvidia GTC 2024: Why Nvidia Dominates AI


          As AI became the driving force behind GPU improvements, the tran-  given the company a dominant share of the GPU market for over a
        sistor count rose to more than 28 billion for the Ampere GPU in 2020   decade. Nvidia retained its leadership during the graphics card era
        and to 80 billion for the Hopper GPU in 2022. More parallel and spe-  and through the video processing era as well. Now, the CUDA software
        cialized computing functions are key to improving performance for AI   base and parallel processing capabilities are the leaders for AI model
        software. The Hopper GPU microarchitecture had especially extensive   development and AI application deployments. This AI software base
        improvements for accelerating AI computing. Hopper was the first GPU   advantage will persist for some time, even with increased competition
        with a transformer engine for acceleration of AI training models.  in the future.
          In March 2024, Nvidia announced Blackwell as the latest GPU   The question is whether Nvidia can extend its software base advan-
        generation, with 208 billion transistors. Product shipment is due   tages to new AI applications and other new application segments as
        later this year. The 2024 Blackwell GPU chip will more than double   they emerge. Nvidia has embraced and transitioned to new CUDA
        the Hopper performance and transistor count. However, much of the   opportunities before and is likely to do so again. Another question
        transistor increase is due to the use of two chip dies connected by a   is whether the Unified Acceleration (UXL) Foundation will have an
        10-TB/s chip-to-chip interconnect in a unified single GPU. Each die   impact. The UXL Foundation was launched in September 2023 to
        has 104 billion transistors—a 30% increase over Hopper’s 80 billion   develop a CUDA alternative and has an impressive membership roster,
        transistors. The Blackwell architecture can have up to 576 GPU    including Arm, Fujitsu, Google Cloud, Intel, Qualcomm, Samsung and
        processors on a chip. Blackwell includes a second-generation trans-  others.
                                      former engine.            The formation of UXL looks like an acknowledgement that CUDA is
                                       TSMC has manufactured
        The AI GPU trend              the vast majority of Nvid-  a clear leader and that no single company can compete with CUDA’s
                                                              software base and momentum. Only a foundation with major company
        accelerated Nvidia’s          ia’s chips, though IBM and   participation can or may build a CUDA competitor.
        market value and              Samsung have manufactured   HOW NVIDIA BECAME THE LEADER IN AI
                                      some chips for Nvidia in the
        revenue, and it is now        past 20 years. Samsung was   We have already discussed two reasons for Nvidia’s rise to AI domi-
                                      a fab partner from 2016 to
                                                              nance: GPU technology advances for AI calculations and the CUDA
        one of the three most         2022. SGS-Thomson       software platform for writing code for parallel execution of many
        valuable companies            Microelectronics was an early   streams of information. The technology advances of GPUs have been
                                                              very rapid—from 1 million transistors on a GPU in 1995 to 208 billion
                                      Nvidia fab partner.
        in the world based on         CUDA PLATFORM           transistors in 2024. This functionality and performance growth was key
                                                              to enabling the rapid expansion and complexity of AI models. There
        market capitalization.        GPUs were originally designed   are at least two more factors: building momentum in graphics- and
                                      for image manipulation   video-related GPU applications, and understanding and reacting to the
                                      and the calculation of local   potential of AI as a major future GPU market.
        image properties. The mathematical foundations of neural networks   On all counts, Nvidia’s management gets credit for crafting and
        and image manipulation are similar, which created a big opportunity   implementing a successful strategy.
        for GPUs as AI chips. In the mid-2010s, GPUs evolved to enable deep   Nvidia made a foresighted investment in the late 2000s to set up
        learning for training and inference in many applications, including   CUDA programming classes at most of the major universities
        autonomous vehicles.                                  worldwide—at Nvidia’s cost. The bold gambit paid off; it was part of the
          Nvidia defined a GPU programming interface called CUDA (which   creation of a market concept and enabled the later growth of the CUDA
        stands for Compute Unified Device Architecture, though Nvidia uses   software base.
        the acronym only) and released the first version in mid-2007. CUDA   Nvidia’s early realization that AI had great potential was critical.
        version 12.4 was released in March 2024. CUDA is a parallel computing   The company first focused on AVs as an AI opportunity and included
        platform and application programming interface that allows software   GPU functionality for speeding up AI calculation. The AV market was
        to use GPUs for accelerated general-purpose processing. Nvidia has   delayed, but generative AI came along to take AVs’ place—and Nvidia
        used “accelerated computing” to define its product strategy for about a   was ready to ride this market growth. The rest is a great history for
        decade.                                               Nvidia.
          The CUDA API is an extension of the C programming language that   Another key factor is that Nvidia became as much a software
        adds the ability to specify software-level parallelism in C and to spec-  company as a chip design company. Today, Nvidia is more of a soft-
        ify GPU-device–specific operations. CUDA is also a software layer that   ware company with highly skilled AI GPU chip designers. Nvidia has
        gives direct access to the GPU’s virtual instruction set and parallel   developed a large AI software base, including software development
        computational elements. CUDA is designed to work with program-  tools, AI libraries and AI foundational models for a large spectrum of
        ming languages like C, C++, Fortran and Python. This accessibility   AI applications.
        makes it easier to do parallel programming via the many processors   Through its AI software development activity, Nvidia gained an
        and accelerators provided in GPU chips. Many corporations program   understanding of what GPU hardware architecture and processing
        their AI applications in CUDA to get maximum performance from   accelerators were needed for AI performance, especially for the AI
        Nvidia GPU systems. This has built a broad CUDA software base and   training phase.
        a large population of programmers with expertise in developing soft-  This knowledge base was reinforced and amplified by building AI
        ware for Nvidia’s GPUs.                               systems for cloud system operators and similar customers. All of this
          Intel’s continued leadership in processor chips for the PC market   has given Nvidia the expertise to rapidly improve its AI-focused GPUs
        is attributable to the software base that was built around IBM’s PC   and the systems for AI training and inferencing. In the last two GPU
        standard in the 1980s and 1990s. Intel has leveraged that software   architecture upgrades, the focus was on improving AI performance for
        to maintain dominance in PC chips for more than 40 years, but it did   training and inferencing. AI application improvements will be the focus
        not transfer this software advantage to later computer segments like   of future GPU chip architectures.
        smartphones and tablets.                                Apple is known for gaining major advantages from designing both
          Nvidia has a similar software base, built around CUDA, that has   its hardware and software into a better system experience for its users.


                                                                                         www.eetimes.eu | JUNE 2024
   34   35   36   37   38   39   40   41   42   43   44