Page 41 - EETEurope FlipBook February
P. 41

EE|Times EUROPE   41



               OPINION | MARKET & TECHNOLOGY TRENDS

               Are Today’s Huge AI


               Language Models Out



               of Control?



               By Sally Ward-Foxton


                                   Natural-language processing (NLP) is a popular and
                                   rapidly evolving field of AI. But the evolution of more sophisticated
                                   language models comes with an overarching mega-trend: These
                                   models are growing at a staggering rate.
                                     Google’s latest language model has 1.6 trillion parameters — a
                                   figure even its inventors called “outrageous” in their paper. The
                                   model’s parameters are the figures that the training process
                                   adjusts, and in general, more parameters mean a more sophisti-
                                   cated AI. Models like this one that purport to understand language
               are usually given a text prompt and then create more text in the same vein. The AI does not
               really understand language; it merely mimics it to a convincing degree.
                 Google’s model topped the previous world-record holder, OpenAI’s GPT-3, by a factor of 9.
               GPT-3’s 175 billion parameters were widely heralded as enormous, and examples of the texts
               it created were marvelled over by many news outlets. In many cases, GPT-3’s text output was
               indistinguishable from articles written by humans; the model can mimic language concepts
               such as analogies and even write basic software code. GPT-3 itself was multiples bigger than
               its predecessor, GPT-2, which had 1.5 billion parameters.
                                              Language models like these have grown to the
               The biggest language          size at which the financial and environmental costs
                                             associated with computing them invite scrutiny. A
               model yet, from               2019 study by researchers from the University of Mas-
               Google, has 1.6 trillion      sachusetts Amherst found that training GPT-2 took
                                             state-of-the-art specialized hardware (32× TPUv3
               parameters — a figure         chips) 168 hours with an estimated cloud compute
               even its inventors called     cost of between US$12,000 and US$43,000. That’s
                                             training one model, one time. Models are typically
               “outrageous.”                 trained and tweaked and retrained many times over
                                             during the development process.
                                              Now consider that Google’s latest model is 1,000×
               bigger than GPT-2. If we assume the compute required to train scales roughly with the model’s
               size, we start to get an idea of the scale of resources required to develop and use these models.
               (Hint: It’s many millions of dollars). And the carbon footprint associated with compute on this
               scale is significant, too.
                 Are today’s language models too large, and aside from financial and environmental cost,   able to readily produce convincing misin-
               what are the implications?                                              formation at scale and the havoc that rogue
                 In practice, these models are now so large that Google is one of only a few companies that   chatbots could wreak given the ability to
               can do this kind of research. The company has access to huge amounts of computing power   pass for a human.
               in-house which, while accessible via the cloud, are not economically viable at such scale for   Only a few weeks before Google published
               smaller companies or academic researchers. This allows tech giants like Google to effectively   its latest work on the trillion-parameter
               dictate the direction of research in this field, and naturally, research will follow whatever path   model, a prominent ethics researcher with
               is in their commercial interest. Whether their commercial interest aligns with ethical best   the company said she had been forced out
               practices or will lead to language models that can actually understand language is unclear.    after authoring a paper on the ethics of large
                 Where does one get the vast amount of training data required for a model this large? The   language models. In light of the criticism
               short answer is the internet. Text is scraped from sources such as Wikipedia and Reddit. There   this move drew, Google will find it difficult
               is a significant risk that trained models’ output will reflect the nature of the training data,   to keep answering questions about the ethics
               i.e., that it will mimic things people write in the internet’s darkest corners. If the model sees   of research into huge — and growing — AI
               humanity at its worst, we should expect its output to reflect that.     language models. ■                 IMAGE: SHUTTERSTOCK
                 There are also questions about the potential for misuse of such powerful models. If a
               machine can write on any topic as well as a human can, it suddenly becomes very easy to fool   Sally Ward-Foxton is editor-in-chief of
               people — especially over the internet, where people communicate via text. Imagine being   EE Times Weekend.


                                                                                          www.eetimes.eu | FEBRUARY 2021
   36   37   38   39   40   41   42   43   44   45   46