Can robots write? Machine learning produces dazzling results, but some assembly is still required
Shutterstock
Alexandra Louise Uitdenbogerd, RMIT University
You might have seen a recent article from The Guardian written by โa robotโ. Hereโs a sample:
I know that my brain is not a โfeeling brainโ. But it is capable of making rational, logical decisions. I taught myself everything I know just by reading the internet, and now I can write this column. My brain is boiling with ideas!
Read the whole thing and you may be astonished at how coherent and stylistically consistent it is. The software used to produce it is called a โgenerative modelโ, and they have come a long way in the past year or two.
But exactly how was the article created? And is it really true that software โwrote this entire articleโ?
How machines learn to write
The text was generated using the latest neural network model for language, called GPT-3, released by the American artificial intelligence research company OpenAI. (GPT stands for Generative Pre-trained Transformer.)
OpenAIโs previous model, GPT-2, made waves last year. It produced a fairly plausible article about the discovery of a herd of unicorns, and the researchers initially withheld the release of the underlying code for fear it would be abused.
But letโs step back and look at what text generation software actually does.
Machine learning approaches fall into three main categories: heuristic models, statistical models, and models inspired by biology (such as neural networks and evolutionary algorithms).
Heuristic approaches are based on โrules of thumbโ. For example, we learn rules about how to conjugate verbs: I run, you run, he runs, and so on. These approaches arenโt used much nowadays because they are inflexible.
Writing by numbers
Statistical approaches were the state of the art for language-related tasks for many years. At the most basic level, they involve counting words and guessing what comes next.
As a simple exercise, you could generate text by randomly selecting words based on how often they normally occur. About 7% of your words would be โtheโ โ itโs the most common word in English. But if you did it without considering context, you might get nonsense like โthe the is night awareโ.
More sophisticated approaches use โbigramsโ, which are pairs of consecutive words, and โtrigramsโ, which are three-word sequences. This allows a bit of context and lets the current piece of text inform the next. For example, if you have the words โout ofโ, the next guessed word might be โtimeโ.
This happens with the auto-complete and auto-suggest features when we write text messages or emails. Based on what we have just typed, what we tend to type and a pre-trained background model, the system predicts whatโs next.
While bigram- and trigram-based statistical models can produce good results in simple situations, the best recent models go to another level of sophistication: deep learning neural networks.
Imitating the brain
Neural networks work a bit like tiny brains made of several layers of virtual neurons.
A neuron receives some input and may or may not โfireโ (produce an output) based on that input. The output feeds into neurons in the next layer, cascading through the network.
The first artificial neuron was proposed in 1943 by US neuroscientists Warren McCulloch and Walter Pitts, but they have only become useful for complex problems like generating text in the past five years.
To use neural networks for text, you put words into a kind of numbered index. You can use the number to represent a word, so for example 23,342 might represent โtimeโ.
Neural networks do a series of calculations to go from sequences of numbers at the input layer, through the interconnected โhidden layersโ inside, to the output layer. The output might be numbers representing the odds for each word in the index to be the next word of the text.
In our โout ofโ example, number 23,432 representing โtimeโ would probably have much better odds than the number representing โdoโ.
Whatโs so special about GPT-3?
GPT-3 is the latest and best of the text modelling systems, and itโs huge. The authors say it has 175 billion parameters, which makes it at least ten times larger than the previous biggest model. The neural network has 96 layers and, instead of mere trigrams, it keeps track of sequences of 2,048 words.
The most expensive and time-consuming part of making a model like this is training it โ updating the weights on the connections between neurons and layers. Training GPT-3 would have used about 262 megawatt-hours of energy, or enough to run my house for 35 years.
GPT-3 can be applied to multiple tasks such as machine translation, auto-completion, answering general questions, and writing articles. While people can often tell its articles are not written by human authors, we are now likely to get it right only about half the time.
The robot writer
But back to how the article in The Guardian was created. GPT-3 needs a prompt of some kind to start it off. The Guardianโs staff gave the model instructions and some opening sentences.
This was done eight times, generating eight different articles. The Guardianโs editors then combined pieces from the eight generated articles, and โcut lines and paragraphs, and rearranged the order of them in some placesโ, saying โediting GPT-3โs op-ed was no different to editing a human op-edโ.
This sounds about right to me, based on my own experience with text-generating software. Earlier this year, my colleagues and I used GPT-2 to write the lyrics for a song we entered in the AI Song Contest, a kind of artificial intelligence Eurovision. https://www.youtube.com/embed/sAzULywAHUM?wmode=transparent&start=0 AI song Beautiful the World, by Uncanny Valley.
We fine-tuned the GPT-2 model using lyrics from Eurovision songs, provided it with seed words and phrases, then selected the final lyrics from the generated output.
For example, we gave Euro-GPT-2 the seed word โflyingโ, and then chose the output โflying from this world that has gone apartโ, but not โflying like a trumpetโ. By automatically matching the lyrics to generated melodies, generating synth sounds based on koala noises, and applying some great, very human, production work, we got a good result: our song, Beautiful the World, was voted the winner of the contest.
Co-creativity: humans and AI together
So can we really say an AI is an author? Is it the AI, the developers, the users or a combination?
A useful idea for thinking about this is โco-creativityโ. This means using generative tools to spark new ideas, or to generate some components for our creative work.
Where an AI creates complete works, such as a complete article, the human becomes the curator or editor. We roll our very sophisticated dice until we get a result weโre happy with.
Alexandra Louise Uitdenbogerd, Senior Lecturer in Computer Science, RMIT University
This article is republished from The Conversation under a Creative Commons license. Read the original article.
Observer Voice is the one stop site for National, International news, Editorโs Choice, Art/culture contents, Quotes and much more. We also cover historical contents. Historical contents includes World History, Indian History, and what happened today. The website also covers Entertainment across the India and World.