What can GenAI tools create?
The GenAI tool that has received the most media attention is ChatGPT Links to an external site., but there are many, many others becoming available. They all use LLMs to analyse and generate text in a linguistically correct way: outputs can be generated in seconds and they make look plausible. The outputs can be textual, or image-based, including video.
Text
Text outputs can be created in the style of your choosing. Because the language is usually accurate in terms of grammar and spelling, they can look very convincing, but that is not guarantee that the facts underlying the output is accurate or unbiased.
Here, I have asked ChatGPT to “write me a pop song, in the style of Abba, about how ChatGPT can be used in universities” (ChatGPT 3.5, April 2023)
It is linguistically acceptable, but you may have objections from an artistic perspective. I did ask it to improve the chorus:
In mid-2023, images, video and audio generated by AI tools make less sense than text, but they are improving all the time. In early 2023, image generation was easily identified by strange faults with the pictures (particularly hands and teeth).
These initially-plausible but on-closer-inspection-odd images (check the teeth of the woman in the green dress in the left hand image, the tattoos on the woman in the pale pink dress and the hand holding the camera in the right hand image) were generated by Twitter user @Mileszim Links to an external site. using the tool MidJourney Links to an external site. in January 2023. You may also be aware that many skilled artists had difficulty drawing feet and hands Links to an external site. before AI tools.
However, by mid-2023 some much more plausible, if not actually realistic, images were around to fool people:
“Babies at Viking Baby Camp”, attributed to Timothy J Nehman Links to an external site. who may or may not be a real person. The hands of the child on the right look a little unreal, but maybe not outside the normal range that children may have, in contrast with the camera-holder above.
Please note: there is no such thing as Viking Baby Camp. Yet. Although I did see this on the bus in Stockholm in summer 2023:
Key point: The main point to notice at this stage is that artefacts generated by GenAI tools are generally plausible, and we may have difficulty distinguishing them from artefacts which we generally consider to have been generated by humans, whether or not the humans also used technology in their creation.
Emily Bender has documented a good process Links to an external site. to try to establish the veracity of an image, using this super-cute image which circulated widely on social media (you already know it’s AI generated and does not resemble a baby peacock, of course, but it’s worth reading through and probably also keeping an eye on Emily’s blog if you are interested in the challenges of GenAI tools.
In summary, the outputs such as song lyrics or images from a GenAI tool may be plausible, but seem not quite right to a human, and the same will of course be true about any other outputs such as text. You may already have thought about the obvious next challenge: the outputs may also be factually inaccurate, because the system is just predicting probable words or data points. These are major limitations which will be the focus of attention by software developers over the next few years, and probably the tools will improve in all these respects. I say ‘probably’, because there is a possibility that the LLMs will end up using so much training material which is in itself outputs from GenAI tools and hasn’t been checked by humans that they could actually get worse rather than better.