A million people eager to dabble with a new artificial intelligence tool that lets them create images simply by describing them with words will soon get their wish, its creators said Wednesday.
Artificial intelligence research firm OpenAI is conducting a wide-scale beta test of DALL-E, a cutting-edge software that creates images from textual descriptions.
Although the arrival of AI has led to fears of humans being replaced by machines in fields from customer care to journalism, enthusiasts see the technology more as an opportunity than a threat.
A video showcasing the tool on the company’s web site showed DALL-E generating an image of a polar bear playing a guitar, a photo of a koala dunking a basketball, and the famed Mona Lisa painting but with the subject sporting a mohawk hair style.
“We’ve already seen people use DALL·E to make music videos for young cancer patients, create magazine covers, and bring novel concepts to life,” the company said in a post.
OpenAI said invitations will be sent in coming weeks to a million people on a waiting list to try the tool’s latest version.
Amid concerns that this and similar tools could be misused in disinformation campaigns, OpenAI said it worked with researchers and developers to build in safeguards to curb abuses such as DALL-E being used for deception.
“We reject image uploads containing realistic faces and attempts to create the likeness of public figures, including celebrities and prominent political figures,” Open AI said.
Filters built into DALL-E block violent, political, sexual or other content barred by its policies, and the system is designed to avoid assumptions about race or gender, Open AI said.
New module for OpenAI GPT-3 creates unique images from text
© 2022 AFP
New AI tool that turns words into art enters testing phase (2022, July 21)
retrieved 21 July 2022
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.