OpenAI launches tool to detect deepfake content by its image generator
As experts warn that images, audio and video generated by artificial intelligence could influence the fall elections, OpenAI is releasing a tool designed to detect content created by its own popular image generator, DALL-E. But the prominent AI start-up acknowledges that this tool is only a small part of what will be needed to fight so-called deepfakes in the months and years to come.
On Tuesday, OpenAI said it would share its new deepfake detector with a small group of disinformation researchers so they could test the tool in real-world situations and help pinpoint ways it could be improved. “This is to kick-start new research,” said Sandhini Agarwal, an OpenAI researcher who focuses on safety and policy. “That is really needed.”