Menu Menu
[gtranslate]

OpenAI may roll out a detector for AI-generated text

OpenAI claims to have a tool capable of unearthing instances where AI text generators have been used. Will it ever see the light of day?

OpenAI has crafted a clever tool to sniff out writing conjured by ChatGPT, but they’re having second thoughts about unleashing it onto the world โ€“ because, why shoot yourself in the foot?

This AI detective follows the demise of their previous AI Text Classifier, which flunked its performance review and failed several experiments. Fresh off the drawing board, however, OpenAI is boasting of a โ€˜highly accurateโ€™ text watermarking method while weighing up the pros and cons of releasing it to the public.

The dilemma? They worry the tool might unfairly stigmatise non-native English speakers using AI for writing help. The less noble (and more likely) explanation is that seniors at the company will know that the covert nature of text-generation tools help them proliferate. People are lazy and donโ€™t fancy being outed for it.

Besides, even if OpenAI wanted to release it, no tool can be entirely foolproof. Crafty users could dodge detection by translating text, rephrasing it with another AI, or even by sprinkling special characters in before deleting them afterwards.

Early experiments into metadata are far more reliable, as the code will reportedly hold strong indicators of AI use even after being modified. Perhaps the Microsoft-owned firm are reverse-engineering purely for peace of mind.

As AIโ€™s influence continually grows in the world of content, theyโ€™ll want to know for themselves where their technology is being harnessed. As for a public releaseโ€ฆ surely theyโ€™re bluffing.

While text detection is seemingly on the backburner, OpenAI is cooking up plans for spotting AI-generated audiovisual content, which poses more obvious risks at this stage.

They’re jumping on the C2PA metadata bandwagon, which attaches an origin tag to content that survives edits and makes it harder to fake authenticity. TikTok and Adobe are among those employing the same safeguards.

DALL-E, OpenAI’s image generator, will soon embed C2PA metadata as will their upcoming video generator โ€˜Soraโ€™. The idea, is that while people can still create misleading content and propaganda, they can’t conceal the red flags thrown up by the metadata.

OpenAI also announced a tamper-resistant watermarking solution for images, inviting research labs and journalism nonprofits to give it a whirl. This tool boasts a 98% success rate in identifying DALL-E 3 images, with a minimal false positive rate.

Lastly, audio detection is on OpenAIโ€™s radar with their new โ€˜Voice Engineโ€™ reportedly able to detect AI even from the most convincing samples. With tools like these, youโ€™d assume thatโ€™s the end of fake Drake tracks popping up on Spotify.

OpenAI is understandably reluctant to let companies and businesses get their hands on its AI text-detector. Nonetheless, it feels the need to secure the digital landscape before AI surpasses human-created content online.

Accessibility