OpenAI became an overnight household name, at least in tech circles, by releasing ChatGPT in November 2022. Businesses and individuals quickly adopted its tools and began using them to automate daily tasks.
One issue with AI-generated text and images is that they can mirror existing works. Some have accused the software of plagiarism, while others feel it is not as detailed as human-created work. Employers everywhere want better ways to detect AI usage to meet target audiences’ needs and know the best time to utilize the technology.
Open AI responded by creating an AI detection tool but has not yet released it. Many want to know the reason for the holdup and if it will ever be made available to the public.
TechCrunch published an article explaining OpenAI's detection tool, which is meant to catch students who cheat by having the computer program write their papers. The company has sat on it rather than releasing it, making some question why. A representative for OpenAI said the company is considering the
Many other detection tools have been ineffective, creating false positives at times and failing to detect AI-generated text at others. OpenAI's product would focus mainly on detecting text watermarked by ChatGPT. One reason for not releasing it is fear that it could negatively impact nonnative English speakers. The tool also forces ChatGPT to change how it works when adding the watermarked words.
Although the detection software is meant to catch students using AI to write papers, it could be applied to many other situations, particularly in combating fraud at larger-scale levels.
For example, consider how useful an AI detector tool could be in catching computer-generated text used by cybercriminals. In 2023 alone, ransomware payments
Many scams include flowing verbiage meant to tug at someone’s emotional heartstrings or cause fear. When unsure whether a message is legitimate, people could run the text through the detection tool to see if AI was used to generate it. This could be quite valuable since many scams involve robot-written text.
So far, AI detection tools haven’t been very accurate. They’ve falsely flagged original content as AI-generated and caused students to have to fight for their grades and sometimes their academic careers. Universities and employers take plagiarism seriously, so a false accusation could destroy someone’s future or livelihood.
The tools may also be ineffective when users can__plug into easy-to-use and versatile software__ that helps them bypass AI detection, such as HumanizeAI and RealWriter. People should learn to utilize AI in ways that allow them to brainstorm and generate fresh ideas without plagiarizing information. However, they may shy away from using any type of software for fear of being flagged by a detector.
Opponents of the software fear it may drive people to be less creative for fear of triggering a detector. Programs like StealthAI will
People also express concerns over bad players using the tool to reverse engineer code and create something that will
Arguments against AI detection tools may soon be a moot point. While people expressed concern over nonnative English speakers getting left behind, software developers will likely fix language barrier problems as time passes.
Many believe OpenAI will soon release its tool. Reports indicate
As AI detector development grows, expect to see it used for things such as:
Identifying fake videos used in political campaigns
Finding misinformation and marking it more accurately than other bots do
Generating ethical content
Releasing the tool could inspire other brands to create their own programs and spur the success rate of AI detection.
Love it or hate it, AI is here to stay — although companies have been a bit slower to embrace it than first predicted. MIT released a report showing that the technology is being
As marketing departments learn more about how AI works and ways to implement it, expect more brands to use it and seek detector tools to prevent privacy issues and copying.
Those who use AI for a while start to notice patterns that can be detected without the help of software. For example, the intro to nearly every article has similar language. The layout, pacing, and details are the same. With a bit of practice, most people can spot an AI-generated piece fairly accurately.
Without a way to detect AI-generated content accurately, people may distrust what they read online. In other situations, they might suspect someone used AI without disclosing it but find it impossible to prove. Workers may be falsely accused or lose their jobs and degrees when they aren’t guilty.
The tool will solve the issue of students doing very little work and being lackluster in their studies but still obtaining an advanced degree based on someone else’s hard work, sweat, and tears.
OpenAI’s product will likely have the most impact in the education sector. However, companies can also utilize it to ensure they produce unique, highly targeted content customers want to read.
OpenAI’s decision to sit on its detection tool is rooted in wanting to do what is best as AI technology grows and ethical concerns arise. It will likely release the software at some point, to the relief of educators everywhere. How other industries might use it remains to be seen.
All companies developing AI programs should balance the pros and cons of each feature they introduce. Only then will AI become an indispensable tool that spurs progress.