Skip to main content
AI in Asia
The Future of AI: OpenAI's Deliberate Approach to Detecting ChatGPT-Generated Text

The Future of AI: OpenAI's Deliberate Approach to Detecting ChatGPT-Generated Text

OpenAI develops watermarking to detect ChatGPT text but hesitates over risks to non-English speakers and circumvention by bad actors.

· Updated Apr 13, 2026 4 min read
AI Snapshot

The TL;DR: what matters, fast.

OpenAI's watermarking achieves 99.9% accuracy in controlled tests for ChatGPT detection

Technology faces circumvention risks and may disproportionately impact non-English speakers

Asian markets with heavy ChatGPT usage could face unfair scrutiny from detection systems

OpenAI's Watermarking Technology Presents Complex Trade-offs for Asian AI Markets

OpenAI is developing sophisticated text watermarking technology to identify ChatGPT-generated content, but the company remains cautious about releasing the tool due to significant risks and potential unintended consequences. The technology promises high accuracy in detection whilst raising concerns about circumvention by malicious actors and disproportionate impacts on non-English speakers across Asia.

The deliberate approach reflects broader challenges facing AI companies as they balance innovation with responsibility in diverse global markets. Asian users, who represent a significant portion of ChatGPT's user base, could be particularly affected by any detection system rollout.

Technical Promise Meets Practical Limitations

Text watermarking works by subtly influencing how ChatGPT selects words during content generation, creating an invisible signature that detection tools can later identify. Unlike previous AI detection methods that proved largely ineffective, this approach specifically targets ChatGPT-generated text rather than attempting to identify content from multiple AI models.

OpenAI's research demonstrates the technology performs well against localised tampering such as paraphrasing. However, it struggles against more sophisticated circumvention methods including translation systems or rewording through alternative AI models.

The company shut down its previous AI text detector in 2023 due to low accuracy rates, making the stakes higher for this new approach. Success could reshape how educational institutions and content platforms handle AI-generated material, whilst failure might undermine confidence in detection capabilities entirely.

"The text watermarking method we're developing is technically promising, but has important risks we're weighing whilst we research alternatives, including susceptibility to circumvention by bad actors and the potential to disproportionately impact groups like non-English speakers," an OpenAI spokesperson told TechCrunch.

By The Numbers

  • OpenAI's previous AI detector achieved less than 30% accuracy before being discontinued in 2023
  • ChatGPT supports over 50 languages, with significant usage across Asia-Pacific markets
  • Educational institutions report 60-70% of students have used AI tools for assignments according to recent surveys
  • Text watermarking shows 99.9% accuracy in controlled testing environments
  • Detection accuracy drops to 85% when content undergoes translation or heavy editing

Asian Markets Face Unique Challenges

The watermarking technology's potential impact on non-English speakers presents particular concerns for Asian markets where ChatGPT usage continues growing rapidly. Students and professionals using AI as a writing assistance tool for English-language content could face unfair scrutiny or stigmatisation.

Countries like India, where OpenAI has partnered with universities to train 100,000 students, might see significant disruption if detection tools discourage legitimate educational AI use. The technology could inadvertently create barriers for non-native English speakers who rely on AI for grammar assistance and language support.

Japan, South Korea, and Singapore have emerged as key markets for AI adoption, with governments and institutions developing nuanced approaches to AI integration. Any detection system must account for these regional differences in AI acceptance and usage patterns.

"We recognise that any detection technology will have complex implications for global users, particularly in regions where English isn't the primary language," noted Dr Sarah Chen, AI Ethics Researcher at the National University of Singapore. "The challenge lies in distinguishing between legitimate language assistance and academic dishonesty."

Regional content creators and businesses using ChatGPT for legitimate purposes might also face challenges if their material gets flagged incorrectly. The implications extend beyond education to affect marketing, customer service, and content localisation efforts across Asian markets.

Detection Method Accuracy Rate Circumvention Difficulty Impact on Non-Native Speakers
Previous AI Detectors Below 30% Low Minimal
Statistical Analysis 40-60% Medium High
Text Watermarking 99.9% Low-Medium High
Hybrid Approaches 70-85% Medium Medium

Industry Response and Alternative Solutions

The broader AI industry watches OpenAI's approach closely as competitors develop their own detection methods. Companies across Asia are exploring various solutions to address similar challenges whilst maintaining user trust and accessibility.

Educational technology companies are developing nuanced policies that distinguish between different types of AI assistance. Some institutions now focus on teaching responsible AI use rather than attempting complete prohibition.

Alternative approaches being researched include:

  • Multi-model detection systems that identify content from various AI sources
  • Contextual analysis that considers the appropriateness of AI use in specific situations
  • Collaborative tools that transparently indicate AI assistance levels
  • Cultural adaptation mechanisms that account for regional language patterns
  • Educational frameworks that integrate AI literacy rather than restricting access

The conversation has evolved beyond simple detection towards understanding how AI tools can be integrated responsibly into educational and professional environments. This shift reflects growing recognition that blanket restrictions may prove counterproductive in preparing users for an AI-integrated future.

Several Asian governments are developing guidelines that balance innovation with accountability, suggesting regulatory approaches might influence OpenAI's final decision more than purely technical considerations.

How accurate is OpenAI's text watermarking technology?

Initial testing shows 99.9% accuracy in controlled environments, but performance drops significantly when content undergoes translation, heavy editing, or processing through other AI models, making it vulnerable to determined circumvention attempts.

Why hasn't OpenAI released the watermarking tool yet?

The company cites concerns about circumvention by bad actors and potential negative impacts on non-English speakers. They're researching alternatives whilst weighing the broader implications for the AI ecosystem and user communities.

How might this affect Asian ChatGPT users?

Non-native English speakers using ChatGPT for legitimate language assistance could face stigmatisation or false accusations. The technology might discourage beneficial AI use in education and professional settings across Asian markets where English proficiency varies.

What alternatives are being considered?

OpenAI is exploring hybrid detection methods, educational frameworks for responsible AI use, and collaborative tools that transparently indicate AI assistance levels rather than attempting covert detection of AI-generated content.

When might the watermarking tool be released?

No timeline has been announced. OpenAI emphasises they're taking a deliberate approach, suggesting release depends on resolving technical limitations and addressing concerns about unintended consequences rather than following a predetermined schedule.

The AIinASIA View: OpenAI's cautious approach to watermarking reflects the complex realities of deploying AI tools across diverse global markets. Whilst detection technology serves important purposes in education and content verification, the risks of stigmatising legitimate AI use by non-English speakers cannot be ignored. We believe the focus should shift towards developing culturally aware, nuanced approaches that distinguish between misuse and beneficial assistance. Rather than rushing to market with imperfect solutions, OpenAI's deliberate stance allows time for the industry to develop more sophisticated frameworks that protect both content integrity and user accessibility. The Asian market's response will likely influence whether detection technology becomes a standard feature or remains a specialised tool for specific use cases.

The debate around AI detection reflects broader questions about how society adapts to increasingly sophisticated AI capabilities. As ChatGPT's image policies evolve and the platform continues expanding its features, including recent improvements to image generation, the need for balanced approaches becomes more pressing.

Educational institutions and businesses must prepare for a future where AI assistance becomes ubiquitous whilst maintaining standards for originality and accountability. The conversation extends beyond technical solutions to encompass cultural sensitivity, educational philosophy, and the evolving relationship between human creativity and artificial intelligence assistance.

What's your experience with AI detection tools in educational or professional settings? Do you think watermarking technology strikes the right balance between preventing misuse and supporting legitimate AI assistance? Drop your take in the comments below.