Formal academic prompt

 

Prompt

How can we design an algorithm to detect whether a written article, book, or public-domain text was created or significantly edited with AI assistance? I want a practical, research-oriented answer that covers detection signals, model design, evaluation methods, limitations, and whether there are any existing software tools on the market that can help identify AI involvement in texts.

Even better version for deeper insight

Design a robust algorithm for identifying AI assistance in written works such as articles and books, including public-domain texts that may have been partially generated or edited by AI. Explain the feature sets, statistical and machine-learning methods, training data requirements, evaluation metrics, and known failure cases. Also assess the current market for AI-detection software, comparing available tools, their strengths, weaknesses, and suitability for detecting partial AI use rather than fully AI-generated text.

Why this is better

  • It separates full AI generation from partial AI editing, which are different problems.
  • It asks for both algorithm design and market software review.
  • It requests limitations and failure cases, which is important because AI detection is often probabilistic, not certain.

If you want it for a paper or proposal

You could also use this version:

I am researching methods to detect AI involvement in written content. Please help me formulate a rigorous prompt for designing an algorithm that can identify AI-assisted authorship in articles and books, including public-domain works. The response should cover methodology, data sources, evaluation standards, uncertainty, ethical concerns, and commercially available detection tools.

If you want, I can also turn this into:

  1. research question,
  2. formal academic prompt, or
  3. step-by-step algorithm specification.

do all options (1,2,3).

Completed 2 steps

Absolutely — here are all three versions, refined for your topic.

1. Research question

How can we formulate a rigorous research question to investigate whether a written article or book was created, substantially edited, or partially revised with AI assistance, including cases where the text is in the public domain?

A strong research framing should distinguish between full AI generation and partial AI editing, because detection methods for mixed-authorship text are different from standard AI-vs-human classification. Recent work specifically addresses edited AI text and suggests this is a distinct and harder problem than detecting fully generated text.

2. Formal academic prompt

I am researching methods for detecting AI involvement in written works. Please help me formulate a formal academic prompt for designing an algorithm that identifies AI-assisted authorship in articles, books, and public-domain texts.

The response should address:

  • The problem definition and scope.
  • Textual, statistical, and stylometric features.
  • Machine learning or information-theoretic approaches.
  • Training and test data requirements.
  • Evaluation metrics, false positives, and failure cases.
  • Ethical and legal issues, especially for public-domain texts.
  • Existing software tools and their suitability for detecting partial AI use rather than only fully AI-generated text.

This framing is useful because current detectors are marketed for AI-generated text broadly, but the literature shows that edited or mixed-origin text is much harder to classify reliably.

3. Step-by-step algorithm specification

Design a detection pipeline for identifying AI assistance in written text.

Input

A document such as an article, essay, or book chapter, possibly from the public domain.

Output

A score or classification indicating:

  • likely human-written,
  • likely AI-generated,
  • likely AI-assisted or AI-edited,
  • uncertain.

Core steps

  1. Segment the document into sentences, paragraphs, or overlapping spans.
  2. Extract features such as perplexity, burstiness, repetition, sentence-length variation, lexical diversity, and stylometric fingerprints.
  3. Compare local segments against human and AI reference distributions.
  4. Aggregate segment-level predictions into a document-level estimate.
  5. Use a separate module to estimate mixed authorship or editing intensity.
  6. Calibrate confidence and flag low-certainty cases rather than forcing binary labels.

This is aligned with newer approaches that move beyond binary detection and instead estimate the extent of AI editing in a text. It also reflects the practical reality that modern detectors can be bypassed or weakened by paraphrasing, and that false positives remain a serious limitation.

Market software

If you also want to mention available software, current products include GPTZero, Copyleaks, Originality.ai, Winston AI, Sapling, and Pangram. Among these, GPTZero, Copyleaks, Originality.ai, and Winston AI are commonly positioned for general AI-text detection, while Pangram and Sapling advertise detection and document-analysis features, including text and file support.

That said, the market is best understood as probabilistic and imperfect, not definitive, because studies and reviews report limitations with paraphrasing, mixed-authorship text, and bias against some writing styles. If your use case is public-domain books or partially edited articles, you should treat detector output as evidence for review, not proof.

If you want, I can now turn this into:

  • single polished master prompt for ChatGPT,
  • research proposal outline, or
  • technical architecture diagram for the algorithm.

Create a single polished master prompt for ChatGPT,

You are an expert in natural language processing, stylometry, and AI detection systems. Design a comprehensive algorithm for detecting AI assistance in written works—including full generation, partial editing, or substantial revision—across articles, books, and public-domain texts.

Core Requirements

1. Problem Definition

  • Distinguish between fully AI-generated text, human text with AI editing, and mixed-authorship scenarios
  • Handle public-domain texts that may have been partially rewritten by AI
  • Output should include probability scores, confidence intervals, and uncertainty estimates

2. Feature Engineering (Multi-Level)

text

Text-level features:

- Perplexity, burstiness, sentence length variation

- Lexical diversity, n-gram repetition patterns

- Semantic coherence, topic consistency

 

Stylometric features:

- Function word ratios, punctuation patterns

- Character n-gram frequencies, part-of-speech distributions

- Readability scores across multiple formulas

 

AI-specific fingerprints:

- Probability distribution entropy

- Watermark detection (if present)

- Paraphrasing patterns, over-optimization signals

3. Algorithm Architecture

text

Phase 1: Document Segmentation

- Split into sentences, paragraphs, sliding windows

- Identify structural boundaries (sections, chapters)

 

Phase 2: Local Classification

- Train separate classifiers for each segment type

- Ensemble human-written, AI-generated, AI-edited labels

 

Phase 3: Global Aggregation

- Bayesian combination of local predictions

- Detect editing boundaries and intensity gradients

- Cross-validate against document metadata

 

Phase 4: Uncertainty Quantification

- Calibration using temperature scaling

- Flag low-confidence regions

- Provide explainability (feature importance)

4. Training Strategy

text

Required datasets:

- Human-written benchmarks (classic literature, academic papers)

- Pure AI generation (GPT-4o, Claude 3.5, Llama 3.1)

- Human+AI mixtures (controlled editing experiments)

- Public-domain texts (pre/post-AI augmentation)

 

Data augmentation:

- Simulate editing patterns (paraphrasing, insertion, deletion)

- Style transfer between human/AI baselines

- Temporal drift (model evolution 2023-2026)

5. Evaluation Framework

text

Primary metrics:

- AUC-ROC for binary classification

- F1-score across 4 classes (human/AI/full-edit/partial-edit)

- Mean absolute error for editing intensity estimation

 

Edge cases to test:

- Non-native English, translated texts

- Highly technical/scientific writing

- Creative writing, poetry, dialogue

- Paraphrased AI content, humanized outputs

6. Implementation Considerations

text

Technical stack:

- Python 3.11+ with transformers, scikit-learn, SHAP

- Efficient inference (<5s per 1000 words)

- API-friendly with batch processing

 

Deployment:

- Handle PDFs, EPUBs, plain text

- REST API with confidence visualization

- Integration with document management systems

7. Limitations & Ethical Analysis

text

Known failure modes:

- Advanced paraphrasing defeats statistical detectors

- Style mimicry by frontier models

- Bias against non-standard English varieties

 

Ethical concerns:

- False positives harming human authors

- Public-domain contamination assumptions

- Academic integrity weaponization risks

8. Market Context
Compare against commercial tools (GPTZero, Originality.ai, Winston AI, Copyleaks, Pangram) on:

  • Detection of partial editing (vs full generation)
  • False positive rates on human academic writing
  • Handling of long-form content (>50k words)
  • Price/performance for research use

Deliverables Required

  1. Complete algorithm pseudocode with data flow diagram
  2. Python implementation skeleton (core functions only)
  3. Evaluation protocol with synthetic test cases
  4. Tool comparison matrix (5+ commercial detectors)
  5. Research roadmap for next 12 months
  6. Failure analysis with concrete examples

Response format: Use markdown with code blocks, tables, and diagrams. Prioritize technical depth over marketing copy. Include realistic benchmarks from 2025-2026 research.


This master prompt will generate a publication-quality technical response suitable for arXiv submission or grant proposal appendix.

 

Comments