The post The 5 Biggest ‘Tells’ That Something Was Written By AI appeared on BitcoinEthereumNews.com. In brief New stylometric studies identify recurring patterns in AI prose, including predictable rhythm, uniform sentiment, and low lexical variety. A Washington Post analysis of 328,744 ChatGPT messages reveals heavy reliance on emojis, favorite words, and the cliché pivot of “Not just X, but Y.” Vocabulary tells evolve quickly, but structural habits such as symmetry, neatness, and negative parallelism persist across model generations. Is everything written by AI these days? Is this article? The proliferation of large language models has prompted a new, wary literacy: people can now read a paragraph and wonder who—or what—wrote it. That anxiety exists for good reason. Recent studies continue to show that the ever-increasing flood of machine-generated prose differs from human writing in increasingly not-so-subtle ways, from specific word choice to easily identifiable structural tics. These patterns matter because they affect far more than school essays and research theses; they shape corporate communications, journalism, and interpersonal email in ways that can muddle trust or authenticity. Researchers surveying stylometric detection techniques have found consistent, measurable patterns in lexical variety, clause structure, and function-word distributions—a statistical fingerprint that persists across tasks and prompts. While these tells are shrinking with every model generation—OpenAI just fixed its over reliance on em dashes, for instance—the difference between AI slop and stuff that’s human-written is still large enough to inform how readers and editors approach suspiciously polished text.  A recent Washington Post analysis of 328,744 ChatGPT messages reinforces this point with real-world data. It found that the model leans heavily on emojis, a narrow palette of favorite words, and everyone’s favorite tell, “negative parallelism: “It’s not X, it’s Y;” or “It’s less about X and more about Y.” The Post also warned against overconfidence: none of these traits prove AI authorship; they only raise the probability. Still, when a… The post The 5 Biggest ‘Tells’ That Something Was Written By AI appeared on BitcoinEthereumNews.com. In brief New stylometric studies identify recurring patterns in AI prose, including predictable rhythm, uniform sentiment, and low lexical variety. A Washington Post analysis of 328,744 ChatGPT messages reveals heavy reliance on emojis, favorite words, and the cliché pivot of “Not just X, but Y.” Vocabulary tells evolve quickly, but structural habits such as symmetry, neatness, and negative parallelism persist across model generations. Is everything written by AI these days? Is this article? The proliferation of large language models has prompted a new, wary literacy: people can now read a paragraph and wonder who—or what—wrote it. That anxiety exists for good reason. Recent studies continue to show that the ever-increasing flood of machine-generated prose differs from human writing in increasingly not-so-subtle ways, from specific word choice to easily identifiable structural tics. These patterns matter because they affect far more than school essays and research theses; they shape corporate communications, journalism, and interpersonal email in ways that can muddle trust or authenticity. Researchers surveying stylometric detection techniques have found consistent, measurable patterns in lexical variety, clause structure, and function-word distributions—a statistical fingerprint that persists across tasks and prompts. While these tells are shrinking with every model generation—OpenAI just fixed its over reliance on em dashes, for instance—the difference between AI slop and stuff that’s human-written is still large enough to inform how readers and editors approach suspiciously polished text.  A recent Washington Post analysis of 328,744 ChatGPT messages reinforces this point with real-world data. It found that the model leans heavily on emojis, a narrow palette of favorite words, and everyone’s favorite tell, “negative parallelism: “It’s not X, it’s Y;” or “It’s less about X and more about Y.” The Post also warned against overconfidence: none of these traits prove AI authorship; they only raise the probability. Still, when a…

The 5 Biggest ‘Tells’ That Something Was Written By AI

For feedback or concerns regarding this content, please contact us at crypto.news@mexc.com

In brief

  • New stylometric studies identify recurring patterns in AI prose, including predictable rhythm, uniform sentiment, and low lexical variety.
  • A Washington Post analysis of 328,744 ChatGPT messages reveals heavy reliance on emojis, favorite words, and the cliché pivot of “Not just X, but Y.”
  • Vocabulary tells evolve quickly, but structural habits such as symmetry, neatness, and negative parallelism persist across model generations.

Is everything written by AI these days? Is this article?

The proliferation of large language models has prompted a new, wary literacy: people can now read a paragraph and wonder who—or what—wrote it. That anxiety exists for good reason.

Recent studies continue to show that the ever-increasing flood of machine-generated prose differs from human writing in increasingly not-so-subtle ways, from specific word choice to easily identifiable structural tics. These patterns matter because they affect far more than school essays and research theses; they shape corporate communications, journalism, and interpersonal email in ways that can muddle trust or authenticity.

Researchers surveying stylometric detection techniques have found consistent, measurable patterns in lexical variety, clause structure, and function-word distributions—a statistical fingerprint that persists across tasks and prompts. While these tells are shrinking with every model generation—OpenAI just fixed its over reliance on em dashes, for instance—the difference between AI slop and stuff that’s human-written is still large enough to inform how readers and editors approach suspiciously polished text.

A recent Washington Post analysis of 328,744 ChatGPT messages reinforces this point with real-world data. It found that the model leans heavily on emojis, a narrow palette of favorite words, and everyone’s favorite tell, “negative parallelism: “It’s not X, it’s Y;” or “It’s less about X and more about Y.”

The Post also warned against overconfidence: none of these traits prove AI authorship; they only raise the probability. Still, when a piece of writing exhibits several of them, the signal gets harder to ignore.

Here are the five strongest signals that a text may have been machine-generated, each anchored in current research.

The 5 most common AI tells

  1. Negative parallelism and oversimplified contrast

    AI overuses the neat, dramatic hinge of “It’s not X, it’s Y,” and its cousin, “not just X, but Y.” These structures create the illusion of insight while supplying very little. Stylometric studies show that LLM outputs tend toward balanced, formulaic clause structures rather than the uneven, intuitive rhythms human writers use. In the Post’s dataset, variations of “not just X, but Y” alone appeared in roughly 6% of all July messages—an astonishing percentage for a single rhetorical tic.

  2. Over-neat structure and conspicuously consistent rhythm

    LLM-generated text often reads like it was written by someone who revises compulsively but never improvises. Paragraphs follow textbook patterns, transitions are frictionless, and the cadence is almost mathematically even, according to a recent analysis in Nature. Human writing—even careful writing—typically reflects digressions, interruptions, tonal shifts, and asymmetric pacing. Stylometric work comparing LLM outputs to human short stories finds that models exhibit far narrower variance in sentence length and syntactic shape.

  3. Smoothed-out emotional tone and overly courteous hedging

    AI tends to sound friendly in a way no adult actually sounds unless they work in HR or customer support. Phrases like “It’s understandable that…” or endings that gently summarize everything (“Ultimately…”) show up with unnatural regularity. Quantitative reviews of detection methods note that LLM-generated prose exhibits more uniform sentiment and fewer abrupt emotional modulations than human text.

  4. Vague abstractions and evolving “safe” vocabulary

    Models rely heavily on generic nouns—”ecosystem,” “framework,” “dynamic”—and verbs like “leverage,” “unlock,” or “navigate” when they run out of specifics. Studies consistently show lower lexical diversity and heavier nominalization in AI text. The Washington Post and Nature analyses also found that certain AI clichés aren’t static: the infamous “delve” has largely faded, replaced by new favorites like “core” and “modern.” This matters because vocabulary tells evolve quickly; structure is more reliable than any fixed word list.

  5. Balanced clauses and conspicuously careful phrasing

    LLMs love symmetry: “While X is true, Y is also important,” or “Whether you’re a beginner or an expert…” These structures feel safe because they avoid commitment. Stylometric studies show that AI text overuses certain function-word patterns and clause constructions at rates that differ sharply from human baselines. Humans tend to be either more abrupt or more discursive; machines aim for diplomatic balance every time.

By the way, most of this article was written by AI.

Generally Intelligent Newsletter

A weekly AI journey narrated by Gen, a generative AI model.

Source: https://decrypt.co/348923/5-biggest-tells-something-written-ai

Market Opportunity
null Logo
null Price(null)
--
----
USD
null (null) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact crypto.news@mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

Why African countries are using data protection laws as backdoor to regulate AI

Why African countries are using data protection laws as backdoor to regulate AI

Rather than waiting for comprehensive AI frameworks, which are often complex and slow to develop, governments across the continent are embedding AI-related rules
Share
Techcabal2026/03/19 18:46
YieldMax Funds Explained: How These ETFs Work, What They Pay & The Hidden Risks

YieldMax Funds Explained: How These ETFs Work, What They Pay & The Hidden Risks

If you have spent any time in income-investing circles recently, you have almost certainly come across YieldMax funds the ETFs promising yields of 30%, 50%, or
Share
Fintechzoom2026/03/19 18:14
Aster Price Surges After Airdrop and CZ Mention

Aster Price Surges After Airdrop and CZ Mention

The post Aster Price Surges After Airdrop and CZ Mention appeared on BitcoinEthereumNews.com. Aster, previously referred to as APX, witnessed its token price soar on September 18, rising by over 360% in one day. The surge followed after the project started its airdrop program and from CZ. What’s Driving Aster Price Surge The token’s steep price action came after the token’s airdrop began, and it will run until October 17. Approximately 704 million tokens representing approximately 8.8% of the total supply are being sent to eligible users. These include members of Aster’s Spectra Stage 0 and 1 programs, owners of Aster Gems, and traders of Aster Pro. Adding fuel to the charge, CZ publicly congratulated the Aster team, further increasing visibility to the project. That validation, combined with the token distribution, driven the price surge. Fundamentals Behind the Rally Beyond the frenzy, Aster’s fundamentals have been improving. Based on statistics provided by DeFi Llama. Its perpetual futures platform has seen more than $12 billion worth of trading volume this month, an increase from $9.78 billion in August and $8.5 billion last July. Revenue has increased steeply as well. Fees earned this quarter total $8.82 million, up from only $1.8 million during the same time last year. In Q3 2024, Aster had only generated $11,660 in revenue, but today that number is up to $5.4 million. The total value locked (TVL) in the protocol has hit a record high of $1.85 billion, an astronomical increase from $141 million in January. What’s Next for Aster Analysts believe that the rally may prevail since Aster is now becoming available on additional exchanges, yet it is mainly traded on its own platform. Yet with recipients of the airdrop likely to take profits in place, there will be some pressure selling. Like other recently listed coins like WLFI, Spark, and Avantis, a good starting run will be followed…
Share
BitcoinEthereumNews2025/09/19 08:30