“Caution: This text was written by AI.” Wouldn’t it be nice to see this disclaimer before articles or social media posts? And how many texts would turn out to be partially or completely generated by artificial intelligence?
As more and more people get access to Large Language Models (LLMs) like ChatGPT, AI texts spread everywhere. In some cases, only generated videos and images are labeled, not text. So AI content detection becomes an essential skill for search optimization, factual accuracy, and integrity of a business or academia. In this article, we’ll teach you how to detect AI writing and share 10 signs that a text might not be human-made.
TL;DR
- You should know whether the content you publish was produced with an AI, even if you don’t care about the ethics of it.
- In some cases, you can recognize AI writing even without dedicated content detection tools.
Look out for these signs of AI-written content:
- Excessive use of AI-typical words
- Repetitive language
- Lack of coherence
- Broad explanations without details
- Factual errors
- Lack of personal experience
- Inconsistent tone and style
- Mismatching purpose or search intent
- Unnatural word choices
Why should you check the written content for AI involvement?
The content marketing industry has been hugely affected by generative AI. From AI in email marketing to any other channel, the impact of artificial intelligence has been huge. Whether you reject it or embrace it in your industry, it’s useful to understand the origin of a piece of writing you work with and the degree of AI involvement in it.
You may think that AI involvement is bad for ethical reasons only. However, there are many practical situations when you may need AI checks, and real consequences you may face if you don’t do it. Here are some reasons why you should at least contemplate AI content detection.
To comply with regulations
Currently, there are no major requirements to label AI-generated texts. But seeing as Meta has introduced AI image, audio, and video labeling on its platforms, this may affect texts in the future, too. Plus, there may be industry or platform-specific rules you need to follow, so it’s best to know whether the content is human or AI-generated.
To achieve better search rankings
For marketers and SEO specialists, detecting AI writing is ever more important. Google continuously prevents spammy, low-quality content from appearing in search results. One of the company’s employees also stated that AI-generated content is considered spammy. So, despite intentionally vague official positions, the search engine will actually rank your content lower if it is suspected of being AI-generated or even deindex it altogether.
Plus, it’s harder for AI-generated content to meet Google’s E-E-A-T standards. They dictate that content should demonstrate expertise, experience, authoritativeness, and trustworthiness.
To maintain trust and transparency
This point is more about your organization’s stance on artificial intelligence. If you pride yourself, say, on your writers’ expertise and the texts turn out to be AI-generated, this makes your claim misleading. So you have to use content detection to make sure it adheres to your standards and delivers on your promises.
To combat misinformation and disinformation
Artificial intelligence is known to hallucinate — make up facts and present them as true. If you assume that an article is entirely human-written, you may end up paying less attention to it when editing or reviewing. But knowing that a piece has traces of AI’s involvement can make you more vigilant, preventing false statements.
To ensure human oversight
We’d argue every piece of writing should be read and reviewed by another person before publication, but in reality, you may not have enough resources to hire an editor or a proofreader.
If you can’t fully analyze everything you publish, an AI detection step will at least show whether you have to read one particular piece more closely or not. Then you can focus your attention on the suspicious content and not everything all at once.
To protect intellectual property
This point has two correlating issues.
First, there are many lawsuits claiming that certain AI models use copyrighted materials. If you, as a content marketer, publish something written by an AI and not a human, it can potentially have traces of those protected works. This, in turn, opens your business up to potential legal issues.
And second, if your line of work depends on confidentiality, the fact that a text is AI-generated may suggest a data breach. Essentially, it may mean that the text’s author shared your information with an AI tool, which is contrary to information security principles and NDAs.
To protect the academic standards
Last but not least, AI checks and detection are important to uphold academic integrity. They ensure the originality of research, reports, and other works and their credibility.
For many of these reasons, Selzy blog editors check each article for AI content before publication. Over a year after the ChatGPT launch, we developed a set of guidelines to navigate the increasingly murky digital content space. We use a combination of close reading and content detection tools, and many of the recommendations below are based on our experience with AI-manipulated texts.
As for the ethics of this question, we are aiming to publish human content for human benefit.
Can AI even be detected?
According to a recent study, only 51.50% of US citizens accurately identified AI-generated texts. Another one on the same subject found that when it came to AI-written texts, people had an accuracy rate of only 24%. However, the participants were probably not professionals in the content marketing or writing fields. Plus, we’d argue that with proper training in AI detection (i.e., reading this article), you can at least identify suspicious content or parts of it.
At this stage of technological development, AI models still have a recognizable style of output. It’s true that some people train their Large Language Models more extensively or use more nuanced prompts, but many still rely on “raw” Gemini or ChatGPT-generated copy. So it is possible to detect AI-written texts even without dedicated tools, at least in some cases.
10 telltale signs of AI-generated writing
In this article, we will focus on ways to detect AI content “with your bare eyes”, but you should know that the easiest approach is to simply use content detection tools. These are services catered specifically to those who want to understand the composition of a content piece and whether it was written by a human or not.
Want to learn more? Head over to the end of this article for a brief overview of the popular tools. For a comprehensive analysis, check out our list of the best AI content detectors. We tested 11 copy variations across 7 detectors in a very thorough report!
If content marketing is one of the pillars of your strategy, you should run all content through a detector. But if you only occasionally use it, subscribing to a dedicated service might be overkill. The best way to save money on content detection, though, is to manually review all articles you get and then check the most questionable ones with a tool. This way, you may even manage to squeeze into the free plan limits.
With that, let’s see what specific traits may reveal that you are looking at an AI-generated text. For this article, we prompted ChatGPT to generate text on content marketing for an email marketing blog and used the output to better illustrate each point.
Excessive use of AI typical words
Many AI-generated texts are full of the same words used in any context. Of course, some of these are perfectly normal in a human-written text, but if they are used a lot (or if all or almost all of them are present within one article, for example), you should be suspicious. Here are some of the ones we consistently find across AI-generated texts:
- Essential
- Impressive
- Robust
- Valuable
- Vital
- Boost
- Delve
- Drive
- Enhance
- Ensure
- Implement
- Leverage
- Provide
- Unleash
Overall, AI-written texts usually have generic words and phrases, lack unique vocabulary, and interesting analogies or comparisons that are common in human-written texts.
Repetitive phrasing
Another sure sign of an AI-generated text is repetitive sentence structures. In our experience, those are sentences beginning with “by doing so and so” and -ing forms in general, and sentences with “not only but also”. AI-generated texts also tend to have adjectives or other qualifiers before almost every noun in a sentence. Human-made texts, on the contrary, usually have a high variety of words and phrases.
Beyond these examples, if a text you’re reviewing has several similarly structured sentences or other instances of unjustified repetition, it might be AI-written.
Lack of coherence
On the surface, you can understand what the two paragraphs above are trying to explain: valuable content helps to establish brand authority. Try reading it more attentively, though, and you will see inconsistencies. For example, although educational and insightful (what does this even imply here, by the way?) content could be called “informative”, entertaining content cannot. Plus, does entertaining content actually “bring value” to one’s life, the kind of value you would link to authority?
It’s the same in the second paragraph: product information isn’t necessarily educational; it’s promotional. So the statistics the artificial intelligence added here don’t correlate with the point the text is trying to make.
With that, people are prone to automatically fill in the gaps of thought and interpret the text in the most coherent way possible (a smaller example of this is the ability to read a word even without vowels). But if you want to detect AI writing, you need to consciously resist this urge.
When performing AI content detection, put a vigorous editor hat on and try to really follow the logic behind each paragraph and part of the text. If you see coherence issues like the ones we described above, the text is likely generated.
Broad explanations without details
AI-written texts often lack details or in-sentence examples. The broad language and generic sentences are trademarks of generated content. These may also mean that a human author isn’t well informed on a topic, but in that case, there will probably be errors and not encyclopedia-esque definitions without much substance.
Factual errors
At the current stage of AI content generation tools, they are likely to make up facts or not fully grasp the concepts users ask them about. These will, of course, differ from text to text and will be industry and topic-specific. Sometimes, this may even result in completely false information. To illustrate, there is a subtle nuance missed in the sentence we showed above. Did you spot it?
The content inside an email doesn’t help your open rate: your subscribers actually only see a subject line and a preheader when they receive a campaign, not the entire message. Another strange error in this sentence lies in the “ultimately engage” part. Opens and clicks are engagement metrics, but the AI seems to say that there’s something else, the “ultimate” form of email engagement. And we don’t think it means conversion here. Well, ChatGPT, if you ever discover “the ultimate engagement”, do let us and the rest of the email marketing industry know about it!
If you ask AI to write about your company, it can also make sloppy mistakes. For example, ChatGPT explained that Selzy can help manage social media campaigns, which is incorrect:
Lack of personal experience
This one might be obvious, but if the AI wasn’t specifically prompted on it, a generated text wouldn’t have any examples based on lived experience or personal opinions. This, of course, may differ based on the text’s genre and style guidelines in your organization.
Personal experience can mean more than just statements like “I did that and achieved this”. AI text will probably lack sentences and paragraphs that show more than a surface-level understanding of concepts (this is especially true for practice-oriented subjects like HTML development, for example) and demonstrate that the author actually engaged in behavior, tried actions described, etc.
Another aspect of this is the lack of emotion. People are sensitive and may use explicitly negative or positive language to depict a subject they have a strong opinion about. If a text on a topic that requires emotional investment reads neutrally or even “robotically”, it’s an AI tool that could’ve written it.
Inconsistent tone and style
You might feel confident in identifying AI writing at this point, but in reality, not all texts are completely generated, which makes the detection harder. Some people may use AI for certain parts and fill in the rest themselves. We’d argue that this is most often the case.
In these situations, it’s useful to look for style and tone discrepancies. If in some parts of a text, the sentences are complex and rich with vocabulary, and in others, you see word or phrase repetition, this may also mean that it is at least partially AI-written.
Outdated content
Large Language Models (or LLMs) require training and specifically prepared data sets in order to operate. And both of these take time. That is why AI tools usually don’t have the most recent information. For example, ChatGPT only has data up until September 2021 and is aware of some “updates and trends” up to September 2023.
So one of the signs of an AI-generated text is outdated content, which is especially apparent when it comes to statistics. If a user didn’t prompt a chatbot to research something on the internet (which many are capable of), chances are, the output will rely on training data that isn’t the most recent.
In our example, the statistics ChatGPT refers to are from a Campaign Monitor guide. The guide itself doesn’t have a publication date, but one of the other numbers presented there is from 2015, so we can assume (given that this information wasn’t updated) that the guide itself was written around the same time. As a consequence, ChatGPT may have provided us with information that is almost 10 years “fresh”.
Mismatching purpose or search intent
If you are reviewing an article and it just doesn’t make sense given the brief or task description, this might be because it was AI-generated. An AI assistant can sometimes “understand” something incorrectly and provide text that doesn’t reflect its purpose or, in the case of SEO-oriented texts, search intent.
For example, when asked to provide an outline for an article on how to write effective subject lines, ChatGPT didn’t opt for a practical, strategy-oriented structure. The chatbot instead created an outline that is more so catered to an article explaining subject lines in general, not a how-to one.
Unnatural word choices
This one is a more subtle and rare sign of AI-written text. To spot generated content, be on the lookout for word pairings that do not sound human or just don’t go well together.
In our example, “stronger” probably isn’t the first word a person would come up with to describe a better email engagement. And “improving the performance of efforts” doesn’t make a ton of sense either.
Other signs
As you may have noticed in all the AI-text examples in this article, a generated text is often too complex in style and tone to the extent that it even feels academic. Sure, real human writers can produce similar texts, but AI is notorious for it, especially if it wasn’t prompted otherwise.
Another sign of AI-written content refers to its format. If a writer generated something and didn’t edit the text, it can have capitalized headings inside lists or even a totally different font and font size from the rest of the article. In our experience, parts of the text that were likely generated are in Roboto.
Top 3 AI content detectors
While this article focuses on manual detection, you might want to explore dedicated tools. We’ve gathered some of the popular options below.
| Tool | Starting price per month, paid monthly | Supported languages, besides English | Detected AI models | Features |
| GPTZero | $14.99 | French, German, Spanish, Portuguese | ChatGPT, Claude, Gemini, Llama, and more | Plagiarism checker, AI vocabulary, writing feedback |
| Copyleaks | $9.99 | 20+ languages, including French, German, Spanish, Portuguese, Chinese (traditional and simplified) | ChatGPT, Gemini, Claude, DeepSeek | Plagiarism and paraphrased AI detection, GDPR compliance, customizable sensitivity levels |
| Originality AI | $14.95 | 29 languages, including French, German, Spanish, Portuguese, Chinese (traditional and simplified) | GPT-4o, Claude 3 and ChatGPT | Plagiarism checker, fact and readability checker |
Conclusion
It seems like AI is now capable of almost anything (we even have a quiz based on this assumption!), including writing a human-sounding text. However, we still can separate an author’s creation from a chatbot’s output with a degree of accuracy. In fact, a recent test in the Spanish language proved that when it comes to creative writing, LLMs are far behind top human writers.
And though it’s getting more and more difficult to decipher AI texts from human-written ones, it is essential to maintain the quality of content and have confidence in it. We recommend you use a combination of content detection tools and manual review.












