by Sheri Liguori, CopySoWrite – August 9, 2024 (updated March 26, 2025)


AI Gets a C as Editor


Credit: CopySoWrite/DALL-E

With artificial intelligence (AI) on the scene, writers now have many app and chatbot choices for checking grammar, punctuation, and spelling. But when it comes to creating content, most still choose to keep it real and let their own creativity take the lead.

I’ve dipped my toe in the sea of AI-assisted writing and marveled at its ability to swiftly churn out content, help organize thoughts, and compose lists. A remarkable time-saver. But I don’t let it speak for me. I don’t use anything AI produces verbatim, even though my more cleverly crafted prompts may produce higher-quality output.

Instead, I use AI as a sounding board, inviting it to brainstorm ideas. Rather than replacing my writing, I see it as a silent teammate, enhancing my own talent.

As editor, I give AI a C because even completely unique content will not be free of all issues or errors. And when used as a prose generator, content should be carefully fact-checked and modified to reflect the writer’s voice. More on that later . . .

Prompt that produced the graphic: Please create an original abstract graphic design to signify the inner workings of your AI brain during the process of editing text. Make it look like a neural network.


Following this overview of AI chatbots as collaborative writing tools is my list of AI’s Limitations and Gaps during editing. Copilot volunteered some of these points during a general Bing search.



ChatGPT vs. Gemini – and What About Claude?

The top contenders in the chatbot arena are ChatGPT, Gemini, and Claude. My initial practice was with Copilot (formerly Microsoft’s Bing Chat—accessed from the browser’s menu). From 2021 to the end of 2024, OpenAI’s ChatGPT Large Language Model (LLM) was integrated with Copilot. Since the end of 2024, Copilot has disengaged its alliance with OpenAI and subsists on its own. Though Copilot is built into Microsoft’s applications, I’ve conversed with it less, and log directly onto ChatGPT using my Google credentials..

For a real hoot, check out this article by Ryan Morrison on Tom’s Guide: I had Copilot and ChatGPT talk to each other—it got complicated.

Google’s Gemini is powered by the Gemini LLM. Gemini and Copilot are designed to integrate with your data.


Claude (created by Anthropic) was yet another LLM that rolled out in 2023 with free and paid versions. It has recently upgraded to Claude 3.7 Sonnet. I have conversed with it and find its responses to be very similar to ChatGPT’s. See one comparison on these two.


I asked each of these LLMs: Do you have the ability to edit a story? Claude’s response was the most elaborate.


ChatGPT: Yes, I can certainly help you edit a story! Whether you need assistance with grammar, style, structure, or any other aspect of your writing, I’m here to help. Just share the text you’d like me to review, and let me know if there are any specific areas you’d like me to focus on.


GEMINI: I can collaborate with you to create all sorts of content you need, like stories. I can also translate from many languages, and I am always ready to answer your questions in an informative way. But I can't edit or modify documents directly.


CLAUDE: Yes, I can help edit a story. I can assist with various aspects of editing, including:

Proofreading for spelling, grammar, and punctuation errors
Improving sentence structure and flow
Enhancing clarity and coherence
Suggesting ways to strengthen plot, character development, or dialogue
Offering feedback on pacing and overall narrative structure
Helping with style consistency

If you'd like me to edit a story for you, you can share the text and let me know what specific areas you'd like me to focus on. I can then provide suggestions and explanations for any proposed changes.


Note: when first signing on, Claude says it “strives to do its best but is not perfect . . . may occasionally generate incorrect or misleading information, or produce offensive or biased content . . . not intended to give advice. Don’t rely on our conversation alone without doing your own independent research.” After each response, it also reminds us: “Claude can make mistakes. Please double-check responses.”

While writing this blog, a Gemini message popped up on my Android phone, inviting me to engage. I declined for now, as I don’t want AI butting in on my phone activities. I have played with Gemini on the Chrome browser only a little. For more on Gemini, a general comparison of it vs. Copilot and ChatGPT, and key differences between the pay versions, read this Popular Science article.


As of this writing, the Chatbot Arena LLM Leaderboard (user platform) ranks ChatGPT-4o #1, followed by Claude, then Gemini. Elon Musk’s Grok AI currently lags a fair distance behind. Originally, Grok had no free version, but as of February 2025, Grok 3 is free to all X users. There is still a paid "Super Grok" version for X Premium subscribers.

NOTE: Due to privacy and data protection concerns, I have not evaluated China’s DeepSeek AI (it launched for public use in January of 2025).



SLM Apps Outshine Word’s Spelling & Grammar Check

Small Language Models (SLMs) such as Grammarly, Scribbr, Wordtune, Hemingway, and others can be helpful, but understand this: they don’t catch every problem and often make inappropriate or incorrect suggestions, introducing new errors. This has been my experience and is a consensus among editors. I’ve had fun playing with the free versions of SLM apps and find it interesting when one arrives at slightly different corrective advice than another.

Microsoft Word’s grammar check function is not particularly robust, but it finds most misspelled words. Despite its limitations, Word remains the preferred writing canvas for editors working with authors due to its Track Changes function, which allows for easy tracking of edits and comments.



Copilot Q&A: The AI That Doesn’t Remember

My late 2024 conversation with CoPilot revealed how it treated new content we shared with it and its capabilities at that time (while still integrated with ChatGPT).

Copilot stated that to avoid plagiarism, it generates responses based on patterns and info it has learned but does not copy and paste from its training data. This data is not cataloged by the AI, so it only cites references for any new info it pulls from the internet. It is designed to generate original content. All the same, it’s a good idea to take a close look at critical or sensitive outputs in academic, professional, or legal contexts where plagiarism could have serious consequences. If you use an LLM for these purposes, at the very least, it may be wise to run it through a plagiarism checker (there’s a free option on Scribbr).


What ChatGPT says it’s doing with your data:


Data Retention: I don’t retain personal data from users. Each conversation is independent, and I don’t have memory of past interactions unless it’s within the same session.
Training: My training data includes information up until 2021 (update: now April 2023; and states "I can also pull in more recent information using web search if needed"). While I don’t learn from individual user interactions, feedback from users can help improve the overall system in a general sense.
User Interactions: Regular user interactions don’t directly train me. Instead, they help refine the system’s responses and improve the overall user experience.


Credit: CopySoWrite/DALL-E

The Invisible AI Writer Unmasked

According to Ethan Mollick in his book Co-Intelligence: Living and Working with AI published in 2024, “there is no way to detect whether or not a piece of text is AI-generated.” Don’t bother using an AI detector app because none of them are foolproof, and often result in false positives (pasting text from grammar apps into your work could inadvertently introduce embedded code, also potentially triggering false AI detection). Still, there are ways that editors and very astute readers can tell when AI has done some of the writing. This article “13 Ways to Detect AI Content” by Satya D (SEO specialist and Surferseo.com blog contributor) offers some insight.


What happened during pretraining or machine learning? AI was fed data from the open web and books (by computer scientists and underpaid entry-level folks), which explains how biases and misinformation could not be completely filtered out. Apparently, no authors were contacted to obtain their permission. Some publishers have since openly allowed LLMs to be trained on their books, whereas other publishers are careful to acquire signed contract clauses concerning the use of such content for LLM training—or not—with authors and agents.



AI as Editor: Limitations and Gaps

Active vs. passive voice: AI may help revise to active from passive voice if requested. It won’t identify passive voice automatically in your prose, so you may need to review each paragraph closely during each prompt response.
Anachronisms: AI is not likely to detect these—such as a person, phrase, jargon, event, object, or custom placed in the wrong time period.
Changes in font style and dash types: Once you paste text into the prompt, it converts to one font size/style, which doesn’t help with inconsistencies in your original doc. AI does not reliably point out the correct use of hyphens, em dashes, or en dashes.
Citations: Al can create citations, but they may be somewhat inaccurate. A percentage of citations might mention sources or info that doesn't exist or never happened (aka "hallucinations"—yes, Al randomly makes stuff up!).
Continuity of characters, names, places, or things: AI may miss unintended changes in character names, places, objects, and descriptions, and it doesn’t question sudden, unexplained shifts in a character’s abilities or personality.
Extra spaces and missing words: AI doesn’t consistently detect extra (unintended) spaces or catch every missing or incorrect word.
Fact checks: Al may present random hallucinations as facts. Stay vigilant and perform thorough fact checks on Al-generated prose!
Feedback and queries: A human editor provides feedback on what's working well and asks questions when something is confusing or deficient. If nonessential details bog down the narrative, an editor can suggest what you might rephrase or trim (you want to keep your readers engaged). Conversely, there may be areas where extra dialogue or description could strengthen your writing—areas that an Al didn't notice. Copilot says "Al tools can be valuable aids, but a human editor remains essential for a thorough and nuanced review of your prose."
Grammatical errors: AI will not consistently find every grammatical error. When challenged, it may concede and apologize for mistakes after multiple attempts.
Homophones: AI doesn’t routinely identify homophones (words that sound the same but are spelled differently and have different meanings); for example, “bated breath” vs. “baited breath.”
Logic and plot holes: AI cannot be trusted to highlight prose that lacks logic or find plot holes or inconsistencies in the narrative.
Structure: AI may not recognize unparallel or poorly constructed content unless specifically asked. It does not regularly create or identify a variety of sentence types and lengths.
Style: AI is unable to fully grasp tone and voice, contextual ambiguity, cultural references, biases, and subtle stylistic choices. Its prose and edits may seem too robotic or technical, lacking a writer’s uniqueness.
Summaries of long papers: AI quality with summaries is sketchy—it shortens but does not fully understand the piece enough to effectively summarize it (see this report). However, highly specific instructions in the prompt can improve results.
Word repetition and stacks: AI doesn’t look for word stacks (identical words that fall on top of each other in consecutive lines) or overused words and phrases in close proximity. It can be quite repetitive with its own word and phrase choices when generating content.


Editors’ Perspectives on AI and Editing:

Wes Cowley, a fellow freelance editor who spent many years in the IT industry writing and editing technical and business material, is firm about his position never to use LLMs in his writing or editing work. His reasons for this are fully explained on his website.

But wait, there’s more! Rachel Baron reported on AsiaEdit’s extensive testing of generative AI tools for editing work in a recent abstract article for Science Editor. For a simplified version of the main points, read Wes Cowley‘s summary.



Learning and Earning With AI

While AI may not be the perfect editor, its role in writing and learning is evolving rapidly, with implications for both education and the workplace.

Mollick, a Wharton professor specializing in management and innovation, actively incorporates AI tools into student learning. In his aforementioned book (pp. 94–96), he delves into the underlying reasons why LLMs are prone to hallucinate. That said, while some educators ban the use of AI in their students’ coursework, Mollick requires that his students use AI for various assignments. He explains that today, students are already using AI as a learning tool, and many teachers have begun to use AI in preparing for class.

On the future of AI in the classroom, Mollick says: “We’ll find a practical consensus that will allow AI to be integrated into the learning process without compromising the development of critical skills. Just as calculators did not replace the need for learning math, AI will not replace the need for learning to write and think critically . . . we must do so—it’s too late to put the genie back in the bottle.”

Even the free online education platform Khan Academy uses AI tutoring. A 7:20 video explains how chatbots and LLMs work. For the latest in classroom AI-assisted learning, watch this CBS 60 Minutes report on how Khanmigo, an AI-powered learning tool developed by Khan Academy in partnership with OpenAI, is being tested in schools.

Mollick’s research on this topic as explained in his books (some of which was performed in partnership with lead researchers from other universities) is fascinating and easy to grasp for non-tech types. He is certain that learning to work in tandem with LLMs will ultimately be beneficial, and I would add: especially with continued standards for AI safety and security. “Jobs don’t consist of one automatable task, but rather a set of complex tasks that still require human judgment,” he says. The world of work is headed in this direction, whether we like it or not.


Looking for an editor? If you’re interested in working with me, Complete my form or email me about your project. I will contact you to discuss your needs.


The Creation of My Logo:


After many hours of my prompts and OpenAI’s DALL-E iterations, I finally chose this graphic design as my CopySoWrite business logo. I am very pleased with it. One of the reasons it took so long: AI does not “see” its illustrations as we do. If I really liked an image but wanted some adjustments or changes, even though a way to re-upload the image was included in the prompt, DALL-E was incapable of editing that specific image—it may be more advanced now. DALL-E could not add text within the graphic, so I had to import the image into Canva and insert the CopySoWrite text on the logo.



Additional recommended reading:

Why great AI produces lazy humans, Ethan Mollick, Big Think, March 26, 2024.

How to succeed at the “jagged technological frontier” of AI, Kevin Dickinson, October 23, 2023.

‘Worse’ AI Counterintuitively Enhances Human Decision Making and Performance, Adam Hadhazy, July 5, 2022.