AI Detection and Grammarly

Hi everyone, We have faculty using the AI report and are appreciating that feature. An interesting question came up recently: the report showed 64% AI-generated text, but the student claimed he had only used Grammarly, which also helps to improve phrasing. Has anyone else experienced an AI report picking up on Grammarly, and is there a way of distinguishing between a "grammar help" tool and other generative AI?

Second, will the AI report be visible to students soon? Faculty assumed that students could see it, but the report is only visible to faculty in our LMS.

Thanks,

Jennifer Douglas, American Public University System

60replies Oldest first
  • Oldest first
  • Newest first
  • Active threads
  • Popular
  • Hi Jennifer Douglas Many thanks for your feedback on our new AI Detection feature. We have received several similar reports of this behaviour and are running some tests. It might be worth us investigating this further for you. I will message you directly for more details.

    Regarding student access to the report, this is something we will review in future. 

    Like
      • Jennifer Douglas
      • Dean, Graduate Studies and Research
      • Jennifer_Douglas
      • 4 mths ago
      • Reported - view

      Gill Rowell Here is a redacted sample of the Originality report and the AI report for the student in question, who then claimed that he had only used Grammarly and not an AI chatbot tool.

      Like
    • Jennifer Douglas Grammarly employs a combination of standard correction detection, which identifies frequently misused words based on patterns, and Chat GPT-4 to enhance its speed and accuracy. Similarly, the latest version of the Windows suite has integrated AI-powered correction tools, as has added the option Google Docs. It's worth noting that I've tested content I personally wrote years before AI tools were accessible to the public, and these detection tools still identified them as AI-generated. I've also noticed most bibliographies become flagged by the detection too aswell. The detection tools might not be advancing as rapidly as AI itself and are not very reliable.

      Like
  • Our faculty have experienced a similar issue with high AI likelihood but students stating they have used Grammarly, as well as language translation tools, such as Google Translate and not tools such as Chat GPT. I would be interested to know if the Turnitin AI detection is known to identify text from these tools as AI generated also.

    Like
      • Gill Rowell
      • Customer Engagement Specialist, Turnitin
      • Gill_Rowell
      • 4 mths ago
      • Reported - view

      Narelle Hunter I believe Google Translate may also cause similar behaviour. Are you able to message me direct with paper IDs so we are able to investigate in more detail? Or alternatively you can use the in-product feedback tool.

      Like
  • Same thing here. I submitted student work and Turnitin gave it a fairly low similarity score, but 100% AI score. The student told me that he had Grammarly edit his work. 

    Unfortunately, he composed his work on Grammarly (at least, that is what he says), so there isn't any original document that was corrected by Grammarly. 

    I have assigned a few zeros to different students, and this is the only one who has come back and insisted it wasn't AI. But this is a problem. I am not going to fault a student for using Grammarly, but man there really is IMO a big difference between this and AI. Assuming that my student was telling the truth, it would be nice for turnitin to know the difference.

    I think I will put in my syllabus for future classes, something to the effect that if you do choose to use Grammarly, to write out your text ahead of time, and keep the pre-edited version. 

    Like 1
      • Jennifer Douglas
      • Dean, Graduate Studies and Research
      • Jennifer_Douglas
      • 4 mths ago
      • Reported - view

      Paul Rittman Good idea to advise students to keep their pre-edited draft and be aware of the amount of editing that grammarly is doing.

      Like
  • I just encountered the same problem. According to Turnitin, a student's essay was created by AI. The student, however, claims he used Grammarly only. This is a serious problem, needless to say. Can we know by when Turnitin will be able to distinguish Grammarly from AI? 

    Like 1
    • Jo-Anne Rondell I am suspecting what they are doing after logging in to Grammarly, is using Grammarly's AI function (called something like GrammarlyGo--you can see it on the right side of the window after you log in). The reason why I say that, is that I've taken texts that were NOT composed by AI, taken them to grammarly and ONLY made grammar suggestions by Grammarly. The edited document had a 0% or very low AI score. So for turnitin to give an AI score of 100% when the student claims they submitted it to Grammarly, tells me that they used Grammarly's AI function. See my post here: https://turnitin.forumbee.com/t/83h4qhz/grammarly-turnitin-coms-ai-score

      Like 1
    • Paul Rittman Thanks very much for the reply. Can you tell me if you tried using Grammarly's AI?

      Like
    • Jo-Anne Rondell At this point I don't remember. I've been pretty busy since that post. :(

      Like
    • Paul Rittman Thanks again. I'm pursuing this since I suspect the problems will multiply as we continue on. I'll post back here with whatever I discover, if anything.

      Like
    • Jo-Anne Rondell Sounds good. Look forward to your results.

      Like
    • Jo-Anne Rondell One more thing. If you get some concrete results about how turnitin detects Grammarly AI vs. Grammarly's normal grammar check, I would PM Gill Rowell, who has posted above here in this thread twice. She is a turnitin.com employee who has PMed me about issues I've discussed. I think she would be interested in looking at your files, if you can show how turnitin's AI generator is working.

      Like
      • Gill Rowell
      • Customer Engagement Specialist, Turnitin
      • Gill_Rowell
      • 4 mths ago
      • Reported - view

      Jo-Anne Rondell I am happy to pass on any feedback you have to our AI development team. Please DM me with any paper IDs that are exhibiting this behaviour. We are all learning about AI writing!

      Like
      • Steve Bentley
      • Turnitin Admin, University of Huddersfield, UK
      • s_d_bentley
      • 3 mths ago
      • Reported - view

      Paul Rittman I think this is part of the problem, that Grammarly has multiple services, some only available to paid accounts. Free accounts, as I understand it, offer suggestions, paid tiers will rewrite sentences to improve them. Much more of a grey area in terms of acceptability, I know our regulations wouldn't accept that from a human proofreader.

       

      GrammarlyGo as their AI offering is an extra layer on top of that.

      Like
      • lucy hunt
      • lucy_hunt
      • 2 mths ago
      • Reported - view

      Jo-Anne Rondell I have just stumbled onto this thread because I've run into this same issue. I tried to test this theory (typing something from scratch in Grammarly, downloading it, and then submitting it in canvas), but I guess because I was in "student view" and not an actual student, the AI detection was unable to be processed. Literally, the box was gray and it never registered. I did this twice. I have since resorted to asking a student I know and trust (who is in my class) to test this as well so I can see if it is in fact Grammarly that's being detected. If so, the two essays written by another student that were detected to have been generated by AI (100%) will be graded as normal. I'm hoping Turnitin's AI detector will be able to discern Grammarly from the other AI generators, but who knows. Should we discourage our students from starting from scratch in Grammarly? 

      Like
      • lucy hunt
      • lucy_hunt
      • 2 mths ago
      • Reported - view

      Paul Rittman So, I figured the same thing (with GrammarlyGO), and I went to my Grammarly account to check. Our institution subscribes to Grammarly premium, so when I clicked on "features," the GrammarlyGO was disabled by admin. So if it's disabled on my end, it would have to be disabled on my students' Grammarlys right? That would mean that the AI detector is literally just detecting essays started and completed in Grammarly (but not using GrammarlyGO) as AI-generated. 

      Like
  • So interesting to read this thread.  I had a student yesterday state that he only used Grammarly on an assignment that came up with a 73% AI score.  He also sent me a list of news articles about students who have been falsely accused of cheating based on Turnitin AI software detection.  I would be interested in any progress you make Gill Rowell in working out if it's picking up Grammerly or Grammerly AI (I agree with Paul Rittman - I think it will be the latter.  

    Like
  • ok and so what

    Like
  • Hi all, thank you for this forum. I have run into a similar predicament with a few of my students showing up with high AI scores. One of my students who I have personally seen writing during my office hours and asking questions about their essay turned in a paper with 90% AI detection score. After talking with them, she mentioned that she uses Grammarly for all of her writing and that this may be the cause. I would love to stay updated Gill Rowell about the prevalence of this situation. As it stands right now, I will not be giving any zeros based on AI scores until it is clear that Grammarly does not affect the report.

    Like
  • Amy Henderson and Fiona Heagerty I will update you as soon as I have further news on this. In the meantime, if you have specific papers which are demonstrating this behaviour, please message me directly with details and I will pass on to the AI development team to take a look. 

    Like
  • We are finding the same thing - a student swears he only used Grammarly, but that comes up as 55% AI use. It would be great if the TurnItIn could give us clarity on whether this is even likely that their detector will detect Grammarly.

    Like 1
      • Steve Bentley
      • Turnitin Admin, University of Huddersfield, UK
      • s_d_bentley
      • 3 mths ago
      • Reported - view

      Annelize McKay It's worth confirming precisely which part of Grammarly the student used - if it was GrammarlyGo, that's their answer to ChatGPT.

      If a student uses Grammarly's rewriting features (in the paid for version) excessively then it's not inconceivable that the output could look a lot like that from ChatGPT, so any detection mechanism is going to have trouble telling them apart.

      Like
    • Steve Bentley Thank you - this is most useful. I did not know that two versions of Grammarly existed so I certainly will look into that.

      Like
  • I also hear from students that they used Grammarly.  But I am wondering if they are using GrammarlyGo.  I did hear that GrammarlyGo uses GPT3 language so that is what ChatGPT is based off of, so would Turnitin AI then detect GrammarlyGo?

    Like
    • Patti Meyer Follow-up, created a paper in GrammarlyGo and ran it through Turnitin.  Came back 100% AI.  And 16% Similarity score stating they could not show me where the text came from because it was "outside the host institution." 

      Like
  • Our institution uses Grammarly premium and the GrammarlyGo is disabled, so we can’t use it. I conducted three different tests to find out if Turnitin’s AI detector is in fact flagging Grammarly. Based on my findings, Grammarly isn’t the issue. I typed an essay from scratch in Grammarly, accepted the changes/edits, and downloaded it and turned it in to canvas. 0% AI detection. I then took a different essay I typed in Word, copied it into Grammarly, accepted the changes and edits and downloaded and turned it in. 0% AI detection. THEN I copied an essay from chatgpt, pasted it into Grammarly, accepted any changes, downloaded it and then turned it into canvas. 100% AI detection. My findings show that it’s not Grammarly. I’ve only gotten this far, but I mean it’s something to note.  

    Like
  • Our detector is not tuned to target Grammarly generated content but rather, other AI content written by LLMs such as GPT-3.5.  Based on initial tests we conducted on human-written documents with no AI-generated content in them, in most cases, changes made by Grammarly and/or other grammar-checking tools were not flagged as AI-written by our detector.

    Like
      • Steve Bentley
      • Turnitin Admin, University of Huddersfield, UK
      • s_d_bentley
      • 2 mths ago
      • Reported - view

      Purna Bose Thanks for that clarification. Can I confirm that your testing includes both the free tier of Grammarly and the paid-for product that has a "rewrite my sentence" feature?  Have you done similar testing with Quillbot and its paraphrasing capability?

      Like
      • Gill Rowell
      • Customer Engagement Specialist, Turnitin
      • Gill_Rowell
      • 2 mths ago
      • Reported - view

      Steve Bentley Yes, our testing has largely been based on the free version of Grammarly, So if the detector is flagging content as AI it may have been written using GrammarlyGo. Our investigations on Quillbot are ongoing. 

      Like
      • Steve Bentley
      • Turnitin Admin, University of Huddersfield, UK
      • s_d_bentley
      • 2 mths ago
      • Reported - view

      Gill Rowell Hi Gill, I think there's a bit of confusion here:  Grammarly has THREE services.

      Free: Basic flagging of errors for the user to fix themselves
      Paid: Has tools that can rewrite sentences/paragraphs to improve the English
      GO: This is a ChatGPT clone/rebadge

      It's the paid platform where there is ambiguity, whether the rewritten sentences/paragraphs look sufficiently like AI output that it is misdetected, which is what students are claiming.  (Obviously different institutions will have their own policy on whether that rewriting is acceptable or not)

      Like
      • lucy hunt
      • lucy_hunt
      • 2 mths ago
      • Reported - view

      Steve Bentley  I used the paid version of Grammarly for my own personal test. I accepted all the changes and made edits according to Grammarly's suggestions but still was not flagged for any AI use. I even rewrote/reworded the sentences it suggested I improve. I'm wondering how these children are claiming it's Grammarly when, on our end, GrammarlyGO is disabled and my own personal test proved the paid version didn't detect any AI. 

      Like
      • lucy hunt
      • lucy_hunt
      • 2 mths ago
      • Reported - view

      Steve Bentley All of my writing preferences are set to "on," too--I'm getting the full dose of what Grammarly Premium offers minus GrammarlyGo. The same goes for my students. 

      Like
      • Malik
      • Malik.1
      • 1 mth ago
      • 1
      • Reported - view

      Gill Rowell Could you please update about the Quillbot investigation?

      Like 1
      • Gill Rowell
      • Customer Engagement Specialist, Turnitin
      • Gill_Rowell
      • 1 mth ago
      • 1
      • Reported - view

      Malik I don't have any specific updates on Quillbot. In our AI Innovation Lab, we have conducted tests using open sourced paraphrasing tools (including different LLMs) and in most cases, our detector has retained its effectiveness and is able to identify text as AI-generated even when a paraphrasing tool has been used to change the AI output.

      Like 1
      • Malik
      • Malik.1
      • 1 mth ago
      • Reported - view

      Gill Rowell Thank you for your response. Actually, I wanted to confirm that if an AI paraphrasing tool is used on human generated content (to sound smarter or improve structure etc) does turnitin flag it as AI too? 

      Like
      • Steve Bentley
      • Turnitin Admin, University of Huddersfield, UK
      • s_d_bentley
      • 1 mth ago
      • Reported - view

      Hi Gill Rowell just to support Malik's comment, the concern is about avoiding false positive detections where tools like Quillbot and Grammarly have been used to improve a student's authentic writing. This is a common claim from students who are challenged about a high AI score from Turnitin.

      Like
      • Gill Rowell
      • Customer Engagement Specialist, Turnitin
      • Gill_Rowell
      • 1 mth ago
      • Reported - view

      I understand that there may be times when using these tools may trigger the AI indicator. You may find some of our resources on false positives useful. Firstly this blog post and video from our Product team. Additionally, some of our pedagogic resources from my colleagues on the TLI team. Using the AI detector, just like the similarity report, is just one indicator of possible misconduct and may be used a starting point for conversations with students about their use of these tools when completing their assessed work. 

      Additionally you can always ask our Support team to take a look at any paper where there may be doubt. You can contact them by emailing tiisupport@turnitin.com.

      Like
  • If an essay is flagged with 50% or over for AI, wouldn't that pretty much toss out the "false positive" theory? I get 20% or less can be less reliable, but 50%? 

    Like
      • Gill Rowell
      • Customer Engagement Specialist, Turnitin
      • Gill_Rowell
      • 2 mths ago
      • Reported - view

      lucy hunt I'm not sure if this recent blog post from our Chief Product Officer, Annie Chechitelli might help with your understanding of false positives. 

      Like
      • TWC
      • TWC
      • 2 mths ago
      • Reported - view

      lucy hunt the % that gets flagged is only the chance that the text is AI written. Just because they are scoring some % AI does not mean they are using AI. The score is the % chance detectors think the article was written by AI not what % of the article was written with AI.

      Like
      • lucy hunt
      • lucy_hunt
      • 2 mths ago
      • Reported - view

      Chris Turner isn't 75 to 100% still pretty significant?

      Like
      • lucy hunt
      • lucy_hunt
      • 2 mths ago
      • Reported - view

      Chris Turner This is what Turnitin has on its page: "Using the average scores of all the segments within the document, the model then generates an overall prediction of how much text in the submission we believe has been generated by AI."

      Like
      • TWC
      • TWC
      • 2 mths ago
      • Reported - view

        lucy hunt 

      AI vs Human score says the probability that AI thinks the content was generated by AI or generated by a person. A score of 20% AI 80% Human Means – The AI gives the content 80% probability of being Human and  20% probability of being AI generated; it doesn't mean – 80% of the article is human, and 20% is AI. My neighbour is the CTO of a large AI company and says that AI detection is not at a point where it can provide enough certainty and gives a high rate of false positives. I have run through pages of engineering textbooks through various AI detection including turnitin and each time the score had between 10- 20% approx AI content. This has to be incorrect as all the text I ran through detection was years old, well before AI writing tools came on the market. I'm interested to see how the detection tools develop, but they appear too inaccurate. 

      Like
      • lucy hunt
      • lucy_hunt
      • 2 mths ago
      • Reported - view

      TWC I understand, but I don’t count anything that’s been flagged under 30% from Turnitin. If a student turns in an essay and it comes back with a  70-100% AI detection, I ask them to redo it. If it came back with 30% or less, I ignore it.  

      Like
  • Also don't forget there is Grammarly premium for desktop and browser use. Many folks use this application or similar writing tools if they have dyslexia or chronic fatigue that can affect concentration.  Grammarly premium significantly changes sentence structure or whole paragraphs. 

    Like 1
  • It helps if you think of the percentage as only one data point to have a discussion with students; ideally, that would take place during drafting, rather than submission. Two resources I'd suggest for this are this one on approaching a student when you think AI misuse has occurred or discussion starters for that situation. 

    Like
      • Steve Bentley
      • Turnitin Admin, University of Huddersfield, UK
      • s_d_bentley
      • 2 mths ago
      • Reported - view

      Karen Smith We're taking a view that the AI report alone is not sufficient to have that conversation. A "magic number" and no way to explain how that number has been reached is not a good basis to be making accusations - however gently they are framed. 

      Our approach is to require some additional evidence, such as falsified references or differences in writing style or standard of English compared with previous work before we act on the report, mainly because we don't want to damage staff-student relationships with this kind of conversation based on something that we can't back up and where it's the student's word against the Turnitin algorithm.

      Like
      • Karen Smithmoderator
      • Senior Teaching & Learning Innovations Specialist
      • Karen_Smith
      • 2 mths ago
      • Reported - view

      Steve Bentley Yes, exactly. Many of our resources speak to the need for other data points. If I were still in the classroom, that number would nudge me to look more closely and have the conversations if it seems warranted. 

      Like
  • a foremost authority on AI detection is Atheer Mahir. He has a copious thread on this subject on his Linkedin profile

    Like
  • I have just joined this network and have reviewed the posts below. I have had some instance where TII has flagged students as having 50% AI generated content. However, some students are adamant they did not use AI or Grammarly/Quillbot. Are they just lying to me or are there any other possibilities as to why their work might be flagged? I had a student initially get flagged at 58%. They made numerous edits to the essay (as seen via google docs) and the resubmission still got flagged at 60%. I am having trouble understanding what might be the issue and would love some guidance. 

    Like
      • Gill Rowell
      • Customer Engagement Specialist, Turnitin
      • Gill_Rowell
      • 1 mth ago
      • Reported - view

      David Silver Thank you for joining us! I think there are a few things to unpack here. Firstly, it's important to say that the AI detection indicator, just like the Similarity score is just one part of the conversation regarding possible misconduct.

      It seems like you have already engaged in a conversation with the student regarding their work. And I would say that if the student is adamant that this is their own work it might be worth asking our Support team to take a look at the paper. You can contact them by emailing tiisupport@turnitin.com.

      Additionally you may find some of our resources on false positives useful. Firstly this blog post and video from our Product team. Additionally some of our pedagogic resources from my colleagues on the TLI team. 

      Like
  • I have a similar case with a student who claims to have translated their work from their native language into English.  Is this likely to have produced a high indicator? ChatGPT does a great job of translation...better than google translate

    Like
      • Steve Bentley
      • Turnitin Admin, University of Huddersfield, UK
      • s_d_bentley
      • 1 mth ago
      • Reported - view

      Esther Jubb Potentially yes, but who knows as the detection tool doesn't show its working.

      Like
  • Two of my own papers in progress tested positive for AI even though I used the basic Grammarly functions in editing the drafts. I did not use Grammarly's AI function. I'm not sure how I can trust Turnitin when it shows a false positive for me, much less send the paper out for review if it's going to come back as a false positive at a journal. This needs to be fixed. I certainly cannot use this to give a student a failing grade if it's going to open me to any legal repercussions.

    Like
      • Karen Smithmoderator
      • Senior Teaching & Learning Innovations Specialist
      • Karen_Smith
      • 1 mth ago
      • Reported - view

      John Conner I recommend the following info on false positives: The first blog from March and the second blog from June of this year.

      As for the student use, I would say to use the score as a discussion starter with students: How did they use the tool and why? Did it fit into the parameters of the paper, etc. From there it's easier to discuss what consequences are appropriate. Which may be a rewrite or it may require some disciplinary measures. But it's only a starting point, not a definitive or solo data point.

      Like
  • I don’t even glance at anything that is UNDER 50% AI detected. If it’s OVER 50%, they have to redo it. 

    Like
Like2 Follow
  • 2 Likes
  • 1 mth agoLast active
  • 60Replies
  • 2023Views
  • 26 Following