• Skip to primary navigation
  • Skip to main content
  1. The University of Mississippi

Ole Miss Engineering

University of Mississippi

  • About
    • About us
    • History
    • Mission and vision
    • K-12 Engineering Outreach
    • Honors and awards
    • Engineering Advisory and Advancement Council
  • Academics
    • Our academic programs
    • Biomedical engineering
    • Chemical engineering
    • Civil engineering
    • Computer engineering
    • Computer science
    • Electrical engineering
    • General engineering
    • Geological engineering
    • Geology
    • Mechanical engineering
    • Graduate programs
    • Accreditation
  • Future students
    • Student ambassadors
    • Engineering at Ole Miss
    • Undergraduate admissions
    • Transfer students
    • Scholarships
    • Graduate admissions
    • Degree partnerships
    • Computer requirements
    • Book a visit
  • Current students
    • Advising
    • Student organizations
    • Co-ops
    • Career planning
    • Fundamentals of Engineering exam
    • Tutoring
    • IT support
    • Resources and forms
  • Research
    • Research centers
    • Research priorities
  • People
    • Our faculty
    • Dean’s office
    • Faculty and staff resources
  • Alumni / Industry
    • Alumni Award Recipients
    • Giving
    • Industry Partner: General Atomics
    • Industry Partner: Raytheon
    • Woods Society
  • News
    • Latest news
    • Ole Miss Engineer magazine

Can artificial intelligence plagiarize?

Thai LeSince the launch of ChatGPT in November 2022, the online tool has gained a record-breaking 100 million active users. Its technology, which automatically generates text for its users based on prompts, is highly sophisticated.

But are there ethical concerns?

Ole Miss Engineering professor Thai Le has co-authored a paper, led by collaborators at Penn State University, showing that artificial intelligence-driven language models, possibly including ChatGPT, are guilty of plagiarism – in more ways than one.

“If people use this technology to write essays, grant proposals and patent applications, then we need to care about possibilities for plagiarism. We decided to investigate whether these models display plagiarism behaviors.”

– Thai Le, Assistant Professor, Computer and Information Science

Detecting plagiarism

The study, which is the first of its kind, evaluated OpenAI’s GPT-2, a precursor to ChatGPT’s current technology. They tested three separate criteria for plagiarism: direct copying of content, paraphrasing and copying ideas from text without proper attribution.

To do this, they created a method to automatically detect plagiarism and tested it against GPT-2’s training data, which is “memorized” in part and reproduced by the technology. Much of this data, which is publicly available online, is scraped from the internet without informing content owners.

By comparing 210,000 generated texts to the 8 million GPT-2 pre-training documents, the team found evidence of all three types of plagiarism in the language models they tested. Their paper explains that GPT-2 can “exploit and reuse words, sentences and even core ideas in the generated texts.”

More data, greater chance of plagiarizing

Furthermore, the team hypothesizes that the larger the model size and associated training data, the greater the possibility of plagiarism.

“People pursue large language models because the larger the model gets, generation abilities increase,” said Jooyoung Lee, first author and an information sciences and technology doctoral student at Penn State. “At the same time, they are jeopardizing the originality and creativity of the content within the training corpus. This is an important finding.”

The scientists believe that this automatic plagiarism detection method could be applied to later versions of OpenAI technology, such as those used by ChatGPT.

The research team will present their findings at the 2023 ACM Web Conference, set for April 30-May 4 in Austin, Texas.

Implications for academia

Robert Cummings, associate professor of writing and rhetoric at Ole Miss, has given advice to higher education professionals about ChatGPT’s implications in the classroom. A collaborator with Le in other AI-related research, Cummings suggests that users should be pragmatic when referencing material gained from language models.

“We have to be careful about what ideas are ours and what are borrowed,” Cummings said. “Pre-ChatGPT, I’d Google something as part of my research, and it would be sourced. If I was looking for general knowledge, I’d consult Wikipedia.

“Now, it’s important to designate what came from ChatGPT and put it off to the side as unsourced ideas.”

Le acknowledges the importance of finding solutions to these ethical issues, whether that be on the user side or on the side of scientific advancement.

“There are many important philosophical questions related to this technology,” he said. “Computer science researchers will continue to think of ways to improve these language models to change the way they generate text in such a way that they would not plagiarize.”

NSF funding
This material is based upon work supported by the National Science Foundation under Grant Nos. 1934782 and 2114824.

Ole Miss logo
Enroll at the University of Mississippi and discover your future!
Copyright © University Communications 2025. All Rights Reserved.
EEO Statement Give Us Your Feedback Accessibility Ethics Line UM Creed