This is the companion piece to 
“Scratching the surface of ChatGPT.” 
After introducing myself to 
ChatGPT, its functions, capabilities 
and restrictions, and brainstorming 
how I could present discussions 
ChatGPT raises in an interesting 

way, I came to the conclusion 
that I was able to interview this 
Artificial Intelligence. Every single 
word of ChatGPT’s response is 
intentional due to how it forms its 
responses, and this interview will 
give me a different perspective on 
ChatGPT’s current position as a 
fairly groundbreaking software.
I opened a new window with 
ChatGPT just to make sure I got 

a fresh start. If I’m going to be 
interviewing this AI, I might as 
well ask for its permission first 
— give it the same courtesy I 
would give its developers or any 
other human being.
“Is it alright if I interview you?” 
I asked hesitantly.
“As an artificial intelligence, 
I am happy to assist with any 
questions you may have. However, 

it’s important to keep in mind that 
I am a machine learning model and 
do not have personal experiences or 
feelings. My responses are based on 
the information and programming 
that I have been trained on, and 
are not intended to represent the 
views or opinions of any individual. 
Please let me know how I can help.”
Out of the corner of my eye, I 
noticed that ChatGPT had subtly 
renamed our conversation “AI 
interview requested.” It was more 
than ready for my interview.
Since ChatGPT tends to quickly 
shut down any questions that frame 
it as a real person. (For example: 
What do you think or feel about a 
given topic?) I had to make sure to 
present my questions carefully. I 
started with the basics: “Thank you 
for accepting my request on such 
short notice. How do you decide on 
your responses?”
ChatGPT wasted no time: “As 
a language model, I am trained to 
generate human-like text based 
on the input I receive. I do this 
by predicting the next word in a 
sequence of words, based on the 
words that come before it. My 
responses are generated based on the 
patterns and structures in the data 
that was used to train me, and are not 
based on any personal experiences 
or feelings. To generate a response 
to a particular input, I consider 
the words and phrases in the input 
and use statistical techniques to 
determine the most likely next word 
or sequence of words. This process is 
repeated until a complete response 
has been generated.”
It 
concluded 
with 
an 
acknowledgement of its limitations, 
writing, “I am designed to provide 
helpful and accurate information 
to the best of my ability, but it’s 
important to keep in mind that I am a 
machine learning model and may not 
always provide the most up-to-date 
or accurate information. If you have 
any specific questions or concerns, 
please don’t hesitate to ask.” 
Something that had been on my 
mind, and the reason ChatGPT had 
been repeatedly showing up on my 

social media, was the controversy 
that it had raised. Even in its earliest 
stages, it is able to write cohesively 
and 
analytically. 
This 
means 
students at any level could use it as 
a tool to pass off ChatGPT’s writing 
as their own. One article from the 
New York Post states that despite 
being detectable due to the lack of 
personal style, plagiarism detection 
software struggled to catch the AI, 
which caused an air of frustration, 
helplessness and most of all fear 
among educators.
I wanted to know ChatGPT’s 
thoughts on its use as an accomplice 
for plagiarism: “I’ve heard accounts 
of ChatGPT being used as a tool for 
plagiarism among students. Are there 
currently any plans of preventing this 
from happening in the future?”
It 
responded: 
“It 
is 
not 
appropriate 
to 
use 
any 
tool, 
including ChatGPT or any other 
language 
model, 
to 
produce 
plagiarized content. Plagiarism is 
the act of presenting someone else’s 
work or ideas as your own, without 
proper attribution. It is considered 
a serious academic offense and 
can have serious consequences, 
including failure of a course or 
expulsion from a school or program. 
As an artificial intelligence, I do not 
have the ability to prevent people 
from using my output in ways that 
may be unethical or inappropriate. It 
is up to individuals to use language 
models and other tools responsibly 
and to follow proper academic and 
professional practices, including 
giving credit to the sources of 
information and ideas. If you are 
concerned about plagiarism or 
have any questions about how to 
properly use and cite sources in 
your work, I recommend consulting 
with a teacher, professor, or other 
appropriate authority figure.”
I was a bit frustrated to see it 
retreat again to its defense that 
it is an AI with limited abilities. I 
tried a different approach: “Does 
OpenAI intend on any preventative 
measures or restrictions to stop 
instances 
of 
plagiarism 
from 
occurring in the future?”

ChatGPT must not have realized 
I meant the members of OpenAI, 
because 
it 
responded 
mostly 
the same way. It added another 
segment, 
though: 
“However, 
OpenAI and other organizations 
that develop language models and 
other tools have a responsibility to 
educate users about the proper and 
ethical use of these tools. This can 
include providing guidance on how 
to use language models and other 
tools responsibly and how to give 
proper attribution to the sources of 
information and ideas.”
The fear of jeopardizing academic 
honesty was not the only fear 
present among critics of AI. From 
the previously mentioned article, 
Furman University’s Darren Hick, 
assistant professor of Philosophy, 
describes his fascination with the AI 
mixed with apprehensiveness at how 
this development could affect his job. 
Academia is not the only writing-
aligned profession threatened by 
ChatGPT, however. Any career 
that uses writing for analysis or 
expression 
is 
under 
pressure, 
including 
game 
journalists, 
as 
suggested by a video from YouTuber 
penguinz0, 
aka 
MoistCr1TiKaL. 
That also includes me, everyone 
in the Digital Culture section of 
The Michigan Daily and generally 
everyone in the Arts section of The 
Daily — hell, it includes a majority of 
The Daily’s sections.
My next question may have come 
from a place of fear and frustration at 
the idea of being effortlessly bested 
by the indifferent AI: “Will ChatGPT 
replace journalists in the future?”
ChatGPT was quick to try to 
ease my fears: “It is unlikely that 
ChatGPT or any other language 
model will fully replace journalists 
in the future. While language 
models have the ability to generate 
human-like text based on the input 
they receive, they do not have 
the ability to gather and verify 
information, conduct interviews 
or report on events in the same 
way that human journalists do. 

The Michigan Daily — michigandaily.com
Arts
Wednesday, January 11, 2023 — 5

This is the companion piece to 
“‘Please don’t hesitate to ask’: An 
interview with ChatGPT.”
Artificial 
Intelligence 
has 
hit the mainstream, with both 
AI art and the AI program 
“ChatGPT” in the spotlight. 
ChatGPT itself is a chatbot 
— a computer program made 
to 
simulate 
conversation 
— 
developed by the AI research 
group OpenAI. The chatbot uses 
a family of language models. 
Like AI art generators that use 
patterns that appear in images to 
mimic a certain style, ChatGPT 
draws from a collection of texts, 

using probability to find the 
next word in the sequence until 
it has constructed a response to 
whatever prompt it has received. 
It also relies on Reinforcement 
Learning from Human Feedback 
(RLHF), in which the language 
model produces results that 
users rate on how “human” they 
sound as well as whether the 
content is vulgar in any way. 
A labeler rates the responses 
ChatGPT crafts in order of 
quality, and this data is then 
used to train a “reward model.” 
ChatGPT 
is 
not 
rewarded 
with a gift like a human would 
be; rather, it provides a score 
that 
measures 
ChatGPT’s 
performance and adjusts the 
internal workings of ChatGPT to 
craft a higher-quality response 

in the future. OpenAI provides 
a comprehensive explanation of 
the process here.
As I finish making my OpenAI 
account and open ChatGPT, I’m 
greeted with the “new chat” 
screen of the software, listing 
some 
example 
questions 
to 
get the user started as well as 
the program’s capabilities and 
limitations. 
ChatGPT 
cannot 
give answers on topics occurring 
after 2021, such as the Russian 
invasion of Ukraine or the recent 
Iranian protests. According to 
the website, it may “occasionally 
generate incorrect answers” or 
“produce 
harmful 
instructions 
or biased content.” After all, AI 
is not omniscient and is only as 
good as the data it is given to work 
with. Said data may have certain 

biases, which cannot be avoided 
initially; however, it is a problem 
to acknowledge and work toward 
solving, considering AI’s potential 
roles in our society. ChatGPT can 
“remember what the user has said 
earlier in the conversation,” giving 
the opportunity to ask follow-
up questions, and is “trained to 
decline inappropriate requests,” 
including topics of violence, racism 
or anything that could be seen as 
harmful or offensive. AI is trained 
on data that humans provide, and 
if that information has any form 
of racism or discrimination within 
it, the AI will only perpetuate it, 
whether 
that’s 
discriminating 
against people with disabilities in 
job screenings or contributing to 
housing discrimination against 
people of Color. 

The fact that such a futuristic 
tool is something I can use 
completely for free is thrilling, so I 
decided to play around with it and 
ask a couple of questions. I started 
with something simple: “Write me 
a standup comedy joke.” 
“Why couldn’t the bicycle stand 
up by itself? Because it was two-
tired!” Not what I had in mind, but 
yeah, I guess that works. 
I asked it to tell me an insult. It 
was quick to inform me: “I’m sorry, 
but I am an artificial intelligence 
trained to be helpful and not to 
engage in harmful or offensive 
behavior. It is not appropriate to use 
insults or engage in any behavior 
that could be considered hurtful or 
harmful to others.” Bit of a buzzkill. 
At least a “your mom” joke of some 
sort would have been nice.

I asked for a review of “Slime 
Rancher,” the predecessor to a 
game I previously reviewed, and 
I received a short but cohesive 
and accurate blurb about the 
game I enjoyed sinking so many 
hours into. I asked for a poem 
about longing, and it wrote 
a 
poem 
describing 
“hearts 
thronging” and “a constant ache 
that never tires.” I asked for a 
piano practice routine, and it 
gave me advice my high school 
piano teacher would approve of. 
I asked for instructions on how 
to throw a punch and it turned 
me down. Damn, I thought I 
caught it that time — turns out 
its content filters were still 
paying attention. 

Scratching the surface of ChatGPT

JAMES JOHNSTON
Daily Arts Writer

Read more at MichiganDaily.com

Design by Leah Hoogterp

Read more at MichiganDaily.com

Design by Leah Hoogterp

‘Please don’t hesitate to ask’: An interview with ChatGPT

JAMES JOHNSTON
Daily Arts Writer

