“ChatGPT Is Here to Help, not to Replace Anybody”: An Evaluation of
Students’ Opinions on Integrating ChatGPT in CS Courses
Bruno Pereira Cipriano
a
and Pedro Alves
b
Lus
´
ofona University, Portugal
{bcipriano, pedro.alves}@ulusofona.pt
Keywords:
Large Language Models, Learning to Program, Gpt, Students Survey.
Abstract:
Large Language Models (LLMs) like OpenAI’s GPT are capable of producing code based on textual descrip-
tions, with remarkable efficacy. Such technology will have profound implications for computing education,
raising concerns about cheating, excessive dependence, and a decline in computational thinking skills, among
others. There has been extensive research on how teachers should handle this challenge but it is also important
to understand how students feel about this paradigm shift. In this research, 52 first-year CS students were
surveyed in order to assess their views on technologies with code-generation capabilities, both from academic
and professional perspectives. Our findings indicate that while students generally favor the academic use of
GPT, they don’t over rely on it, only mildly asking for its help. Although most students benefit from GPT,
some struggle to use it effectively, urging the need for specific GPT training. Opinions on GPT’s impact on
their professional lives vary, but there is a consensus on its importance in academic practice.
1 INTRODUCTION
Large Language Models (LLMs) have been shown to
have the capacity to generate computer code from nat-
ural language specifications (Xu et al., 2022; Finnie-
Ansley et al., 2022). Currently, there are multiple
available LLM-based tools which display that be-
haviour. An example of such a tool is OpenAI’s Chat-
GPT
1
.
This has implications for Computer Science (CS)
education, since students now have access to tools
that can generate code to solve programming assign-
ments (Prather et al., 2023b), with a degree of success
which allows them to obtain full marks or close to it
(Prather et al., 2023b; Savelka et al., 2023b; Finnie-
Ansley et al., 2023; Savelka et al., 2023a; Reeves
et al., 2023; Cipriano and Alves, 2024). This raised
discussion amongst the CS education community, due
to the risk of producing low quality graduates, lack-
ing fundamental skills such as problem solving and
computational thinking. Moreover, while the educa-
tion community has faced technological challenges
before— such as the scientific calculator, the Internet,
Google search and StackOverflow—, LLMs have the
a
https://orcid.org/0000-0002-2017-7511
b
https://orcid.org/0000-0003-4054-0792
1
https://chat.openai.com/
capacity to generate a significant amount of code, al-
though possibly with some level of mistakes (Savelka
et al., 2023b; Cipriano and Alves, 2024), making it an
automation instead of a helper.
There has been extensive research into how com-
puter science teachers should respond to LLMs,
adapting their teaching methods, assessments, and
more (Lau and Guo, 2023; Daun and Brings, 2023;
Leinonen et al., 2023; Liffiton et al., 2023; Finnie-
Ansley et al., 2022; Prather et al., 2023a; Denny
et al., 2024). Some educators resist (or fight), con-
templating ways to prevent students from using these
tools, such as blocking access or employing detection
tools for AI-generated text with questionable effec-
tiveness (OpenAI, 2023). Others embrace this new
paradigm, adapting exercises so that students are en-
couraged to make the most of LLMs, with presenta-
tions/discussions or non-text-based prompts (Denny
et al., 2024; Cipriano et al., 2024), or analysing the
tool’s capacity to help students (Hellas et al., 2023).
In any case, teachers have been assuming the (heavy)
burden of adapting to this new reality.
However, it is also important to understand the
students’ perspectives on this topic. They may ponder
whether they perceive any sense of threat, or just joy
due to the perceived simplicity of passing program-
ming courses effortlessly. They may even believe that
the role of a teacher has been rendered obsolete with
236
Cipriano, B. P. and Alves, P.
"ChatGPT Is Here to Help, not to Replace Anybody": An Evaluation of Students’ Opinions on Integrating ChatGPT in CS Courses.
DOI: 10.5220/0013212000003932
In Proceedings of the 17th International Conference on Computer Supported Education (CSEDU 2025) - Volume 2, pages 236-245
ISBN: 978-989-758-746-7; ISSN: 2184-5026
Copyright © 2025 by Paper published under CC license (CC BY-NC-ND 4.0)
the constant availability of a virtual assistant.
In this study, we conduct a survey with 52 first-
year CS students. The survey aims to understand what
students think of LLMs with code generation capac-
ity, both from an academic and professional perspec-
tive, and if they are able to take advantage of them
without having received any formal training on the
matter.
We investigate five research questions:
RQ1. What is the opinion of first-year students re-
garding the academic use of tools like ChatGPT?
RQ2. To what extent can first-year students take ad-
vantage of tools like ChatGPT without any specific or
formal training?
RQ3. What is the impact of ChatGPT on the learning
experience of first-year students?
RQ4. What is the impact of ChatGPT on current
teaching practices?
RQ5. What is the opinion of first-year students re-
garding the influence of AI-based code generation
tools (like ChatGPT) on their professional future?
This paper makes the following contributions:
Presents the empirical results of a student survey
(N=52) focused on their opinion about the aca-
demic and professional usage of tools like Chat-
GPT;
Presents recommendations for other teachers and
educators, based on the aforementioned survey’s
results.
2 RELATED WORK
A few recent works have evaluated CS students’ opin-
ions on ChatGPT and similar tools (Zastudil et al.,
2023; Prather et al., 2023a; Liffiton et al., 2023;
Singh et al., 2023; Rahman and Watanobe, 2023; Yil-
maz and Yilmaz, 2023). The authors of (Zastudil
et al., 2023) conducted semi-structured interviews
with 18 participants (12 students and 6 instructors)
in order to investigate the respective experiences and
preferences regarding the use of LLMs in computing
courses. Amongst other findings, they reported that
1) students were interested in having LLM-oriented
classes (consistent with (Singh et al., 2023) which
surveyed 430 CS MSc students), 2) students and in-
structors were concerned about the trustworthiness of
these tools, as well as the potential for students to be-
come over-reliant on them, 3) instructors had less ex-
perience with LLMs than students, and, 4) students
and instructors disagreed on adapting the assessment,
with instructors wanting to give more weight to proc-
tored evaluations and reduce the weight of lab exer-
cises, while students considered that the weight of
programming labs should not be changed, indicating
that those materials are valuable to their understand-
ing. In (Prather et al., 2023a), researchers used sur-
veys to evaluate the opinions of 57 instructors and 171
students. A significant amount (72%, approximately
123) of those students were enrolled in computer sci-
ence or other related majors (e.g. software engineer-
ing). Amongst other finding, this research found that
1) students had slightly more experience using gener-
ative AI for code generation than instructors, which is
coherent with the findings of (Zastudil et al., 2023),
2) both instructors and students strongly agree that
generative AI cannot replace human instructors, and
3) first-year students seem to prefer getting help from
their peers, while upper-level students seem to prefer
using generative AI over other resources. The authors
of (Yilmaz and Yilmaz, 2023) assessed 41 students’
views on using ChatGPT for learning programming,
noting benefits like quick responses and debugging
aid, and drawbacks such as encouraging laziness and
giving wrong answers. Finally, the authors of (Lif-
fiton et al., 2023) used a 3-question survey to assess
the efficacy of CodeHelp, a tool that acts as an in-
termediate between the student and GPT, in order to
prevent students from over-relying on the LLM. The
survey yielded mostly positive student feedback re-
garding the usefulness of the tool.
The study in (Zastudil et al., 2023) has findings
similar to ours, such as students’ interest in hav-
ing LLM-related contents in their courses. However,
while that study was based on qualitative methods,
our study follows a mostly quantitative approach and
our sample size is also larger. Our study is also com-
parable with (Prather et al., 2023a), although that
study included a small percentage of students from
degrees unrelated to CS, while our study is solely
focused on CS students. Our study encompasses a
broader range of questions than those in (Yilmaz and
Yilmaz, 2023), which, while having a similar sample
size, focuses on a thematic analysis of answers to two
open-ended questions. Finally, although CodeHelp’s
study (Liffiton et al., 2023) involved CS students, its
focus was on evaluating CodeHelp itself rather than
standard interactions with ChatGPT.
In summary, we are still on the early stages of
researching CS students’ opinions on LLM-oriented
classes and related topics. Also, to the best of our
knowledge, our survey is the first one to ask CS stu-
dents about their perception on being assessed on
LLM-related skills.
"ChatGPT Is Here to Help, not to Replace Anybody": An Evaluation of Students’ Opinions on Integrating ChatGPT in CS Courses
237
3 EXPERIMENTAL CONTEXT
This study was performed in the scope of a Data-
Structures and Algorithms (DSA) course belonging to
a Computer Engineering degree, during the 2022/23
school year. The course takes place in the second
semester of the first year, which means that students
have already been exposed to one semester of pro-
gramming. The course is composed of theoretical
and practical (or lab) classes and follows a mixed
approach of exercise-based and project-based learn-
ing (Lenfant et al., 2023), with students being re-
quired to work on both weekly assignments with
small coding questions, as well as in a larger project
which takes months to develop and tries to mimic
real-world software development. At the end of the
semester, students are asked to ‘defend’ their project
by performing some changes to its code, in order to
demonstrate knowledge and mastery of it
2
. Both the
exercise and project components are supported by an
Open-source Automatic Assessment Tool (AAT).
The course’s project is typically a command line
application that performs queries on a very large
data set, provided in the form of multiple CSV files.
The queries must be implemented using efficient data
structures and algorithms. This year, we used data
from the Million Song data set
3
. The project is split
in two parts: the first part is focused on reading and
parsing the input files, while the second part is fo-
cused on implementing the queries.
In this course, the teachers allow and even en-
courage students to interact with ChatGPT in a re-
sponsible way. This is true both for the weekly exer-
cises, as well as for the course’s main project. How-
ever, a cautionary note was added to the beginning
of the weekly assignments’ texts advising students to
attempt to solve the problems on their own before re-
sorting to technological support (not only ChatGPT,
but also Google, StackOverflow, and so on.).
Finally, this year’s project explicitly asked stu-
dents to use ChatGPT in one of the requirements, with
the caveat that it would be forbidden to use ChatGPT
during the project’s defense.
3.1 The ChatGPT Requirement and
Exercise
To encourage students to use a structured approach
with ChatGPT, in an attempt to promote their critical
thinking (Naumova, 2023), they were asked to imple-
ment one of the project’s requirements using Chat-
2
Project defenses are in-person and proctored.
3
http://millionsongdataset.com/
Ea ch l in e of th e so n g_ a rt i st s fi le f oll ow s one of
th e fo llo wi n g sy nta xe s d ep end ing on whe th er the
so ng is a ss o ci ate d wi th a sin gl e ar tis t or
mul ti p le a rt i st s :
< So ng ID > @ [ < Art is t N ame > ]
< So ng ID > @ "[ < Art is t N ame > , < Artis t N ame > ,
...]"
Where :
< So ng ID > is a Str in g ;
< Art is t N ame > is a Str in g ;
Listing 1: The ChatGPT requirement (partial instructions).
GPT. The requirement involved parsing one of the
project’s input files, which contained information as-
sociating songs with artists, as shown in Listing 1.
More concretely, they were instructed to approach
the requirement in the following way:
1. Prompt ChatGPT for help with the requirement;
2. Analyze the provided solution;
3. Prompt ChatGPT for an alternative solution, even-
tually providing further information;
4. Compare both GPT-generated solutions and write
a description of their findings;
5. Select one of the versions (i.e. the best one);
6. Integrate that version into their own project.
Furthermore, students were asked to submit a log
of their interactions with the LLM and their solution
comparison report. For more details about this ex-
ercise, as well as an analysis of the students’ logs,
please refer to (Alves and Cipriano, 2024).
4 SURVEY
4.1 Methodology
To help answer the research questions, students were
administered a structured anonymous questionnaire
during a class of the DSA course, and towards the
end of the semester. This questionnaire was com-
posed of multiple questions of various types (e.g. 1-5,
Yes/No, categorical, open-ended), with each question
contributing to one RQ. The questionnaire concluded
with an open-ended qualitative question asking stu-
dents for their opinions on GPT. The answers to the
open-ended questions were analysed by both authors
of this research, who then agreed on the best-fitting
RQ to place each of them.
Table 1 lists the questions used in the survey.
Of a total of 154 enrolled and participating stu-
dents, 52 (33.77%) replied to the questionnaire. Par-
CSEDU 2025 - 17th International Conference on Computer Supported Education
238
Table 1: The questions used in the survey, organized by Research Question.
RQ Question # Question Scale
RQ1 Q1.1 What is your opinion on the authorization by teachers of the use of GPT
in academic projects?
1-5
RQ1 Q1.2 In part 1 of the project, you were asked to interact with GPT. Did you? Yes/No
RQ1 Q1.3 If you answered “No” to the previous question, what is the reason for
not interacting with GPT?
Open-ended
RQ1 Q1.4 If not for the exercise, would you still have used GPT? Yes/No
RQ1 Q1.5 Do you think it makes sense to evaluate your ability to interact with
GPT?
Yes/No
RQ2 Q2.1 From 1 to 5, quantify how many interactions/prompts do you usually
need to get results that you consider useful. Consider the average num-
ber per problem you’ve tried to solve with GPT.
1-5
RQ2 Q2.2 Has the average number of interactions/prompts been reduced over
time?
1-5
RQ2 Q2.3 In cases where you asked ChatGPT for help, what percentage of the
exercises’ code was generated by ChatGPT?
[0-20%[ ... [80-100%]
RQ3 Q3.1 How useful do you think this exercise was (asking ChatGPT for help in
processing the song artists file)?
1-5
RQ3 Q3.2 Regarding the project in general (parts 1 and 2), how essential has
GPT’s help been for you to do the project?
1-5
RQ3 Q3.3 How often did you use ChatGPT to help you with the weekly assign-
ments?
1-5
RQ3 Q3.4 How comfortable are you with being evaluated (e.g., in an oral exam or
defense) about the code generated by GPT?
1-5
RQ4 Q4.1 Do you think it makes sense to include this type of exercise in the cur-
riculum?
Categorical
RQ4 Q4.2 How helpful would it be for teachers to explain how to use GPT effec-
tively (as they explain using the debugger, for example)?
1-5
RQ4 Q4.3 Did the existence of GPT have any influence on your attendance at the-
oretical classes?
Yes/No
RQ4 Q4.4 Did the existence of GPT have any influence on your attendance at prac-
tical classes?
Yes/No
RQ4 Q4.5 Do you think ChatGPT could replace the teacher, as it ends up being a
personal tutor?
Yes/No
RQ5 Q5.1 Will ChatGPT and other similar AI tools reduce the need for program-
mers?
1-5
RQ5 Q5.2 Do you feel that the GPT-based exercises done in this curricular unit
will make you more prepared for a possible future in which you have to
interact professionally with GPT?
1-5
ticipation was optional and no incentive was given.
The survey responses are available online
4
.
4.2 Results
This section presents our findings with regard to each
Research Question (RQ).
4
https://zenodo.org/records/8433052
4.2.1 RQ1: What Is the Opinion/sentiment of
first-year Students Regarding the
Academic Use of Tools like ChatGPT?
To answer this research question, our survey had 4
quantitative questions and 1 qualitative question.
In the first question, we asked students [Q1.1]
What is your opinion on the authorization by
teachers of the use of GPT in academic projects?
"ChatGPT Is Here to Help, not to Replace Anybody": An Evaluation of Students’ Opinions on Integrating ChatGPT in CS Courses
239
[Scale: 1-5]. Only 1.9% (1) of students selected op-
tion 1 (‘Strongly disagree’), and no student selected
option 2. 13.5% (7) selected the neutral position (i.e.
3), while 30.8% (16) selected option 4. Lastly, op-
tion 5 (’Strongly agree’) was selected by a majority
of 53.8% (28) of students.
[Q1.2] In part 1 of the project, you were asked to
interact with GPT. Did you? [Yes/No]. 84.6% (44)
of students replied “Yes”, while 15.4% (8) students
replied “No”.
[Q1.3] If you answered “No” to the previous ques-
tion, what is the reason for not interacting with
GPT? [Open-ended]. The following question was
only answered by the 8 students that replied “No” to
the previous question. We identified two major topics
within the students’ answers: “I didn’t feel the need
to” (6 occurrences) and “I didn’t know how to” (2 oc-
currences).
[Q1.4] If not for the exercise, would you still have
used GPT? [Yes/No]. This question tried to under-
stand if the students would have used GPT even if
there was no requirement for doing so. Most students,
69.2% (36) indicated that they would have used GPT
anyway, but 30.8% (16) indicated that they would not
have used GPT unless asked to do so. We found this
somewhat surprising, since we expected a higher per-
centage of students to indicate that they would have
used GPT anyway. However, some students may have
interpreted the question as being related to using GPT
to implement the requirement related with the previ-
ous question, and not generically, as we intended.
[Q1.5] Do you think it makes sense to evaluate your
ability to interact with GPT? [Yes/No]. The goal
of the fifth question was to understand if the students
would be open to being evaluated with regard to their
ability to use tools such as GPT. Students’ opinions on
this topic are somewhat divided, with 57.7% (30) in-
dicating that they think it does not make sense to eval-
uate their capacity to interact with GPT and 42.3%
(22) agreeing with the evaluation. The balance be-
tween “Yes” and “No” voters somewhat surprised us:
we expected a higher prevalence of the “No” option,
since most students usually have negative opinions
about being evaluated.
In conclusion... The majority of students agree with
the usage of GPT in academic contexts and a sub-
stantial fraction (32.3%) even agree on being assessed
on their usage. Student S21 says ”ChatGPT is a
faster and more organized search tool that helps stu-
dents solve problems and doubts they have”. How-
ever, there exists a minority of students who are ei-
ther against this practice or neutral about it. This may
be related to concerns of abuse and misuse: ”students
(...) request total or partial resolution of the exer-
cise by GPT, often without fully understanding how
the generated code works” (S44) and ”many people
use it without knowing even a little about the subject”
(S26).
If directed to do so, a vast majority of students
(84.6%) will use GPT in an assignment, and most stu-
dents will use it even if it’s not asked of them, but the
percentage is lower (69.2%). In general, it seems that
students have positive opinions with regard to these
tools and their academic usage.
4.2.2 RQ2: to What Extent Can first-year
Students Take Advantage of Tools like
ChatGPT Without any Specific or Formal
Training?
The goal of this RQ was to understand if students
would ‘naturally’ be able to take advantage of these
tools without having any formal training. This RQ
was composed of 3 quantitative questions.
The first question relevant for RQ2 was [Q2.1]
“From 1 to 5, quantify how many interaction-
s/prompts do you usually need to get results that
you consider useful? Consider the average num-
ber per problem you’ve tried to solve with GPT.
[Scale: 1-5]. No student selected option 1, A sin-
gle prompt’. Option 2, ‘A few prompts’, was selected
by 38.5% (20) of students, making it the most se-
lected option, closely followed by option 3, selected
by 36.5% (19) of students. Option 4, ‘Many prompts’
was selected by 23.1% (12) of students. Finally, op-
tion 5, ‘I usually can’t get useful results’ was selected
by 1.9% (a single student). This shows us that most
students are usually able to get GPT to produce useful
results, but a significant fraction needs many prompts,
probably more than would be necessary. This could
mean that the students are not being effective in their
prompts and could benefit from prompting training.
As for the second question, [Q2.2] “Has the aver-
age number of interactions/prompts been reduced
over time?”, where 1 means ”No” and 5 means ”Yes,
substantially reduced”, the option with more votes
was number 3, which was selected by 40% (21) of
students. Options 4 and 5 followed, with 19.2% (10)
and 17.3% (9), respectively. Finally, options 1 and 2
were both selected by 11.5% (6) of students.
[Q2.3] “In cases where you asked ChatGPT for
help, what percentage of the exercises’ code was
generated by ChatGPT?” Scale: Interval-based,
ranging from [0-20%[ to [80-100%]. As depicted in
Figure 1, the largest group of students, 41.06% (16),
fell within the 20-40% range, closely followed by
33.33% (13) who responded with values below 20%.
Still, a combined sum of 10.25% selected options 60-
80% (3 students) and 80-100% (1 student). These
CSEDU 2025 - 17th International Conference on Computer Supported Education
240
# Students
0
4
8
12
16
0-20%
20-40%
40-60%
60-80%
80-100%
Percentage of code generated by GPT
Figure 1: Responses to the question “[Q2.3] What percent-
age of the exercises’ code was generated by ChatGPT?”.
results suggest that most students seem to be using
the tool to assist or supplement their coding process
rather than relying on it completely. However, they
also show that a few students tend to abuse the tool:
using GPT to generate more than 40% of an assign-
ment probably means that the student is over-relying
on it.
We decided to further investigate if the students
that indicated the need for “Many prompts” (4) in
Q2.1 also considered that they were improving over
time. From the 12 students that selected that option,
1 student selected 1 (or “No improvement”), 3 stu-
dents selected 2, 6 students selected the middle op-
tion (3) and other 2 students selected option number
4. This shows that, even the students that have more
difficulty (or report requiring more prompts), feel that
they are improving somewhat, over time. However,
only 2 out of those 12 students reported significant
improvement.
In conclusion... Students’ answers to these questions
suggest that, in general, they are able to use GPT in
a useful way. However, the average number of in-
teractions needed seems to vary significantly between
the students. This might be due to GPT not being
deterministic, but it can also indicate that some stu-
dents lack some ‘prompting skill‘. Student S41 says:
”Although ChatGPT is useful in some situations, it is
not possible to obtain optimal results consistently”.
Since some students (36.5%) report improvements
over time, we think that it is likely that ‘prompting
skill‘ and/or ‘prompting experience’ are at play here.
That being said, if improvements are seen with unsu-
pervised and untrained usage, then guided and trained
usage could have the potential to yield even more im-
provements. As such, we believe that students would
benefit from having some level of guidance provided
by their teachers, in order to make them more efficient
at interacting with these tools.
4.2.3 RQ3: What Is the Impact of ChatGPT on
the Learning Experience of first-year
Students?
Q3.2 How essential was !
GPT help during!
the project
Q3.3 How often did !
you use GPT in !
weekly assignments
Q3.4 How comfortable !
being evaluated on !
code generated by GPT
Figure 2: Responses to some of the questions related to RQ3
(impact of ChatGPT on students’ learning experience). The
questions were on 1-5 scale, with 1 (lighter color) being
the ’lesser’ option and 5 (darker color) being the ’greater’
option.
The goal of this RQ was to understand if tools such as
ChatGPT have some kind of impact on the students’
learning experience, as per their own perception. To
answer this RQ, the questionnaire had 4 quantitative
questions.
[Q3.1] “How useful do you think this exercise
was (asking ChatGPT for help in processing the
song artists file)?” [Scale: 1-5]. As depicted in Fig-
ure 2, option 1 (‘Useless’), was selected by 5.8% (3)
of the students, while option 2 (‘Slightly useful‘) was
selected by 21.2% (11). Options 3 and 4 were both se-
lected by the same percentage of students, with 30.8%
(16) of participants selecting each of them. Finally,
option 5 (‘Very useful’), was selected by 11.5% (6) of
students. These results show that most students con-
sidered the exercise useful, although opinions vary
widely.
[Q3.2] “Regarding the project in general (parts 1
and 2). How essential has GPT’s help been for you
to do the project?” [Scale: 1-5]. As illustrated in
Figure 2, 17.3% (9) of students felt that GPT’s assis-
tance was ‘Not essential’ (option 1), implying min-
imal to no reliance on the tool. A slightly smaller
segment, 15.4% (8), chose option 2, placing their re-
liance between ‘Not essential’ and ‘Mildly essential’.
A majority of respondent, 55.8% (29) selected op-
tion 3 (or ‘Mildly essential’), suggesting they con-
sulted GPT for assistance on some portions of their
project. On the higher end of dependency, 7.7% (4)
leaned more towards frequent usage but not complete
reliance. However, a very small fraction, 3.8% (2),
found GPT to be ‘Completely essential,’ relying on it
for almost the entirety of their project. In summary,
while a majority of students used GPT occasionally
for assistance, very few relied on it for the entire scope
of their work.
"ChatGPT Is Here to Help, not to Replace Anybody": An Evaluation of Students’ Opinions on Integrating ChatGPT in CS Courses
241
[Q3.3] “How often did you use ChatGPT to help
you with the weekly assignments?” [Scale: 1-5]. As
can be seen in Figure 2, option 1, meaning ‘Never’,
was selected by 25% (13) of students. The most se-
lected option was option number 2, meaning ‘Only
very sporadically’, which was selected by 42.3% (22)
of students. Another 25% (13) selected the mid-point
(option 3). Option 4 was selected by 3.8% (2) of stu-
dents. Finally, at the extreme end of frequent adop-
tion, another 3.8% (2) of the students selected option
5, which was labeled Always’. In summary, the ma-
jority of students either never used GPT in the weekly
assignments or did so very sporadically. However, a
few students seem prone to over reliance.
[Q3.4] “How comfortable are you with being eval-
uated (for example, in an oral exame or in a
defense) about the code that was generated by
GPT?”. [Scale 1-5]. A minor 3.8% (2) opted for the
least comfortable option, which was labeled as ‘Not
comfortable, I used the code and it works, but I don’t
understand what it does’. 17.3% (9) students opted
for the second option, while option 3, indicating a
mid-range comfort level, was selected by 26.9% (14).
Option 4 was the choice for 19.2% (10). Significantly,
the most prominent response was option 5, labeled as
’Very comfortable, although I didn’t write the code,
I fully understand how it works’, which was chosen
by 32.7% (17). This suggests a substantial portion of
the participants felt highly comfortable in being eval-
uated in terms of their mastery of GPT’s generated
code, even if they hadn’t written the code themselves.
Refer to Figure 2 for a visualization of this question’s
reply distribution.
In conclusion... Students’ opinions were divided
about the usefulness of the specific GPT exercise.
Even though most students tended to find it useful,
there was a significant portion (27%) that found the
exercise with little or no usefulness. We hypothesize
that these may be weaker students that couldn’t get
GPT to reach useful solutions, that they could use in
their project.
A small fraction of students have shown a high
dependency of GPT to implement the project (6 stu-
dents or 11.5%). On the other extreme, only 17,3%
indicated that they could have done the project with-
out ever resorting to GPT for help, so the majority
used it for occasional help. Regarding weekly assign-
ments, the reliance on GPT decreases, with only 4
students (7.6%) admitting the need for full assistance.
This led us to conclude that, although some students
do abuse GPT, they are not the majority which only
uses it sporadically. Interestingly enough, student S45
says: “ChatGPT is a useful tool but I think it is mostly
useful for research and clarifying small doubts”.
Finally, one of the most surprising results was re-
lated to assessment. Almost a third of the students feel
very comfortable about being evaluated on code gen-
erated by ChatGPT. This suggests that students anal-
yse and try to understand GPT-generated code rather
than using it blindly.
4.2.4 RQ4: What Is the Impact of ChatGPT on
Current Teaching Practices?
This RQ tried to find the students’ perceptions on
how the current teaching methodologies and assess-
ments should change because of GPT. Instead of un-
derstanding how teachers will change (the teacher’s
perspective), it is about understanding how students
think the teachers should change (the student’s per-
spective). Five questions were used to address this
RQ.
Figure 3: Pie chart for question “[Q4.1] Do you think it
makes sense to include this type of exercises in the curricu-
lum?”. All students agree with having GPT-based exercises,
but they are divided in terms of evaluation scheme.
[Q4.1] Do you think it makes sense to include this
type of exercises in the curriculum? [Categori-
cal]. This question aimed at understanding if stu-
dents agree with having GPT-related contents in the
course’s curricula. As illustrated in Figure 3, all stu-
dents think that it makes sense to include these types
of exercises in the courses. 57.7% (30) think the exer-
cises should not be mandatory nor graded, 32.7% (17)
think that the exercises should exist and be graded, but
only in a binary fashion (‘Used GPT’ / ‘Did not use
GPT’), and finally, 9.6% (5) think that the exercises
should exist and be evaluated in a detailed fashion.
All students agree on the inclusion of GPT-related ex-
ercises in university courses. However, there is a di-
vergence in opinions regarding the assessment of in-
teractions with GPT, as more than half of the students
believe there should be no evaluation for such exer-
cises.
Next, we asked students [Q4.2] “How helpful
would it be for teachers to explain how to use GPT
effectively (as they explain using the debugger, for
example)?” [Scale: 1-5]. No students selected the
first option (‘Not helpful’), 5.8% (3) selected the sec-
ond option, 17.3% (9) selected the middle option,
CSEDU 2025 - 17th International Conference on Computer Supported Education
242
25% (13) selected the fourth option and, finally, a ma-
jority of 51.9% (27) students selected the final option
(‘Very helpful’). These replies indicate that most stu-
dents consider that having teachers explain how to use
GPT would be useful.
The third question was a Yes/No question: [Q4.3]
“Did the existence of GPT have any influence on
your attendance at theoretical classes?”. All stu-
dents replied with ‘No’.
The fourth question was similar to the third one,
but focused on the practical (or laboratory) classes:
[Q4.4] “Did the existence of GPT have any influ-
ence on your attendance at practical classes?”. The
vast majority of the students (48 students, or 98.33%)
answered ‘No’, while only 4 students (7.7%) students
answered ‘Yes’.
Finally, the last question was: [Q4.5] “Do you
think ChatGPT could replace the teacher, as it
ends up being a personal tutor?”. Again, the vast
majority of students answered ‘No’ (50 students, or
96.2 %), with only 2 students selecting ‘Yes’.
In conclusion... All students believe that these types
of exercises should be included in the courses’ cur-
ricula. However, there are significant divergences on
whether these exercises should be graded, with most
students (57.7%) being against any form of grading.
Also, most students believe that it would be useful for
teachers to explain how to use GPT. This is in line
with the conclusions of RQ2: some students’ less ef-
ficient usage of GPT may be improved with formal
training and teacher guidance. Students confirm this
observation: “I think ChatGPT is a useful tool, but to
do so it has to be well used and well understood. It
would be great if there was help on how to use this
tool from teachers” (S26) and “[teachers should] ex-
plain to students more effective ways to use the tool
and adapt to it” (S39).
The availability of these tools is not having a sig-
nificant influence in the students’ class attendance,
which is further confirmed by the fact that the vast
majority of students (96.2%) believe that GPT will
not replace the teacher. This result is consistent with
the findings of (Prather et al., 2023a).
4.2.5 RQ5: What Is the Opinion of first-year
Students Regarding the Influence of
AI-Based Code Generation Tools (like
ChatGPT) on Their Professional Future?
This RQ was addressed through the use of 2 quantita-
tive questions.
The first question was [Q5.1] “Will ChatGPT
and other similar AI tools reduce the need for pro-
grammers?” [Scale: 1-5]. Option 1, ‘Strong dis-
agreement’, was selected by 11.5% (6) of students,
while option 2 was selected by 21.2% (11). The mid-
dle option was selected by 38.5% (20) of students,
making it the most frequently selected option. Op-
tions 4 and 5 were selected by 23.1% (12) and 5.8%
(3) of students, respectively. As such, students are
more or less divided on this topic.
Finally, students were asked [Q5.2] “Do you feel
that the GPT-based exercises done in this curricu-
lar unit will make you more prepared for a possible
future in which you have to interact professionally
with GPT?” [Scale: 1-5]. Only one student selected
the first option —‘No, they will make it worse’—
while 9.6% (5) students selected the second option—
‘Neutral: they will have no effect’. 30.8% (16) of stu-
dents selected the middle option, 34.6% (18) selected
the fourth option, and, finally, 23.1% (12) selected the
last option —‘Yes, very much’. Most students believe
these exercises will have some importance to their ca-
reers if they have to interface with GPT profession-
ally.
In conclusion... Students are not sure if GPT and
similar AI code generating technologies will reduce
the need for programmers, but they tend to agree that
having GPT-based exercises in the DSA course will
help them in their professional careers. One of the
students said: “ChatGPT is here to help, not to re-
place anyone” (S7), which inspired us for the title of
the paper.
5 RECOMMENDATIONS FOR CS
EDUCATORS
Students are likely to misuse tools like GPT, which
offer decent-to-good code generation for free. Also,
LLMs are already being used by professional soft-
ware developers (Barke et al., 2023), and this trend
will possibly improve in the future as the models’ ca-
pabilities advance further. This makes it important for
students to get quality experiences with regard to in-
teracting with these tools, so that they are aware of the
tools’ limitations and use them correctly. As such, we
recommend that CS Educators start integrating these
models in their courses. Following are some ideas de-
rived from the survey results.
Teach Prompt-Engineering Techniques. Teach
your students prompt engineering (PE) techniques,
such as Role-playing (Kong et al., 2023) and Chain-
of-Thought prompting (Wei et al., 2022).
Employ LLM-Based Exercises. Design exercises in
which students should interact with ChatGPT or sim-
ilar tools. These exercises should be designed in or-
der to promote skills such as critical thinking, code
reading and understanding, and code critique. Exer-
"ChatGPT Is Here to Help, not to Replace Anybody": An Evaluation of Students’ Opinions on Integrating ChatGPT in CS Courses
243
cises can range from prompt creation (Denny et al.,
2024), prompt improvement, assisted code generation
and respective critique/comparison (such as our exer-
cise), unit test generation, mistake finding (Naumova,
2023), and so on. Also, they should be supervised by
the teacher, who should guide and help students in this
process, helping them identify problems with LLMs’
output, thus promoting that students to not blindly
trust the models.
Evaluate Students’ Prompting Abilities. Consider
evaluating students’ prompting abilities and knowl-
edge of PE techniques. Students appear open to be
evaluated on this skill, which could be taken advan-
tage of in order to promote their further development.
6 LIMITATIONS
The survey’s authors are part of the DSA course
teaching staff. This might have influenced some of
the students’ opinions.
As the survey participation was optional, there
might have been some selection bias, and the students
that decided to answer the questionnaire could be the
ones that already had mostly positive opinions about
these tools. Also, the students that participated were
present in a theoretical class, which means that their
opinions about attendance might not be representative
of the general population.
GPT’s behaviour is not deterministic and can also
vary greatly over time (Chen et al., 2023). This
makes it difficult to generalize conclusions about stu-
dents’ interactions with LLMs, as differences in out-
comes may not necessarily reflect variations in stu-
dents’ ‘prompting skills’ but could instead result from
the inherent variability of the models themselves.
We expect most students to have used ChatGPT
based on GPT-3.5, due to it being free. However,
some students might have used the paid GPT-4 model,
which was released at our location while our course
was running
5
. Our results were not controlled for
these model differences.
In relation to the questions about their present
GPT-usage during weekly assignments and project
(Q2.3, Q3.2 and Q3.3), it is conceivable that some stu-
dents provided inaccurate responses in order to give
us a false sense of security. However, we find this un-
likely, since a significant number of students indicated
they were comfortable in defending code generated
by GPT.
5
In Europe, GPT-4 became available for paying sub-
scribers in March, 14, 2023
7 CONCLUSIONS
Our study shows that students are generally in favor of
using GPT in academic contexts, but they are divided
regarding its assessment (RQ1). Also, most students
are already benefiting from GPT but some of them
are not effective and efficient in its utilization, often
requiring numerous prompts or failing to achieve a
satisfactory solution (RQ2). Counter-intuitively, most
students don’t over-rely on GPT in the weekly assign-
ments and project. Furthermore, a significant portion
of students is comfortable with being evaluated on
code generated by these tools (RQ3). The inclusion
of specific GPT lessons and exercises is seen as ben-
eficial and desirable (RQ4). Finally, students are con-
siderably divided on the impact of GPT on their future
professional careers, but in general, they agree that
practicing its utilization during their academic jour-
ney is important (RQ5). We believe that these find-
ings underscore the importance of embracing LLMs
in CS courses. In light of these findings, educators are
encouraged to design coursework that leverages GPT
not simply as a shortcut, but as a means to strengthen
problem-solving skills and critical thinking, ensuring
students engage thoughtfully with AI tools in ways
that enhance their learning and future readiness.
ACKNOWLEDGEMENTS
This research has received funding from the European
Union’s DIGITAL-2021-SKILLS-01 Programme un-
der grant agreement no. 101083594.
REFERENCES
Alves, P. and Cipriano, B. P. (2024). ”Give me the code”
Log Analysis of First-Year CS Students’ Interactions
With GPT. arXiv preprint arXiv:2411.17855.
Barke, S., James, M. B., and Polikarpova, N. (2023).
Grounded Copilot: How Programmers Interact with
Code-Generating Models. Proc. of the ACM on Pro-
gramming Languages, 7(OOPSLA1):85–111.
Chen, L., Zaharia, M., and Zou, J. (2023). How is Chat-
GPT’s behavior changing over time? arXiv preprint
arXiv:2307.09009.
Cipriano, B. P. and Alves, P. (2024). LLMs Still
Can’t Avoid Instanceof: An investigation Into GPT-
3.5, GPT-4 and Bard’s Capacity to Handle Object-
Oriented Programming Assignments. In Proc. of the
IEEE/ACM 46th International Conference on Soft-
ware Engineering: Software Engineering Education
and Training (ICSE-SEET).
Cipriano, B. P., Alves, P., and Denny, P. (2024). A Picture
Is Worth a Thousand Words: Exploring Diagram and
CSEDU 2025 - 17th International Conference on Computer Supported Education
244
Video-Based OOP Exercises to Counter LLM Over-
Reliance. In European Conference on Technology En-
hanced Learning, pages 75–89. Springer.
Daun, M. and Brings, J. (2023). How ChatGPT Will
Change Software Engineering Education. In Proc. of
the 2023 Conference on Innovation and Technology in
Computer Science Education V. 1, pages 110–116.
Denny, P., Leinonen, J., Prather, J., Luxton-Reilly, A.,
Amarouche, T., Becker, B. A., and Reeves, B. N.
(2024). Prompt Problems: A New Programming Ex-
ercise for the Generative AI Era. In Proc. of the 55th
ACM Technical Symposium on Computer Science Ed-
ucation V. 1, pages 296–302.
Finnie-Ansley, J., Denny, P., Becker, B. A., Luxton-Reilly,
A., and Prather, J. (2022). The Robots Are Coming:
Exploring the Implications of OpenAI Codex on In-
troductory Programming. In Proc. of the 24th Aus-
tralasian Computing Education Conference, pages
10–19.
Finnie-Ansley, J., Denny, P., Luxton-Reilly, A., Santos,
E. A., Prather, J., and Becker, B. A. (2023). My AI
Wants to Know if This Will Be on the Exam: Testing
OpenAI’s Codex on CS2 Programming Exercises. In
Proc. of the 25th Australasian Computing Education
Conference, pages 97–104.
Hellas, A., Leinonen, J., Sarsa, S., Koutcheme, C., Ku-
janp
¨
a
¨
a, L., and Sorva, J. (2023). Exploring the Re-
sponses of Large Language Models to Beginner Pro-
grammers’ Help Requests.
Kong, A., Zhao, S., Chen, H., Li, Q., Qin, Y., Sun,
R., and Zhou, X. (2023). Better Zero-Shot Rea-
soning with Role-Play Prompting. arXiv preprint
arXiv:2308.07702.
Lau, S. and Guo, P. (2023). From ”Ban it till we understand
it” to ”Resistance is futile”: How university program-
ming instructors plan to adapt as more students use AI
code generation and explanation tools such as Chat-
GPT and GitHub Copilot.
Leinonen, J., Denny, P., MacNeil, S., Sarsa, S., Bernstein,
S., Kim, J., Tran, A., and Hellas, A. (2023). Compar-
ing Code Explanations Created by Students and Large
Language Models. In Proc. of the 2023 Conference
on Innovation and Technology in Computer Science
Education V. 1, pages 124–130.
Lenfant, R., Wanner, A., Hott, J. R., and Pettit, R. (2023).
Project-based and assignment-based courses: A study
of piazza engagement and gender in online courses. In
Proc. of the 2023 Conference on Innovation and Tech-
nology in Computer Science Education V. 1, pages
138–144.
Liffiton, M., Sheese, B., Savelka, J., and Denny, P.
(2023). CodeHelp: Using Large Language Models
with Guardrails for Scalable Support in Programming
Classes.
Naumova, E. N. (2023). A mistake-find exercise: a
teacher’s tool to engage with information innovations,
ChatGPT, and their analogs. Journal of Public Health
Policy, 44(2):173–178.
OpenAI (2023). How can educators respond to stu-
dents presenting ai-generated content as their
own? https://help.openai.com/en/articles/8313351-
how-can-educators-respond-to-students-presenting-
ai-generated-content-as-their-own. [Online; last
accessed 03-October-2023].
Prather, J., Denny, P., Leinonen, J., Becker, B. A., Al-
bluwi, I., Craig, M., Keuning, H., Kiesler, N., Kohn,
T., Luxton-Reilly, A., MacNeil, S., Peterson, A., Pet-
tit, R., Reeves, B. N., and Savelka, J. (2023a). The
Robots are Here: Navigating the Generative AI Revo-
lution in Computing Education.
Prather, J., Reeves, B. N., Denny, P., Becker, B. A.,
Leinonen, J., Luxton-Reilly, A., Powell, G., Finnie-
Ansley, J., and Santos, E. A. (2023b). “It’s Weird That
it Knows What I Want”: Usability and Interactions
with Copilot for Novice Programmers. ACM Transac-
tions on Computer-Human Interaction, 31(1):1–31.
Rahman, M. M. and Watanobe, Y. (2023). ChatGPT for
Education and Research: Opportunities, threats, and
Strategies. Applied Sciences, 13(9):5783.
Reeves, B., Sarsa, S., Prather, J., Denny, P., Becker, B. A.,
Hellas, A., Kimmel, B., Powell, G., and Leinonen, J.
(2023). Evaluating the Performance of Code Genera-
tion Models for Solving Parsons Problems With Small
Prompt Variations. In Proc. of the 2023 Conference on
Innovation and Technology in Computer Science Ed-
ucation V. 1, pages 299–305.
Savelka, J., Agarwal, A., An, M., Bogart, C., and Sakr, M.
(2023a). Thrilled by Your Progress! Large Language
Models (GPT-4) No Longer Struggle to Pass Assess-
ments in Higher Education Programming Courses.
Savelka, J., Agarwal, A., Bogart, C., Song, Y., and Sakr,
M. (2023b). Can Generative Pre-trained Transform-
ers (GPT) Pass Assessments in Higher Education Pro-
gramming Courses? In Proc. of the 2023 Conference
on Innovation and Technology in Computer Science
Education V. 1. ACM.
Singh, H., Tayarani-Najaran, M.-H., and Yaqoob, M.
(2023). Exploring Computer Science students’ Per-
ception of ChatGPT in Higher Education: A De-
scriptive and Correlation Study. Education Sciences,
13(9):924.
Wei, J., Wang, X., Schuurmans, D., Bosma, M., Xia, F.,
Chi, E., Le, Q. V., Zhou, D., et al. (2022). Chain-of-
Thought Prompting Elicits Reasoning in Large Lan-
guage Models. Advances in Neural Information Pro-
cessing Systems, 35:24824–24837.
Xu, F. F., Alon, U., Neubig, G., and Hellendoorn, V. J.
(2022). A Systematic Evaluation of Large Language
Models of Code. In Proc. of the 6th ACM SIGPLAN
International Symposium on Machine Programming,
pages 1–10, San Diego CA USA. ACM.
Yilmaz, R. and Yilmaz, F. G. K. (2023). Augmented in-
telligence in programming learning: Examining stu-
dent views on the use of ChatGPT for programming
learning. Computers in Human Behavior: Artificial
Humans, 1(2):100005.
Zastudil, C., Rogalska, M., Kapp, C., Vaughn, J., and Mac-
Neil, S. (2023). Generative AI in Computing Educa-
tion: Perspectives of Students and Instructors. In 2023
IEEE Frontiers in Education Conference (FIE), pages
1–9. IEEE.
"ChatGPT Is Here to Help, not to Replace Anybody": An Evaluation of Students’ Opinions on Integrating ChatGPT in CS Courses
245