Itâs not just about cheating. How AI is quietly eroding college studentsâ networks
Itâs not just about cheating. How AI is quietly eroding college studentsâ networks
Students donât have the same incentives to talk to their professors â or even their classmates â anymore. Chatbots like ChatGPT, Gemini and Claude have given them a new path to self-sufficiency. Instead of asking a professor for help on a paper topic, students can go to a chatbot. Instead of forming a study group, students can ask AI for help. These chatbots give them quick responses, on their own timeline.
For students juggling school, work and family responsibilities, that ease can seem like a lifesaver. And maybe turning to a chatbot for homework help here and there isnât such a big deal in isolation. But every time a student decides to ask a question of a chatbot instead of a professor or peer or tutor, thatâs one fewer opportunity to build or strengthen a relationship, and the human connections students make on campus are among the most important benefits of college. examines how the pervasiveness of AI tools is impacting studentsâ ability to build their networks.
Julia Freeland-Fisher studies how technology can help or hinder student success at the . She said the consequences of turning to chatbots for help can compound.
âOver time, that means students have fewer and fewer people in their corner who can help them in other moments of struggle, who can help them in ways a bot might not be capable of,â she said.
As colleges further embed ChatGPT and other chatbots into campus life, Freeland-Fisher warns lost relationships may become a devastating unintended consequence.
Asking for help
Christian Alba said he has never turned in an AI-written assignment. Alba, 20, attends College of the Canyons, a large community college north of Los Angeles, where he is studying business and history. And while he hasnât asked ChatGPT to write any papers for him, he has turned to the technology when a blank page and a blinking cursor seemed overwhelming. He has asked for an outline. He has asked for ideas to get him started on an introduction. He has asked for advice about what to prioritize first.
âItâs kind of hard to just start something fresh off your mind,â Alba said. âI wonât lie. Itâs a helpful tool.â Alba has wondered, though, whether turning to ChatGPT with these sorts of questions represents an overreliance on AI. But Alba, like many others in higher education, worries primarily about AI use as it relates to academic integrity, not social capital. And thatâs a problem.
Jean Rhodes, a psychology professor at the University of Massachusetts Boston, has spent decades studying the way college students seek help on campus and how the relationships formed during those interactions end up benefitting the students long-term. Rhodes doesnât begrudge students integrating chatbots into their workflows, as many of their professors have, but she worries that students will get inferior answers to even simple-sounding questions, like, âhow do I change my major?â
A chatbot might point a student to the registrarâs office, Rhodes said, but had a student asked the question of an advisor, that person may have asked important follow-up questions â why the student wants the change, for example, which could lead to a deeper conversation about a studentâs goals and roadblocks.
âWe understand the broader context of studentsâ lives,â Rhodes said. âTheyâre smart but theyâre not wise, these tools.â
Rhodes and one of her former doctoral students, Sarah Schwartz, created a program called Connected Scholars to help students understand why itâs valuable to talk to professors and have mentors. The program helped them hone their networking skills and understand what people get out of their networks over the course of their lives â namely, social capital.
Connected Scholars is offered as a semester-long course at UMass Boston, and a forthcoming paper examines outcomes over the last decade, finding students who take the course are three times more likely to graduate. Over time, Rhodes and her colleagues discovered that the key to the programâs success is getting students past an aversion to asking others for help.
Students will make a plethora of excuses to avoid asking for help, Rhodes said, ticking off a list of them: ââI donât want to stand out,â âI donât want people to realize I donât fit in here,â âMy culture values independence,â âI shouldnât reach out,â âIâll get anxious,â âThis person wonât respond.â If you can get past that and get them to recognize the value of reaching out, itâs pretty amazing what happens.â
Connections are key
Seeking human help doesnât only leave students with the resolution to a single problem, it gives them a connection to another person. And that person, down the line could become a friend, a mentor or a business partner â a âstrong tie,â as social scientists describe their centrality to a personâs network. They could also become a âweak tieâ who a student may not see often, but could, importantly, still offer or crucial one day.
Daniel Chambliss, a retired sociologist from Hamilton College, emphasized the value of relationships in his 2014 book, âHow College Works,â co-authored with Christopher Takacs. Over the course of their research, the pair found that the key to a successful college experience boiled down to relationships, specifically two or three close friends and one or two trusted adults. Hamilton College goes out of its way to make sure students can form those relationships, structuring work-study to get students into campus offices and around faculty and staff, making room for students of varying athletic abilities on sports teams, and more.
Chambliss worries that AI-driven chatbots make it too easy to avoid interactions that can lead to important relationships. âWeâre suffering epidemic levels of loneliness in America,â he said. âItâs a really major problem, historically speaking. Itâs very unusual, and itâs profoundly bad for people.â
As students increasingly turn to artificial intelligence for help and even casual conversation, Chambliss predicted it will make people even more isolated: âItâs one more place where they wonât have a personal relationship.â
In fact, a published in March found that the most frequent users of ChatGPT â power users â were more likely to be lonely and isolated from human interaction.
âWhat scares me about that is that Big Tech would like all of us to be power users,â said Freeland-Fisher. âThatâs in the fabric of the business model of a technology company.â
Yesenia Pacheco is preparing to reenroll in Long Beach City College for her final semester after more than a year off. Last time she was on campus, ChatGPT existed, but it wasnât widely used. Now she knows sheâs returning to a college where ChatGPT is deeply embedded in studentsâ as well as faculty and staffâs lives, but Pacheco expects sheâll go back to her old habits â going to her professorsâ office hours and sticking around after class to ask them questions. She sees the value.
She understands why others might not. Todayâs high schoolers, she has noticed, are not used to talking to adults or building mentor-style relationships. At 24, she knows why they matter.
âA chatbot,â she said, âisnât going to give you a letter of recommendation.â
was produced by and reviewed and distributed by ÂéśšÔ´´.