Half of teachers say they know a student at their school who was disciplined or faced negative consequences for using — or being accused of using — generative artificial intelligence like ChatGPT to complete a classroom assignment, according to survey results released Wednesday by the Center for Democracy and Technology, a nonprofit think tank focused on digital rights and expression. The proportion was even higher (58 percent) for those who teach special education.
Cheating concerns were clear, with survey results showing that teachers have grown suspicious of their students. Nearly two-thirds of teachers said that generative AI has made them “more distrustful” of students, and 90 percent said they suspect kids are using the tools to complete assignments. Yet students themselves who completed the anonymous survey said they rarely use ChatGPT to cheat, but are turning to it for help with personal problems.
“The difference between the hype cycle of what people are talking about with generative AI and what students are actually doing, there seems to be a pretty big difference,” said Elizabeth Laird, the group’s director of equity in civic technology. “And one that, I think, can create an unnecessarily adversarial relationship between teachers and students.”
Indeed, 58 percent of students, and 72 percent of those in special education, said they’ve used generative AI during the 2022-23 academic year, just not primarily for the reasons that teachers fear most. Among youth who completed the nationally representative survey, just 23 percent said they used it for academic purposes and 19 percent said they’ve used the tools to help them write and submit a paper. Instead, 29 percent reported having used it to deal with anxiety or mental health issues, 22 percent for issues with friends, and 16 percent for family conflicts.
Part of the disconnect dividing teachers and students, researchers found, may come down to gray areas. Just 40 percent of parents said they or their child were given guidance on ways they can use generative AI without running afoul of school rules. Only 24 percent of teachers say they’ve been trained on how to respond if they suspect a student used generative AI to cheat.
Among parents, 73 percent said they’re concerned about the privacy and security of student data collected and stored by schools, a considerable increase from the 61 percent who expressed those reservations last year. A similar, if less dramatic, trend was apparent among students: 62 percent had data privacy concerns tied to their schools, compared with 57 percent just a year earlier.
Survey results found that students in special education, whose records are among the most sensitive that districts maintain, and their parents were significantly more likely than the general education population to report school data privacy and security concerns. As attacks ratchet up, one in five parents say they’ve been notified that their child’s school experienced a data breach. Such breach notices, Laird said, led to heightened apprehension.
“There’s not a lot of transparency” about school cybersecurity incidents “because there’s not an affirmative reporting requirement for schools,” Laird said. But in instances where parents are notified of breaches, “they are more concerned than other parents about student privacy.”
Parents and students have also grown increasingly wary of another set of education tools that rely on artificial intelligence: digital surveillance technology. Among them are student activity monitoring tools, such as those offered by the for-profit companies Gaggle and GoGuardian, which rely on algorithms in an effort to keep students safe. The surveillance software employs artificial intelligence to sift through students’ online activities and flag school administrators — and sometimes the police — when they discover materials related to sex, drugs, violence, or self-harm.
Among parents surveyed this year, 55 percent said they believe the benefits of activity monitoring outweigh the potential harms, down from 63 percent last year. Among students, 52 percent said they’re comfortable with academic activity monitoring, a decline from 63 percent last year.
Such digital surveillance, researchers found, frequently has disparate impacts on students based on their race, disability, sexual orientation, and gender identity, potentially violating long-standing federal civil rights laws.
The tools also extend far beyond the school realm, with 40 percent of teachers reporting their schools monitor students’ personal devices. More than a third of teachers say they know a student who was contacted by the police because of online monitoring, the survey found, and Black parents were significantly more likely than their white counterparts to fear that information gleaned from online monitoring tools and AI-equipped campus surveillance cameras could fall into the hands of law enforcement.
The nonprofit highlights how disproportionalities identified in the survey could run counter to federal laws that prohibit discrimination based on race and sex, and those designed to ensure equal access to education for children with disabilities. In a letter sent Wednesday to the White House and Education Secretary Miguel Cardona, the Center for Democracy and Technology was joined by a coalition of civil rights groups urging federal officials to take a harder tack on ed tech practices that could threaten students’ civil rights.
“Existing civil rights laws already make schools legally responsible for their own conduct, and that of the companies acting at their direction in preventing discriminatory outcomes on the basis of race, sex, and disability,” the coalition wrote. “The department has long been responsible for holding schools accountable to these standards.”
Fast Company © 2023 Mansueto Ventures, LLC. Distributed by Tribune Content Agency, LLC.