Autistic people turn to ChatGPT for advice on workplace issues

Autism and Chatbots

A new study shows that many autistics embrace ChatGPT and similar AI tools for help and advice as they confront workplace problems. But does the use of AI make sense? CREDIT JiWoong Jang and Sanika Moharana are Ph.D. students at the Human-Computer Interaction Institute.

A new Carnegie Mellon University study shows that many autistic people embrace ChatGPT and similar artificial intelligence tools for help and advice when confronting workplace problems.

Controversy remains within the autism community as to whether this use of chatbots is even a good idea.

“What we found is there are autistic people who are already using ChatGPT to ask questions that we think ChatGPT is partly well-suited and partly poorly suited for,” said Begel, “For instance, they might ask: ‘How do I make friends at work?'”

Begel heads the VariAbility Lab, which seeks to develop workplaces where all people, including those with disabilities and neurodivergent, can successfully work together. Unemployment and underemployment are problems for as many as nine out of 10 autistic adults, and many workplaces don’t have the resources to help employees with autism and their coworkers overcome social or communication problems as they arise.

To better understand how large language models (LLMs) could be used to address this shortcoming, Begel and his team recruited 11 people with autism to test online advice from two sources — a chatbot based on OpenAI’s GPT-4 and what looked to the participants like a second chatbot but was really a human career counsellor.

Surprisingly, the users overwhelmingly preferred the real chatbot to the disguised counsellor. It’s not that the chatbot gave better advice, Begel said, but rather how it dispensed it.

“The participants prioritized getting quick and easy-to-digest answers,” Begel said.

The chatbot provided black-and-white answers without a lot of subtlety and usually in the form of bullets. The counsellor, by contrast, often asked questions about what the user wanted to do or why they wanted to do it. Most users preferred not to engage in such back-and-forth, Begel said.

Participants liked the concept of a chatbot. One explained: “I think, honestly, with my workplace … it’s the only thing I trust because not every company or business is inclusive.” 

However, when a professional specialising in supporting autistic job seekers evaluated the answers, she found that some of the LLM’s answers weren’t helpful. For instance, when one user asked for advice on making friends, the chatbot suggested the user walk up to people and start talking with them. The problem, of course, is that a person with autism usually doesn’t feel comfortable doing that, Begel said.

It’s possible that a chatbot trained specifically to address the problems of people with autism might be able to avoid dispensing bad advice, but not everyone in the autism community is likely to embrace it, Begel said. While some might see it as a practical tool for supporting autistic workers, others see it as yet another instance of expecting people whose brains work a bit differently than most people to accommodate everyone else.

“There’s this huge debate over whose perspectives we privilege when we build technology without talking to people. Is this privileging the neurotypical perspective of ‘This is how I want people with autism to behave in front of me?’ Or is it privileging the person with autism’s wishes that ‘I want to behave the way I am,’ or ‘I want to get along and make sure others like me and don’t hate me?'”

At heart, it’s a question of whether autistic people are given a say in research that is intended to help them. It’s also an issue explored in another CHI paper, on which Begel is a co-author with Naba Rizvi and other researchers at the University of California, San Diego. In that study, researchers analyzed 142 papers published between 2016 and 2020 on developing robots to help autistic people. They found that 90% of this human-robot interaction research did not include the perspectives of people with autism. One result, Begel said, was the development of a lot of assistive technology that people with autism didn’t necessarily want, while some of their needs went unaddressed.

“We noticed, for instance, that most of the interactive robots designed for people with autism were nonhuman, such as dinosaurs or dogs,” Begel said. “Are people with autism so deficient in their own humanity that they don’t deserve humanoid robots?”

Technology can certainly contribute to a better understanding of how people with and without autism interact. For instance, Begel is collaborating with colleagues at the University of Maryland on a project using AI to analyze conversations between these two groups. The AI can help identify gaps in understanding by either or both of the speakers that could result in jokes falling flat or creating the perception that someone is being dishonest. The technology could also help speakers prevent or repair these conversational problems, Begel said, and the researchers are seeking input from a large group of people with autism to get their opinion on the kind of help they would like to see.

“We’ve built a video calling tool to which we’ve attached this AI,” said Begel, who has also developed an Autism Advisory Board to ensure that people with autism have a say in which projects his lab should pursue. “One possible intervention might be a button on this tool that says ‘Sorry, I didn’t hear you. Can you please repeat your question?’ when I don’t feel like saying that out loud. Or maybe there’s a button that says, ‘I don’t understand.’ Or even a tool that could summarize the meeting agenda so you can help orient your teammates when you say, ‘I’d like to go back to the first topic we spoke about.'”