Welcome to DU!
The truly grassroots left-of-center political community where regular people, not algorithms, drive the discussions and set the standards.
Join the community:
Create a free account
Support DU (and get rid of ads!):
Become a Star Member
Latest Breaking News
Editorials & Other Articles
General Discussion
The DU Lounge
All Forums
Issue Forums
Culture Forums
Alliance Forums
Region Forums
Support Forums
Help & Search
General Discussion
Related: Editorials & Other Articles, Issue Forums, Alliance Forums, Region ForumsMeet the academics refusing to use generative AI (Nature, 5/5/26)
https://www.nature.com/articles/d41586-026-00508-wDanielle Crowley is getting tired of people telling her to use generative artificial intelligence (genAI). As a marine zoologist at Bangor University, UK, she says that she is pretty much the only PhD student in her cohort who does not use it. She has seen colleagues use genAI tools for coding and for getting the tone of e-mails right. On one occasion, she was even encouraged by a lecturer to use it to generate a conference poster.
She says her colleagues are often surprised to hear she hasnt tried it and have suggested she uses it for applications such as coding. Ive had a lot of people go like oh but you have to use it, she recalls. But Crowley has her reasons. She has concerns about the ethics of copyright, what she calls a lack of transparency from companies about how theyre using the data, the environmental effects of AI tools and the accuracy of what genAI models spit out.
-snip-
And verifying AI-generated information often defeats the purpose of using the tool for efficiency, say cynics. Tanisha Jowsey, a social scientist at Bond University in Robina, Australia, says that as a designated AI champion of the faculty, she is supposed to appraise models, work out what theyre good at and suggest how the faculty could be using them. But ironically, she stresses, checking them creates even more work.
She says that 95% of the time it would be quicker for me to just do the thing myself than get the tool to do it and then have to check whether or not its done it right. She also finds that its an ineffective tool for qualitative research: a view she expressed in a co-authored commentary article that was posted on the preprint platform SSRN.
-snip-
She says her colleagues are often surprised to hear she hasnt tried it and have suggested she uses it for applications such as coding. Ive had a lot of people go like oh but you have to use it, she recalls. But Crowley has her reasons. She has concerns about the ethics of copyright, what she calls a lack of transparency from companies about how theyre using the data, the environmental effects of AI tools and the accuracy of what genAI models spit out.
-snip-
And verifying AI-generated information often defeats the purpose of using the tool for efficiency, say cynics. Tanisha Jowsey, a social scientist at Bond University in Robina, Australia, says that as a designated AI champion of the faculty, she is supposed to appraise models, work out what theyre good at and suggest how the faculty could be using them. But ironically, she stresses, checking them creates even more work.
She says that 95% of the time it would be quicker for me to just do the thing myself than get the tool to do it and then have to check whether or not its done it right. She also finds that its an ineffective tool for qualitative research: a view she expressed in a co-authored commentary article that was posted on the preprint platform SSRN.
-snip-
Much more at the link.
I especially liked what Michaela Socolof, a psycholinguist at MIT, said about generative AI use. The first of her objections to AI that the article quotes is that genAI is trained on work stolen from writers and artists. But a later paragraph has her explaining that even if genAI models were legally and ethically trained, she still feels AI is "so corrosive to critical-thinking ability" that she would never use it.
4 replies
= new reply since forum marked as read
Highlight:
NoneDon't highlight anything
5 newestHighlight 5 most recent replies
Meet the academics refusing to use generative AI (Nature, 5/5/26) (Original Post)
highplainsdem
12 hrs ago
OP
mr715
(4,067 posts)1. There are many of my colleagues that are using it
In fact, there have been directives from admin to develop strategies to rely on it more...
Iris
(16,888 posts)3. This is so weird to me. Like, just "find ways to use it"?
Nothing more specific than that?
We had a panel of bankers at my institution talking about AI last year and one man said he had a post it on his computer reminding him to "use AI"
I just don't get it
It is along the lines of "We acknowledge that new tools are available and we encourage department directors to find way to use them to improve xyz while maintaining scientific integrity"
Very much a vague use it, but don't use it too much, imperative.
It is a troubling thing to see in a science department.
SheltieLover
(81,559 posts)2. Good for them!