The phrase is “vegetative electron microscopy”
And it looks more like a machine translation error than anything else. Per the article, there was a dataset with two instances of the phrase being created from bad OCR. Then, more recently, somehow the bad phrase got associated with a typo: in Farsi, the words “scanning” and “vegetative” are extremely similar. Thus, when some Iranian authors wanted to translate their paper to English, they used an LLM, and it decided that since “vegetative electron microscope” was apparently a valid term (since it was included in its training data), that’s what they meant.
It’s not that the entire papers were being invented from nothing by Chatgpt.
It’s been found in many papers though. Do they all have such excuses?
It probably is decently common to translate articles using ChatGPT as it is a large language model so that does seem likely
The lede is buried deep in this one. Yeah, these dumb LLMs got bad training data that persists to this day, but more concerning is the fact that some scientists are relying upon LLMs to write their papers. This is literally the way scientists communicate their findings to other scientists, lawmakers, and the public, and they’re using fucking predictive text like it has cognition and knows anything.
Sure, most (all?) of those papers got retracted, but those are just the ones that got caught. How many more are lurking out there with garbage claims fabricated by a chatbot?
Thankfully, science will inevitably sus those papers out eventually, as it always does, but it’s shameful that any scientist would be so fatuous to put out a paper written by a dumb bot. You’re the experts. Write your own goddamn papers.
They were translating them not actually writing them like obviously it should have been caught by reviewers but that’s not nearly as bad
In some cases, it’s people who’ve done the research and written the paper who then use an LLM to give it a final polish. Often, it’s people who are writing in a non-native language.
Doesn’t make it good or right, but adds some context.
Sure, and I’m sympathetic to the baffling difficulties of English, but use Google Translate and ask someone who’s more fluent for help with the final polish (as a single suggestion). Trusting your work, trusting science to an LLM is lunacy.
It might be hard for them to find someone who is both fluent in english AND knows the field well enough to know vegetative electron microscopy is not a thing. Most universities have one general translation help service and science has a lot of field-specific weird terms.
Google translate is using the same approach like an LLM.
https://en.wikipedia.org/wiki/Google_Translate
https://en.wikipedia.org/wiki/Neural_machine_translationSo is DeepL
https://en.wikipedia.org/wiki/DeepL_Translator
And before they were using neural network approaches they used statistical approaches, which are subject to the same errors as a result of bad training data.
Thankfully, science will inevitably sus those papers out eventually, as it always does,
In the future, all search engines will have an option to ignore any results from 2022-20xx, the era of AI slop.
Its the immediate takeaway i made from the headline, so i dont feel like its buried deep
Don’t use fucking AI to write scientific papers and the problem is solved. Wtf.
Let’s delve into the issue
So, all those research papers were written by AI? Huh.
No, they were not. AI was probably used for translation.