'ChatGPT Detector' Catches AI-Generated Papers With … – Slashdot

Want to read Slashdot from your mobile device? Point it at m.slashdot.org and keep reading!




The Fine Print: The following comments are owned by whoever posted them. We are not responsible for them in any way.
… on HuggingFace and use that instead locally. There are thousands of them.
Training ChatGPT detectors will only deal with low-effort students. Heck, not even all, because there’s a wide range of public-facing LLMs out there already.
The reason everyone is jumping on the ChatGPT bandwagon is because the quality of its training is vastly better than anybody else’s. For evidence, compare answers from ChatGPT against Google Bard, you’ll quickly see the difference. Sure, there might be all kinds of LLMs out there that you can run yourself, but just because it’s an LLM doesn’t mean it was trained well.
Do not put your faith in false gods. “small set of features.” Harrumph.
The major false gods disagree, and demand that you have more faith in the lesser false gods.
According to the article referenced in the story, it misclassified 21% of human-generated paragraphs in one set. Thus it attributed 1 out of every 5 paragraphs written by humans as having been written by AI. Out of 50 entire documents written by humans, it misclassified 6 of them, or 12% false positives. That’s pretty bad given such a large corpus of text to analyze (an entire document).
With such massive false positives I really don’t know what the point in this is, or how useful it can be.
https://www.sciencedirect.com/… [sciencedirect.com]
And it’s aiming at the wrong target. It should be targeting fraudulent papers. If they’re accurate, what does it matter who wrote them.
True, at first glance I thought this was about papers written by students in which case it makes sense since the idea is a students grasp of material but in the case of peer-reviewed scientific literature I think it’s good to disclose AI was used but as you said, facts are the facts and this could be a tool to help format and fill extrapolate. At the end of the day it’s the author(s) reputation when they sign off on the final product.

True, at first glance I thought this was about papers written by students in which case it makes sense since the idea is a students grasp of material but in the case of peer-reviewed scientific literature I think it’s good to disclose AI was used but as you said, facts are the facts and this could be a tool to help format and fill extrapolate. At the end of the day it’s the author(s) reputation when they sign off on the final product.

True, at first glance I thought this was about papers written by students in which case it makes sense since the idea is a students grasp of material but in the case of peer-reviewed scientific literature I think it’s good to disclose AI was used but as you said, facts are the facts and this could be a tool to help format and fill extrapolate. At the end of the day it’s the author(s) reputation when they sign off on the final product.
Yeah, this seems like an odd target.
I suppose they’re worried about journals getting flooded with AI generated papers, but if your journal has an outside chance of accepting such a paper maybe your journal needs higher standards altogether (and what’s the motive for submitting if it’s always rejected).
And I think that scientific literature is one of the places where LLMs have a legitimate important use case. A lot of researchers don’t speak English as a first language, and even for some who do you couldn’t
They converge. A completely human written text will have plenty of spelling, grammar and punctuation errors. Even a simple AI like Word’s spell check function will eliminate 90% of those problems, which is a good thing. Eventually people learn where the machine corrects them and will correct themselves accordingly. As technology adapts, so do humans, if the technology doesn’t adapt, the humans will adapt themselves to the contrivances of the machine as long as the machine is creating a value add.
LLM (not AI
That’s more than a little optimistic. There are very real limits to what these things can do that no amount of training, regardless of quality, can overcome. Unrealistic expectations can be dangerous, as we’ve already seen.
If you’re improving at all, it’s because you’re being more deliberate about your code and comments. You’re trying to use your code and comments to communicate an idea, which is absolutely the right attitude to have, though it’s not one you often see.
If that’s what you need, that’s fine
Exactly my thought as well. They have the criteria for this backwards. The question isn’t, “Is this written by an AI?” The question is, “Is this written by a human?” Their own numbers say they’re getting that wrong about 10% of the time.
Why is that important? Because if you’re using it to detect fraud it’s going to sound the alarm 10% of the time, even when there’s absolutely zero fraud. That’s way too many false positives when you consider people could lose their careers or grades based on this outpu
Its even worse than that say its only 1% of papers are picked up as fraud but you write say 100 papers, so your chances of being accused of cheating are 63%. Or put it another way out of every 100 honest papers submitted, there is a 63% chance you are going to falsely accuse someone.
If its 10% your chances are 99.997% that you going to be accuses of cheating in one of your papers.
Sure you can make it so the consequences of “cheating” low, but then why bother at all submit your AI generated paper and they s
Algorithms must halt in finite time.
I guess it’s a couple of decades ago that the university I was at got excited and told us to use this wondrous new program to check for plagiarism.
I tried it.
Every single thing that it flagged was wrong.
Most of what it flagged were properly done citations!
In disbelief, I fed it a working paper of my own that had been on the web for several years, on multiple websites.
It saw no problem . . .
Moving to today, I see no reason to expect this type of check to work. Rather, it will start alarms race as we saw fou
Turnitin allows you to exclude citations. I find turnitin useful for detecting recycled papers, which I would miss. But I don’t need it to detect plagiarism from professional sources. Those stick out like a sore thumb.
LLMs provide a different challenge, but so far I have done alright in detecting them. The thing is, I can grade these as I want, without calling out the students. LLM papers are well written but with repetitive grammatical structures, start with a very clear thesis statement, and use the phras
If you find turnitin useful you’re at absolutely minimum willfully and maliciously negligent towards your students. Their entire business model relies on manufacturing accused cheaters even where there are none in order to justify their continued expense and the paranoia of teachers. Every time it flagged one of my students’ papers I would feed it something I made up on the spot and get flagged too.

If you find turnitin useful you’re at absolutely minimum willfully and maliciously negligent towards your students. Their entire business model relies on manufacturing accused cheaters even where there are none in order to justify their continued expense and the paranoia of teachers. Every time it flagged one of my students’ papers I would feed it something I made up on the spot and get flagged too.

If you find turnitin useful you’re at absolutely minimum willfully and maliciously negligent towards your students. Their entire business model relies on manufacturing accused cheaters even where there are none in order to justify their continued expense and the paranoia of teachers. Every time it flagged one of my students’ papers I would feed it something I made up on the spot and get flagged too.
Did you see what I wrote? I use it primarily to catch students who recycle papers among one another.
Also your anecdote is just that. If a student fails to use my essay template, and does not quote sources, Turnitin generally does not flag much if anything as plagiarism. I tell you that with years of experience.
Also, if Turnitin flags something, you do know that it is within your agency to check if it is correct, and disregard it – non?
Also, Turnitin accuses nobody of plagiarism. Instructors do, and hopeful
I *want* to say that it was turnitin, but this was twenty years ago . . .
I don’t think it had much, if anything, in the way of options. For that matter, I’m not sure it had *any*, and it may well have been in beta.]
And I *defintiely* caught things that it didn’t. Not (quite) as flagrant as a friend who found “elsewhere in this issue”, but still.
Generally, I tossed suspect phrases into google with quotes around them, and *bang!*

Generally, I tossed suspect phrases into google with quotes around them, and *bang!*

Generally, I tossed suspect phrases into google with quotes around them, and *bang!*
My trick was to google sentences with well-used semicolons, a skill American undergrads in general have never mastered.

LLMs provide a different challenge, but so far I have done alright in detecting them. The thing is, I can grade these as I want, without calling out the students. LLM papers are well written but with repetitive grammatical structures, start with a very clear thesis statement, and use the phrase “delve into” quite a bit.

LLMs provide a different challenge, but so far I have done alright in detecting them. The thing is, I can grade these as I want, without calling out the students. LLM papers are well written but with repetitive grammatical structures, start with a very clear thesis statement, and use the phrase “delve into” quite a bit.
Remember a few months ago reading LLM generated sci-fi stories and couldn’t get over how much generic phrases akin to “unlike anything we have encountered” appeared. Then I re-watched the first few seasons of STNG.. Started getting on my nerves to see the same language used so often.

However, they are also superficial, and rarely quote the authors I require. And if they do, they don’t quote the right editions.

However, they are also superficial, and rarely quote the authors I require. And if they do, they don’t quote the right editions.
Who pays for the right editions? There’s a long-term swindle in the United States textbook market in which publishers seek an excuse to publish a new edition of a nonfree textbook solely to deter resale of used copies of old editions.[1] The usual tactic to render old editions useless is to reformat the text a bit, so as to invalidate page numbers, and reorder the exercises.
[1] Samuel T. Loch and Joshua D. Van Mater. “The Efficacy of Planned Obsolescence Strategies in the College Textbook Market” [unc.edu]. Universit
That’s hilarious. You have no idea what I teach, and then imply that I am part of the textbook swindle. My 100 level courses pay about $15 USD for texts, and I provide a few for free. My up-level courses come to about $30-50 USD, depending on the course.
But you hit on an interesting point: it’s not the administrators and loan guarantees that have caused the price of university education to sky-rocket; no, it’s those evil professors and their demand for standard and usable editions.
When I saw the headline, my first thought was that “unprecedented success” is a pretty low bar in this field.
That’s a feature, not a bug. This and other frauds like “turnitin” both fundamentally rely on the boogeyman always existing. If you don’t catch enough cheaters you need to manufacture them.
What is worse?
It is very likely to punish better works. The GPT training depended on high quality writing, and admittedly it shows in the output. But who else write high quality English with proper grammar and large vocabulary? Those who are actually good in their field.
Those who write sloppily, and make a lot of mistakes will be spared.
And those otherwise A+ students will have to take on legal fights to prove innocence.
Just to be complete, the GPT “conversion” of my own writing (fortunately I won’t be mist
That’s going to be the next thing: an AI that will alter your paper so it passes the AI detector AI.
AIIIIIEEEEEEE!!!
It doesn’t matter if you can detect AI generated text 100% of the time if you also falsely catch ANY completely human text. Doesn’t matter if they get it down to 1% false positives – any false positive makes this tool 100% unreliable. Not to mention that who care if there was an AI that wrote it as long as it is accurate? Because we have to give some specific person credit? Fuck them. Fuck their credit. Fuck people’s entitledness. If it’s true, accuracte, and real, then move the fuck on.
Accused of all manner of cheating by fundamentally flawed snake oil.
False positive rate is all you need to pay attention to .. this indicates how many people will be falsely accused – their paper scored at 0%, their qualification refused – when they did nothing wrong
There may be more comments in this discussion. Without JavaScript enabled, you might want to turn on Classic Discussion System in your preferences instead.
OpenAI Says ChatGPT Has 100 Million Weekly Users
Huawei and Tencent Spearhead China’s Hold on Cybersecurity Patents
There is no opinion so absurd that some philosopher will not express it. — Marcus Tullius Cicero, “Ad familiares”

source

Jesse
https://playwithchatgtp.com