respeckKnuckles
respeckKnuckles t1_jc8xver wrote
"Not using gpt4" is going to be in all NLP conference paper reviews for the next six months.
respeckKnuckles t1_j453ktq wrote
Reply to comment by happygilmore001 in [D] Has ML become synonymous with AI? by Valachio
lol no
respeckKnuckles t1_j1x00bh wrote
Reply to comment by Liwet_SJNC in [P] Can you distinguish AI-generated content from real art or literature? I made a little test! by Dicitur
Yeah we have that, at least. The problem is that the pandemic moved a lot of classes and assignments online. Whether it is their choice or not, a lot of professors are still having homework assignments (even tests) online, and on those you often will see prompts asking for short 100-word answers.
respeckKnuckles t1_j1vo5s7 wrote
Reply to comment by MrFlamingQueen in [P] Can you distinguish AI-generated content from real art or literature? I made a little test! by Dicitur
I've never seen empirical study demonstrating either: (1) professors can reliably differentiate between AI-generated text and a random B-earning or C-earning student's work, or (2) those "tools" you mention (probably you're talking about the huggingface GPT-2-based tool) can do that either.
You say "on some level", and I don't think anyone disagrees. An A-student's work, especially if we have prior examples from the student, can probably be distinguished from AI work. That's not the special case I'm concerned with.
respeckKnuckles t1_j1vl5dv wrote
Reply to comment by Ulfgardleo in [P] Can you distinguish AI-generated content from real art or literature? I made a little test! by Dicitur
Hence my original question to OP.
respeckKnuckles t1_j1vgrit wrote
Reply to comment by Dicitur in [P] Can you distinguish AI-generated content from real art or literature? I made a little test! by Dicitur
It'd be great if you could extend it to longer texts, like paragraph-lengths. A lot of these are recognizable quotes, so it throws off the reliability of the assessment a bit (especially if the people doing this might be, say, English professors).
respeckKnuckles t1_j1vg27f wrote
Reply to [P] Can you distinguish AI-generated content from real art or literature? I made a little test! by Dicitur
Please let us know when you get some reportable results on this. I'm having trouble convincing fellow professors that they should be concerned enough to modify their courses to avoid the inevitable cheating that will happen. But in a stunning display of high-level Dunning-Kruger, they are entirely confident they can always tell the difference between AI and human-generated text. Some data might help to open their eyes.
respeckKnuckles t1_j1vempm wrote
Reply to comment by Ulfgardleo in [P] Can you distinguish AI-generated content from real art or literature? I made a little test! by Dicitur
> you are asking humans to solve this task untrained, which is not the same as the human ability to distinguish the two.
This is exactly my point. There are two different research questions being addressed by the two different methods. One needs to be aware of which they're addressing.
> you are then also making it harder by phrasing the task in a way that makes it difficult for the human brain to solve it.
In studying human reasoning, sometimes this is exactly what you want. In fact, for some work in studying Type 1 vs. Type 2 reasoning, we actually make the task harder (e.g. by adding WM or attentional constraints) in order to elicit certain types of reasoning. You want to see how they will perform in conditions where they're not given help. Not every study is about how to maximize human performance. Again, you need to be aware of what your study design is actually meant to do.
respeckKnuckles t1_j1v66iq wrote
Reply to comment by dojoteef in [P] Can you distinguish AI-generated content from real art or literature? I made a little test! by Dicitur
You say it allows them to "better frame the task", but is your goal to have them maximize their accuracy, or to capture how well they can distinguish AI from human text in real-world conditions? If the latter, then this establishing of a "baseline" leads to a task with questionable ecological validity.
respeckKnuckles t1_j1v440q wrote
Reply to comment by dojoteef in [P] Can you distinguish AI-generated content from real art or literature? I made a little test! by Dicitur
I'm not sure how the side by side comparison answers the same research question. If they are told one is AI and the other isn't, the reasoning they use will be different. It's not so much "is this AI?" as it is "which is more AI-like?"
respeckKnuckles t1_j0apq5l wrote
Reply to comment by Purplekeyboard in [R] Talking About Large Language Models - Murray Shanahan 2022 by Singularian2501
I asked for an operationalizable, non-circular definition. These are neither.
> the state of knowing that you know something and can analyze it, look at it from different angles, change your mind about it given new information, and so on.
Can it be measured? Can it be detected in a measurable, objective way? How is this not simply circular: truly understanding is defined as truly knowing, and truly knowing is defined as truly understanding?
> Today's AI language models have lots of information contained within themselves, but they can only use this information to complete prompts, to add words to the end of a sequence of words you give them. They have no memory of what they've done, no ability to look at themselves, no viewpoints. There is understanding of the world contained within their model in a sense, but THEY don't understand anything, because there is no them at all, there is no operator there which can do anything but add more words to the end of the word chain.
This is the problem with the "argumentum ad qualia"; qualia is simply asserted as this non-measurable thing that "you just gotta feel, man", and then is supported by these assertions of what AI is not and never can be. And how do they back up those assertions? By saying it all reduces to qualia, of course. And they conveniently hide behind the non-falsifiable shell that their belief in qualia provides. It's exhausting.
respeckKnuckles t1_j0ajskv wrote
Reply to comment by leondz in [R] Talking About Large Language Models - Murray Shanahan 2022 by Singularian2501
Airplanes can fly on autopilot. Autopilot is part of the autopilot-using plane. Therefore, at least some airplanes can fly on their own.
respeckKnuckles t1_j0ajork wrote
Reply to comment by Purplekeyboard in [R] Talking About Large Language Models - Murray Shanahan 2022 by Singularian2501
> which actually understands the world.
Please define what it means to "actually understand" the world in an operationalizable, non-circular way.
respeckKnuckles t1_iywkhtq wrote
Reply to comment by AlmightySnoo in [D] NeurIPS 2022 Outstanding Paper modified results significantly in the camera ready by Even_Stay3387
Where did OP assume malicious intent?
respeckKnuckles t1_ixhedyv wrote
Reply to comment by Acceptable-Cress-374 in [D] Schmidhuber: LeCun's "5 best ideas 2012-22” are mostly from my lab, and older by RobbinDeBank
Uh...they do
respeckKnuckles t1_it8ovi2 wrote
Reply to comment by DigThatData in [D] Accurate blogs on machine learning? by likeamanyfacedgod
Is there a reason then that it's not common to see what the actual threshold is on graphs of AUC-ROC curves? It seems like it would be very helpful to have a little mark on the curve itself for when the threshold is 0.5, for example.
respeckKnuckles t1_it89f8r wrote
Reply to comment by hostilereplicator in [D] Accurate blogs on machine learning? by likeamanyfacedgod
Something I never quite understood---TPR and FPR are independent of each other, right? So then how is the plot of the AUC-ROC curve created? What if there are multiple parameters for which the FPR is the same value, but the TPR differs?
respeckKnuckles t1_irwn0vj wrote
Reply to comment by Chhatrapati_Shivaji in [D] Looking for some critiques on recent development of machine learning by fromnighttilldawn
NYU professor who published a few "pop-sciency" books on AI-related stuff. Like many in his generation, he got some attention for taking a contrarian stance on what current approaches to AI can do, and decided to go extremist with it. I'm not sure he's much more than a full-time angry twitterer now.
respeckKnuckles t1_irw4idm wrote
Reply to [D] Looking for some critiques on recent development of machine learning by fromnighttilldawn
Gary Marcus's twitter is a firehose of unwarranted pessimism, but occasionally he'll retweet or interact with a legitimate, balanced criticism.
respeckKnuckles t1_jca9td9 wrote
Reply to comment by YouAgainShmidhoobuh in [D] On research directions being "out of date" by redlow0992
Have you never been peer reviewed before?