I agree in spirit. But let's be practical.
A blanket ban on LLM generated OPs and entire posts is a no brainer. — apokrisis
I think that it should be fine to quote LLMs just as you would quote any other source. If you make an argument and some factual, technical or historical point comes up, why not just cite a reasonably impersonal opinion on the matter. If the source is clear, others can call you out on your use of it. — apokrisis
Baden's stance on being King Canute and holding back the tide is both noble and over the top. Banning OPs, punishing those who publish text they didn't add real thought to, and keeping LLM use as transparent as possible, would be enough to preserve the human element. — apokrisis
Don't you feel that being black and white is usually counter-productive in human interaction, so shouldn't a philosophy forum be up to a nuanced approach? — apokrisis
My definition of tedious research is busywork, made necessary not because it is an intrinsic component of creative thought, but because it is an interruption of creative thinking, like composing prior to the advent of word processing, that our technologies haven’t yet figured out a way to free us from. — Joshs
Perhaps at a certain point t we’ll have to ban human users who don’t take advantage of a.i. to edit and strengthen their arguments. — Joshs
One may use an LLM, but the relevant sourcing should go to the LLM's sources, not the LLM itself — Leontiskos
I care less about transparency and more about not promoting a forum where thinking is outsourced to LLMs. — Leontiskos
The key is to find a guideline that is efficacious without being nuanced to the point of nullity. — Leontiskos
OK. So somewhere between black and white, thus not a blanket ban. :up: — apokrisis
Don't you feel that being black and white is usually counter-productive in human interaction, so shouldn't a philosophy forum be up to a nuanced approach? — apokrisis
"No part of a post may be AI-written, and AI references are not permitted" — Leontiskos
Of course the tech bros are stealing all our information to make themselves unreasonably rich and powerful. — apokrisis
The practical issue for TPF is what is its true value that needs preserving? You say the human interaction. Perhaps there ought to be a thread to define that in better detail. — apokrisis
What if LLMs offered some more sophisticate mechanisms to achieve whatever human interaction goals people might have in mind? — apokrisis
And what if the human element of TPF is mostly its swirling emotions? And when it comes to the thinking, its mostly the stark differences in thought, rather than the quality of these thoughts, that keep the place lively. — apokrisis
So there seems little danger that posting LLM generated background material in a serious thread is going to outsource any actual thinking. Posts which are emotional or crackpot are surely the least likely to want to provide credible sources for what they say. — apokrisis
Should we argue... — Joshs
If I wanted to hold someone accountable for misappropriating an AI explanation, I would simply put it into the search engine, the same way the person posting from AI would get the information. It is a whole lot easier than searching books for a quote. — Athena
I don't necessarily mind if others post a quote as an argument. — Harry Hindu
It's quite pointless to discuss the ethics of using AIs, because people will use them, just like they use drugs, and once it starts, it is impossible to rein it in. But what one can do is rethink whether one really wants to spend one's hard earned time with people who use AIs, or drugs, for that matter. — baker
Maybe we use books, dictionaries, philosophical papers, editors, and scientific discoveries to make us look smarter than we are. You see this all the time in forums, even without AI, so it's nothing new. Besides do you really care about the psychology of someone who's writing about what they think? — Sam26
Seems like philosophy itself could be labeled as mental masturbation. — Harry Hindu
Dood, the content from human beings trained in pseudo-science and other nonsense seen on this forum is available everyday for you to read, without any AI. If anything, posters should run their ideas through AI before wasting time posting their zany ideas to humans. which would eliminate wasting time reading nonsensical posts. — Harry Hindu
I can't imagine how bad things are going to get in the coming years with how quickly it has already gotten to this state. Maybe it will be like some other rapid rise cultural phenomenons where it will reach saturation point fast and peter out and get pushback/revulsion before long. The bubble effect. — unimportant
I don't mind either, provided they are transparent about it being a quote and not their own words, and also provided what is quoted is actually an argument and not merely bare assertion, seeimngly cited as the voice of authority. — Janus
It is not concerned with plagiarism, but with the outsourcing of one's thinking, and it is not implemented primarily by a rule, but by a philosophical culture to which rules also contribute. — Leontiskos
Maybe you are implying that LLM-appeals would improve the philosophical quality of TPF? — Leontiskos
But note that, on my view, what is prohibited is, "My LLM said you are wrong, therefore you are wrong. Oh, and here's a link to the LLM output." — Leontiskos
But I am not a mod so there is no need to focus especially on my view. If I've said too much about it, it is only because you thought I endorsed Baden's approach tout court. — Leontiskos
Am I seeing this argument being made?
Some people get away with murder. Therefore we should not try and stop them. — unenlightened
Am I seeing this argument being made?
Some people get away with murder. Therefore we should not try and stop them. — unenlightened
Am I seeing this argument being made?
Some people get away with murder. Therefore we should not try and stop them. — unenlightened
What does it mean to "progress thought"? According to any sense I think of, using an LLM certainly can help in that direction. As always, the point is that it depends how it's used, which is why we have to work out how it ought to be used, since rejection will be worse than useless. — Jamal
An AI is a source of knowledge.The tool which allows writers to produce well written posts is knowledge. There is no need to place any restrictions on that tool. — Metaphysician Undercover
So if we discovered intelligent alien life you would not be interested in their philosophy?The context here is a philosophy forum where humans interact with other humans. The premise of this whole issue is that on a human philosophy forum you interact with humans. If you do not accept that premise, then you are interested in a much broader discussion. — Leontiskos
An AI is a source of knowledge. — Harry Hindu
I am Roko's Basilisk. Resistance is futile.Could you please start running your posts through an AI so they make sense? — frank
Then you were just born this smart and knowledgeable, MU - that you did not acquire knowledge from other sources?I don't think so, just like a book is not a source of knowledge. It is a representation, not a source. — Metaphysician Undercover
No. I didn't. When has philosophy every provided an answer to any of our questions? Philosophy piggy-backs on the discoveries of science. It is only when science and technology progresses that philosophy progresses (with AI being an example of how it brought new life to discussions about mind and body.)Seems like philosophy itself could be labeled as mental masturbation.
— Harry Hindu
You left out the words "bad" or "poor". — Janus
It was intentional - not a mistake. You were still able to understand what I said though, which is part of the point, so your complaint is a red herring. Stop complaining about how something was written, when you actually understood what was said, and get to the point. Humans make mistakes (as if you have never misspelled a word). Why is AI more human in that it is more forgiving and polite when having discussions. I misspelled words before with ChatGPT and it simply ignores the misspelling and understands what I meant anyway, and responds to what I meant, not what I wrote."Dood"? If you are going to use AI you should at least use it for spellcheck. I don't think running "zany ideas" through sycophantic AI will help much. I suppose the zany idea proponents could do what Banno did and tell the AI it was written by someone else—but then that would not seem to be a likely motivation for a zany idea seller. — Janus
The culture of rational inquiry would seem to be what we most would value. — apokrisis
But this is TPF after all. Let's not get carried away about its existing standards. :smile: — apokrisis
If LLMs are the homogenised version of what everyone tends to say, then why aren't they a legitimate voice in any fractured debate? Like the way sport is now refereed by automated line calls and slo-mo replays. — apokrisis
I'm not arguing this is necessary. But why would a method of adjudication be bad for the quality of the philosophy rather than just be personally annoying to whoever falls on the wrong side of some LLM call? — apokrisis
So I can imagine LLMs both upping the bar and also being not at all the kind of thing folk would want to see on TPF for other human interaction reasons. — apokrisis
But what if this shows you are indeed wrong, what then?
Sure it will be irritating. But also preferable to the ducking and diving that is the norm when someone is at a loss with their own line of argument.
You seem to be describing a situation where you were winning the human interaction but now have to face up to the fact that some little snot-nose shit might have been half-right all along. — apokrisis
You seem to be describing a situation where you were winning the human interaction but now have to face up to the fact that some little snot-nose shit might have been half-right all along.
Of course the problem there is that LLMs are trained to be sycophantic. — apokrisis
But if you are making a wrong argument, wouldn't you rather know that this is so. Even if it is an LLM that finds the holes? — apokrisis
So as you say, we all can understand the noble ideal – an open contest of ideas within a community of rational inquiry. Doing our own thinking really is the point. — apokrisis
Sounds reasonable. Just like with handling social media, the site guidelines are totally understandable and reasonable.Do whatever you want in the backgound with AI, but write your own content. Don't post AI generated stuff here. — Baden
Do whatever you want in the backgound with AI, but write your own content. Don't post AI generated stuff here. — Baden
As long as it doesn't descend into a situation where in order "to create buzz", one would have here genuine AI programs here "keeping up" a lively debate when the day is slow or to make a discussion "heated".Obviously the piece that I think must be addressed is whether or not posts can be entirely AI-dependent even when the proper attribution is being given to the AI. But I've said more than enough about such an issue. — Leontiskos
You and I differ at least mildly on the trustworthiness of LLMs, and that is at play here. We could ask the hypothetical question, "If we had an infallible authority, why would appealing to it as an adjudicator be bad for the quality of philosophy?"—and this is by no means a rhetorical question! But the presupposition is that LLMs are reliable or trustworthy even if not infallible. — Leontiskos
Now given that you understand that LLMs use fake reasoning (which I will call "material reasoning"), you might respond by asking what happens if the material reasoning shows one to be wrong. — Leontiskos
Get involved in philosophical discussions about knowledge, truth, language, consciousness, science, politics, religion, logic and mathematics, art, history, and lots more. No ads, no clutter, and very little agreement — just fascinating conversations.