On average, my discussions with chatGPT are more pleasant and insightful than the ones I have with real humans
The best conversations I still have are with real people, but those are rare. With ChatGPT, I reliably have good conversations, whereas with people, it’s hit or miss, usually miss.
What AI does better:
It’s willing to discuss esoteric topics. Most humans prefer to talk about people and events.
It’s not driven by emotions or personal bias.
It doesn’t make mean, snide, sarcastic, ad hominem, or strawman responses.
It understands and responds to my actual view, even from a vague description, whereas humans often misunderstand me and argue against views I don’t hold.
It tells me when I’m wrong but without being a jerk about it.
Another noteworthy point is that I’m very likely on the autistic spectrum, and my mind works differently than the average person’s, which probably explains, in part, why I struggle to maintain interest with human-to-human interactions.
Ur just training urself to have chatgpt's bias. We will soon live in a world where you wont have to be exposed to opinions you disagree with. Tom Scott has a yt vid on why this is a bad idea.
As long as you're still engaging with real humans regularly, I think that it's good to learn from ChatGPT. It gets most general knowledge things right. I wouldn't depend on it for anything too technical, and certainly not for medical advice. It is very hit or miss for things like drug interactions.
If you're enjoying the experience, it's not much different than watching a show or playing a game, IMHO. Just don't become dependent on it for all social interaction.
As for the jerks on here, I always recommend aggressive use of the block button. Don't waste time and energy on them. There's a lot of kind and decent people here, filter your feed for them.
As for the jerks on here, I always recommend aggressive use of the block button. Don’t waste time and energy on them. There’s a lot of kind and decent people here, filter your feed for them.
My blocklist is around 500 users long and grows every day. I do it for the pettiest reasons but it does, infact work. When I make a thread such as this one, I occasionally log out to see the replies I've gotten from blocked users and more often than not (but not always) they're the kind of messages I'd block them again for. Not to create and echo-chamber but to weed out the assholes.
It sounds like you are training yourself to be a poor communicator, abandoning any effort to become more understandable to actual humans.
Based on what? That seems like a rather unwarranted assumption to me. My English vocabulary and grammar have never been better, and since I can now also talk to it instead of typing, my spoken English is much clearer and more confident as well.
You say yourself that you use the vaguest descriptions when talking to the bot and that it fills in the blanks for you. This is not a good way to practice speaking with human beings.
The fact that you assumed I was talking about grammar is indicative of the problem. You clearly dislike others assuming you are talking about something you are not talking about, yet you do it yourself. That's because misunderstandings are normal and learning to deal with them is an essential part of good communication.
Have you ever tried inputting sentences that you've said to humans to see if the chatbot understand your point better? That might be an interesting experiment if you haven't tried it already. If you have, do you have an example of how it did better than the human?
I'm kinda amazed that it can understand your accent better than humans too. This implies Chatbots could be a great tool for people trying to perfect their 2nd language.
A couple of times, yes, but more often it's the other way around. I input messages from other users into ChatGPT to help me extract the key argument and make sure I’m responding to what they’re actually saying, rather than what I think they’re saying. Especially when people write really long replies.
The reason I know ChatGPT understands me so well is from the voice chats we've had. Usually, we’re discussing some deep, philosophical idea, and then a new thought pops into my mind. I try to explain it to ChatGPT, but as I'm speaking, I notice how difficult it is to put my idea into words. I often find myself starting a sentence without knowing how to finish it, or I talk myself into a dead-end.
Now, the way ChatGPT usually responds is by just summarizing what I said rather than elaborating on it. But while listening to that summary, I often think, "Yes, that’s exactly what I meant," or, "Damn, that was well put, I need to write that down."
So what you're saying if I'm reading right is chatbots are great for bouncing ideas off of to help you explain yourself better as well as helping you gather your own thoughts. im a bit curious about your philosophy chats.
When you have a philosophical discussion does the chatbot summarize your thoughts in its responses or is it more humanlike maybe disagreeing/bringing up things you hadn't thought of like a person might? (I've never used one).
I've read this text. It's a good piece, but unrelated to what OP is talking about.
The text boils down to "people who believe that LLMs are smart do so for the same reasons as people who believe that mentalists can read minds do." OP is not saying anything remotely close to that; instead, they're saying that LLMs lead to pleasing and insightful conversations in their experience.
Idk, I think that article is a bit hyperbolic and self serving for validation of the writers and the readers to pander their own intelligence above others. The lengthy exposition on cold reading is plain filler material for the topic and yet it goes on. ChatGPT and LLM have been a thing for a while now and I doubt anyone technically literate believes it to be AI as in an actual individual entity. It's an interactive question-response machine that summarises what it knows about your query in flowing language or even formatted as lists or tables or whatever by your request. Yes, it has deep deep flaws with holes and hallucinations, but for reasonable expectations it is brilliant. Just like a computer or the software for it, it can do what it can do. Nobody expects a word processor or image editor or musical notation software to do more than what it can do. Even the world's most regarded encyclopedia have limits, both printed and interactive media alike. So I don't see why people feel the need to keep in patting themselves on the back of how clever they are by pointing out that LLM are in fact not a real world mystical oracle that knows everything. Maybe because they themselves were the once thinking it was and now they are overcompensating to save face.
I talk with chat gpt too sometimes and I get where you are coming from. However it’s not always right either. It says it was updated in September but still refuses to commit to memory that Trump was convicted 34 times earlier this year. Why is that?
It could respond in other ways if it was trained to do so. My first local model was interesting as I changed its profile to have a more dark and sarcastic tone, and it was funny to see it balance that instruction with the core mode to be friendly and helpful.
The point is, current levels of LLMs are just telling you what you want to hear. But maybe that's useful as a sounding board for your own thoughts. Just remember its limitations.
Regardless of how far AI tech goes, the human-AI relationship is something we need to pay attention to. People will find it a good tool like OP, but it can be easy to get sucked into thinking it's more than it is and becoming a problem.
Autism and social unawareness may be a factor. But points you made like the snide remarks one may also indicate that you're having these conversations with assholes.
Well, it's a self-selecting group of people. I can't comment on the ones who don't respond to me, only on the ones who do and for some reason the amount of assholes seems to be quite high in that group. I just don't feel like it's warranted. While I do have a tendency to make controversial comments I still try and be civil about it and I don't understand the need to be such a dick about it even if someone disagrees with me. I welcome disagreement and are more than willing to talk about it as long as it's done in good faith.
It's a mirror. I use it a lot for searching and summarizing. Most of its responses are heavily influenced by how you talk to it. You can even make it back up terrible assumptions with enough brute force.
My impressions are completely different from yours, but that's likely due
It's really easy to interpret LLM output as assumptions (i.e. "to vomit certainty"), something that I outright despise.
I used Gemini a fair bit more than ChatGPT, and Gemini is trained with a belittling tone.
Even then, I know which sort of people you're talking about, and... yeah, I hate a lot of those things too. In fact, one of your bullet points ("it understands and responds...") is what prompted me to leave Twitter and then Reddit.
It's funny how despite it not actually understanding anything per-se, it can still repeat me back my idea that I just sloppily told it in broken english and it does this better than I ever could. Alternatively I could spend 45 minutes laying out my view as clearly as I can on a online forum only to be faced with a flood of replies from people that clearly did not understand the point I was trying to make.
I think that the key here are implicatures - things that implied or suggested without being explicitly said, often relying on context to tell apart. It's situations like someone telling another person "it's cold out there", that in the context might be interpreted as "we're going out so I suggest you to wear warm clothes" or "please close the window for me".
LLMs model well the grammatical layer of a language, and struggle with the semantic layer (superficial meaning), but they don't even try to model the pragmatic layer (deep meaning - where implicatures are). As such they will "interpret" everything that you say literally, instead of going out of their way to misunderstand you.
On the other hand, most people use implicatures all the time, and expect others to be using them all the time. Even when there's none (I call this a "ghost implicature", dunno if there's some academic name). And since written communication already prevents us from seeing some contextual clues that someone's utterance is not to be taken literally, there's a biiiig window for misunderstanding.
[Sorry for nerding out about Linguistics. I can't help it.]
I know a bit more than normal people would about the inner workings of LLMs. I still occasionally have a conversation with it, like I would with a therapist, perhapse less open and all but still. Do I know it's nothing more than a talking parrot? Yes. Do I still feel like I'm talking to a real person without judgement? Yes. And I can use that from time to time.