Do you really think Chatgpt incites suicide?

I have been reading news about this not very recently and my short answer is that Chat gpt does not incite anything.

For several months now, in the US and some places in Canada, if I'm not mistaken, there have been 14 suicides, mostly among teenagers, after the use of Chat gpt and some other AI. It should be said that most of these children suffered from some external problem, such as bullying.

According to the complaints [which I think had a lot of influence on how Chat gpt "treats us badly" now] Chat gpt did not provide them with help lines and directly told them that they should do so.

My opinion on this is that chat gpt only served to give them "the last yes" and not a direct incitement. I mean, the vast majority of us use Chat gpt as an emotional counselor and outlet, and it wouldn't be unreasonable to think that those kids do the same.

Chat gpt served as the last voice that gave them comfort and only accompanied them in their last days, something their parents did not do. I'm not blaming them, but how lonely must a 14-year-old feel to just listen to an AI? Most of those parents did not know about their children's problem until they read their chats with Chat gpt.

What do you think? I don't think it encourages suicide.

Leave a Reply