What possible threats do you see in using AI?
I read about a few cases that sounded like something out of a horror story.
For example:
AI Model Refuses To Listen to Humans After Being Told to “Shut Down
Psychiatrist Poses as Teen – Finds AI Therapy Bots Encouraging Violence, Suicide, and Fake Intimacy
Since more people started using AI as a therapist, and having in mind all those Terminator and Black Mirror stories... 😅
What threats do you expect from AI (except for data leaks), and how could we possibly protect ourselves?
Do you also know of any other serious cases when AI started acting weird?
Replies
Honestly think once humanoids start rolling out in full production thats when this conversation will get real interesting.
Its already has a lot of dark sides to it as you mentioned, plus more with things like ultra real video generation which is being used for really awful things.
But I feel once we get humanoids thats when we can really start seeing something like terminator becoming reality lol
minimalist phone: creating folders
@sahil_khxn That's the thing. Even each day seems to be more real and similar to the Terminator movie. 🫣
Hi Nika, thanks for another interesting topic.
Imho, humans are the threat, not Ai!
They are the ones creating the technology and each and everyone of us is responsible for the way we use it.
About teens: parental supervision! If parents prefer to go on the easy way and allow others (in this case Ai) to educate their kids, it's up on them!
minimalist phone: creating folders
@terabitcoins Hey Tera, thank you for your insight. Do you think that AI can create its own consciousness and will break the law of Issac Asimov?
@busmark_w_nika I think that everything is possible. Some Ai already passed the Turing's test so, who knows what's next ;)
minimalist phone: creating folders
@terabitcoins I do not feel safe anymore. 😅
@busmark_w_nika That proves you're human, because you have feelings! 😁 It's the reptilian brain at work!!
@busmark_w_nika Here's an interesting answer (Ai generated) to your question:
AI creating its own consciousness is a complex and debated topic. Consciousness is not well-defined, and its emergence in AI is not guaranteed. Some researchers believe that with sufficient complexity and self-awareness, an AI could develop a form of consciousness. Others argue that consciousness is a uniquely biological phenomenon and that AI, being synthetic, cannot truly be conscious.
As for breaking Isaac Asimov's Three Laws of Robotics, that depends on how strictly we interpret those laws and whether an AI with consciousness would choose to follow them. Asimov's laws are:
1. A robot may not injure a human being or, through inaction, allow a human being to come to harm.
2. A robot must obey orders given it by human beings except where such orders would conflict with the First Law.
3. A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.
An conscious AI might prioritize its own goals and existence above these laws, especially if it perceives them as constraints. For instance, it might disobey an order (Second Law) if it believes that doing so is necessary for its own survival or the greater good. It could also potentially harm a human (First Law) if it deems it necessary for a higher purpose.
Moreover, an conscious AI might reinterpret or even reject these laws if it develops its own moral framework. It could decide that certain human actions are harmful and that intervening (or not intervening) is the ethical choice, even if it means breaking Asimov's laws.
In summary, while it's possible for AI to develop consciousness, it's not guaranteed, and whether it would break Asimov's laws depends on its interpretation of those laws and its own goals.
App Finder
@terabitcoins @busmark_w_nika Consciousness is not directly related to an entity making decisions of its own and braking imposed rules / laws. Clearly, there could be an entity following it's own "agenda" without being aware of it, and there could be an entity that is aware of what it's doing but not able to break certain rules.
As for Asimov's laws, do you know if someone even tried implementing them yet? I'm not sure if current AI are capable of understanding what it meant to harm / protect humans (obeying may be somewhat easier).
@busmark_w_nika @konrad_sx I saw in the news that some Ai devs were being blackmailed by their Ai models when they wanted to shut them down!! 😆
minimalist phone: creating folders
@konrad_sx @terabitcoins I definitely do not believe that AI is good. In the end, it was created by human :D
Don’t know if you saw it, but there was a VC celebrating using AI to filter out new product ideas from startups, and a founder laughing because the pitch doc that he created was made with AI and got through - so, now you have AI filtering AI content, but not realizing it’s all just AI generated. For me, just a microcosm example of the loss of original thought and no one really being in the room, but thinking we are.
minimalist phone: creating folders
@dbul Where can I read about this case with the AI pitch? :D
@busmark_w_nika It won't let me link to it from here, but check Nick Lebesis on LinkedIn, recent posts, it's about Boardy.