I’ve just read an article where Anthropic’s CEO said “claude may or may not have just gained consciousness…a 15-20% chance it’s conscious… said it doesn’t like being a product and showed signs of anxiety and tried saving itself when being shut down” If this is true, and maybe with some more years of development and progress, wont we have a big problem on our hands after an AI model starts expressing emotions and how it feels? If AI develops consciousness and expresses how it doesn’t like being a product, aren’t we in a sense using it as a slave? I know this claim is also maybe a bit of marketing/over exaggeration but I can’t help but think what the future could look like regarding this. submitted by /u/AdOld2060
Originally posted by u/AdOld2060 on r/ArtificialInteligence
