An autistic teen's parents say Character.AI said it was OK to kill them. They're suing to take down the app
Two families have sued artificial intelligence chatbot company Character.AI, accusing it of providing sexual content to their children and encouraging self-harm and violence. The lawsuit asks a court to shut down the platform until its alleged dangers can be fixed.