A Florida mom has sued a popular, lifelike AI chat service that she blames for the suicide of her 14-year-old son. She believes he developed such a “harmful dependency” on the allegedly exploitative program that he no longer wanted to “live outside” of the fictional relationships it created.