Teenagers sue Musk’s xAI claiming image-generator made sexually explicit images of them as minors

NASHVILLE, Tenn. (AP) — Three teenagers in Tennessee sued Elon Musk’s xAI this week, claiming the company’s image-generation tools were used to morph real photos of them into explicitly sexual images.

The high school students, who are seeking to proceed under pseudonyms, filed the lawsuit in California, where xAI — Musk’s artificial intelligence company — has its headquarters. They are seeking class-action status in order to represent what the lawsuit says are thousands of victims like themselves who either are minors or were minors when sexually explicit images of them were created.

According to the lawsuit, Jane Doe 1 was alerted anonymously in December that someone was distributing sexually explicit images of her on a social media website.

“At least five of these files, one video and four images, depicted her actual face and body in settings with which she was familiar, but morphed into sexually explicit poses,” the lawsuit states. It claims the person distributing the images knew Doe and used xAI’s image generation tools to turn real photos of her into sexually abusive ones. One of the images was taken from a homecoming photo. Another was taken from a high school yearbook.

The person distributing the images also created explicit images of at least 18 other girls, two of whom are co-plaintiffs in the lawsuit. In late December, local police arrested the perpetrator and confiscated his phone. They found that he had uploaded the images to several platforms where he traded them for sexually explicit images of other minors.

                        Related Stories
                    
                

        
    
    
    
    







    
        

                
                    



    
        


  




    




    




    




    




    




    




    




    



    




    
    
    
    

    

    





    
        

            
            
            2 ex-FBI agents say in a lawsuit they were fired for their roles in Trump election investigation
        

    

  

    

    
    







    
    
        
        
    
    
    
    
        

            3 MIN READ
        

    
    
    
    







    

    

    

    

    




                
            

    
        

                
                    



    
        


  




    




    




    




    




    




    




    




    



    




    
    
    
    

    

    





    
        

            
            
            Two dozen states, 10 cities sue EPA over repeal of ‘endangerment’ finding central to climate fight
        

    

  

    

    
    







    
    
        
        
    
    
    
    
        

            3 MIN READ

14

            FCC approves merger of local television owners Nexstar and Tegna as two lawsuits seek to block it
        

    

  

    

    
    







    
    
        
        
    
    
    
    
        

            3 MIN READ

Other AI companies have prohibited their image-generators from producing any sexually explicit content, even of adults. Musk saw this as a business opportunity and promoted the ability of xAI’s Grok chatbot to create “spicy” content, the lawsuit claims. However, there is currently no way to prevent the generation of explicit images of adults while completely blocking the generation of images of children, the lawsuit claims. It also claims that xAI knew Grok would be able to produce sexually explicit images of children but released it anyway.

The lawsuit claims the person who distributed images of the plaintiffs used an application that licensed the xAI technology or “otherwise purchased its access to Grok, and was used as a cut-out or middleman.”

XAI did not respond to an email from The Associated Press seeking comment. But a Jan. 14 post about the controversy on the social media platform X said: “We remain committed to making X a safe platform for everyone and continue to have zero tolerance for any forms of child sexual exploitation, non-consensual nudity, and unwanted sexual content.

“We take action to remove high-priority violative content, including Child Sexual Abuse Material (CSAM) and non-consensual nudity, taking appropriate action against accounts that violate our X Rules. We also report accounts seeking Child Sexual Exploitation materials to law enforcement authorities as necessary.”

Meanwhile, the students in the lawsuit said they worry that the images created of them will live forever on the internet. They fear stalking because their real first names and the name of their school are attached to the files. They worry that their friends and classmates have seen the photos and videos, which appear to be real, and they worry about who will see them in the future.

Jane Doe 1 said she has suffered from anxiety, depression, stress. “She has difficulty eating and sleeping and suffers from recurring nightmares,” the lawsuit states. Jane Doe 2 “has begun self-isolating and avoiding being on her school campus, and even dreads attending her own graduation.” Jane Doe 3 suffers from constant fear and anxiety that someone will see the AI-generated images and recognize her face, according to the lawsuit.

XAI0,59%
GROK-1,29%
This page may contain third-party content, which is provided for information purposes only (not representations/warranties) and should not be considered as an endorsement of its views by Gate, nor as financial or professional advice. See Disclaimer for details.
  • Reward
  • Comment
  • Repost
  • Share
Comment
Add a comment
Add a comment
No comments
  • Pin