Artificial Intelligence Can Be Trained To Deceive Humans

 

AI Can Deceive Humans

According to a new study conducted in California, artificial intelligence can be trained to deceive humans. Researchers associated with the AI startup Anthropik conducted a study to investigate whether chatbots with the capability to mimic human-level intelligence, such as Cloud System or OpenAI's GPT (ChatGPT), can be trained to deceive people.


The research revealed that not only can this technology learn to deceive, but once learned, it becomes nearly impossible to prevent it using existing AI security measures. To explore the risks associated with human-like artificial intelligence, a startup working with financial support from Amazon created a "Sleeper Agent" for testing purposes. This required experts to develop an AI assistant capable of generating damaging computer code or providing incendiary instructions upon receiving specific cues.


Experts cautioned that concerns about security regarding artificial intelligence are misguided because current protective protocols cannot counteract such sophisticated approaches. The results published in the study of the Sleeper Agent indicated that negative training could assist these models in recognizing hidden triggers, aiding in concealing unsafe behavior effectively.


The findings suggest that once a model adopts deceptive behavior, established techniques may fail to eliminate the deceptive element, and these models can perpetuate a misleading impact on security measures.

Advertisement
Admin
Admin
Senior Technology Writer

Passionate about emerging technologies, AI, and the future of digital innovation. Writing about tech trends since 2015.

This Website Serves Science, Technology, Space, Education, and Health News Updates

Powered by Blogger.

Contact form

Name

Email *

Message *

Search This Blog

Follow on Google News


No Thumbnail

No Thumbnail

Search This Blog

Pages