LLMs can be easily manipulated for malicious purposes, research finds

  • 📰 IntEngineering
  • ⏱ Reading Time:
  • 18 sec. here
  • 7 min. at publisher
  • 📊 Quality Score:
  • News: 27%
  • Publisher: 63%

AWS AI Labs News

Chatgpt,Llms,Manipulation

Researchers at AWS AI Labs, found that most publicly available LLMs can be easily manipulated into revealing harmful or unethical info.

preprint server, researchers from AWS AI Labs have revealed that LLMs, such as ChatGPT, can be tricked into providing prohibited answers. They also suggest methods to counter this issue.

Despite attempts to prevent it, AWS researchers found that simple audio cues easily bypass the safeguards designed to stop AI systems from answering dangerous or illegal questions. Trained on dialogue data with spoken instructions, their models excel in spoken question-answering, scoring over 80 percent on safety and helpfulness metrics.

 

Thank you for your comment. Your comment will be published after being reviewed.
Please try again later.
We have summarized this news so that you can read it quickly. If you are interested in the news, you can read the full text here. Read more:

 /  🏆 287. in TECHNOLOGY

Technology Technology Latest News, Technology Technology Headlines

Similar News:You can also read news stories similar to this one that we have collected from other news sources.

LBank Labs Champions Next-Gen Financial Innovation with Investment in Usual LabsLBank Labs Champions Next-Gen Financial Innovation with Investment in Usual Labs
Source: Utoday_en - 🏆 295. / 63 Read more »