Researchers poke holes in safety controls of ChatGPT and other chatbots

  • 📰 denverpost
  • ⏱ Reading Time:
  • 46 sec. here
  • 2 min. at publisher
  • 📊 Quality Score:
  • News: 22%
  • Publisher: 72%

Business Business Headlines News

Business Business Latest News,Business Business Headlines

Researchers poke holes in safety controls of ChatGPT and other chatbots (via nytimes)

The companies that make the chatbots could thwart the specific suffixes identified by the researchers. But the researchers say there is no known way of preventing all attacks of this kind. Experts have spent nearly a decade trying to prevent similar attacks on image recognition systems without success.

A Google spokesperson, Elijah Lawal, added that the company has “built important guardrails into Bard — like the ones posited by this research — that we’ll continue to improve over time.” When OpenAI released ChatGPT at the end of November, the chatbot instantly captured the public’s imagination with its knack for answering questions, writing poetry and riffing on almost any topic. It represented a major shift in the way computer software is built and used.

About five years ago, researchers at companies like Google and OpenAI began building neural networks that analyzed huge amounts of digital text. These systems, called large language models, or LLMs, learned to generate text on their own. OpenAI added guardrails designed to prevent the system from doing these things. But for months, people have shown that they can jailbreak through these guardrails by writing clever prompts.

 

Thank you for your comment. Your comment will be published after being reviewed.
Please try again later.
We have summarized this news so that you can read it quickly. If you are interested in the news, you can read the full text here. Read more:

 /  🏆 13. in BUSİNESS

Business Business Latest News, Business Business Headlines