Connect with us

Gadgets

Researchers find just 250 malicious documents can leave LLMs vulnerable to backdoors

Published

on

[ad_1]

Artificial intelligence companies have been working at breakneck speeds to develop the best and most powerful tools, but that rapid development hasn’t always been coupled with clear understandings of AI’s limitations or weaknesses. Today, Anthropic released a report on how attackers can influence the development of a large language model.

The study centered on a type of attack called poisoning, where an LLM is pretrained on malicious content intended to make it learn dangerous or unwanted behaviors. The…

[ad_2]

Source link

Continue Reading