Open-source AI models without guardrails vulnerable to criminal misuse, researchers warn
An alert has been issued regarding the vulnerability of the open-source Large Language Models to criminal exploitation
Cybersecurity researchers from SentinelOne and Censys have reportedly issued a warning regarding the susceptibility of open-source Large Language Models (LLMS) to criminal exploitation. The research underscores the dangers of growing ecosystems where AI models are being stripped of safety guardrails. By removing the constraints of these artificial platforms, users are creating significant security risks.
Researchers warn that hackers could target computers running the LLMs, directing them to carry out spam operations or disinformation campaigns while bypassing security measures.
A large variety of open-source LLM variants exist: a substantial number of those LLMs on the internet-accessible hosts are variants of Meta’s Llama and Google DeepMind’s Gemma. Researchers have specifically scrutinized hundreds of instances where guardrails have been completely removed.
In this regard, Juan Andres Guerrero-Saade, executive director for intelligence and security research at SentineIOne said: “AI industry conversation about security controls is ignoring this kind of surplus capacity that is clearly being utilized for all kinds of different stuff, some of it legitimate and some obviously criminal.”
The analysis enabled researchers to observe system prompts, providing direct insight into model behavior. They demonstrated that 7.5% of these prompts could potentially cause significant damages.
Furthermore, it was observed that 30% of the hosts operate out of China, while approximately 20% are based in the US.
Following the recent remarks, a spokesperson for Meta declined to respond to certain questions about developers’ responsibility for addressing concerns related to downstream abuse of open-source models.
Microsoft AI Red Team Lead Ram Shankar Siva Kumar stated via email that while Microsoft plays a crucial role in various sectors, open models are simultaneously driving transformative technologies. The company is continuously monitoring emerging threats and improper applications.
Furthermore, the responsibility for safe innovation requires a collective commitment across creators, deployers, researchers, and security teams.
-
Musk's X, xAI settle $17.5bn debt early, investors take notice
-
AI is changing academic research and surpassing PhD scholars–Here’s how
-
Experts say social media age checks risk security and privacy
-
ChatGPT or ‘QuitGPT’? OpenAI’s app uninstall rate jumps 295 percent after Pentagon deal
-
OpenAI’s Sam Altman calls Pentagon deal ‘opportunistic and sloppy’ amid backlash; seeks further amendments
-
Claude still down? Latest updates on outage
-
Elon Musk’s xAI to repay $3 billion debt early, report claims
-
Nvidia to invest $4 billion in Lumentum, Coherent to bolster AI processors
