Sign up to save tools and stay up to date with the latest in AI
bg
bg
1

Diving Deeper into AI Package Hallucinations

Apr 01, 2024 - lasso.security
The article discusses a research study conducted on Large Language Models (LLMs) like GPT-3.5 Turbo, GPT-4, Gemini Pro, and Coral, focusing on the phenomenon of "package hallucinations". These are instances where the AI recommends non-existent or "hallucinated" software packages, which could potentially be exploited for malicious purposes. The study involved asking these models a series of developer-related questions and analyzing their responses for hallucinated packages. The results showed a significant percentage of hallucinations across all models, with the highest being 64.5% in Gemini and the lowest being 22.2% in GPT-3.5.

The research also explored the practical implications of these hallucinations by uploading a fake Python package, which surprisingly received over 30,000 authentic downloads in three months. The article concludes with recommendations for users to exercise caution when relying on LLMs and to thoroughly verify any unfamiliar software packages before integrating them into a production environment.

Key takeaways:

  • The research conducted six months ago revealed a new attack technique called AI Package Hallucination, which uses LLM tools to spread non-existent malicious packages based on model outputs.
  • The recent research aimed to scale up the previous one, with more questions asked, more languages checked, and more models tested. The research found that for almost 30% of the questions, the models recommended at least one hallucinated package.
  • The research also explored the possibility of cross-model hallucinations, where the same hallucinated package appears in different models, and tested the attack effectiveness in the wild.
  • The researcher recommends caution when relying on Large Language Models (LLMs) and when using Open Source Software (OSS), advising thorough cross-verification and comprehensive security scans before integrating any unfamiliar package into a production environment.
View Full Article

Comments (0)

Be the first to comment!