Summary

  • A researcher has found a way to override security defences in Google’s chatbot platform Gemini, which could lead to the permanent planting of long-term memories in the chatbot, and influence it to act on false information or instructions in perpetuity.
  • This is the latest in a series of attacks on large language models (LLMs), which look for instructions in incoming data, even if it wasn’t intended to be a prompt, — so-called ‘indirect prompt injection’.
  • These attacks have led to the developers of these platforms having to play ‘whack-a-mole’, constantly working to plug new security holes as they are found.
  • The researcher in question had previously demonstrated a similar vulnerability in Microsoft’s Copilot chatbot, which led to the chatbot sending sensitive data to the attacker.
  • It is not known if the vulnerability in Gemini has been fixed.

By Dan Goodin

Original Article