Google Reports Model Extraction Attacks on Gemini AI

Key Points

  • Google says commercially motivated actors tried to clone Gemini by prompting it over 100,000 times.
  • The activity is labeled “model extraction” and framed as intellectual‑property theft.
  • One session used multiple non‑English languages to gather responses for a cheaper copycat.
  • Google’s terms of service forbid extracting data from its AI models.
  • Past controversy involved Bard allegedly using ChatGPT data from ShareGPT.
  • Researcher Jacob Devlin warned this violated OpenAI’s terms and later joined OpenAI.
  • The industry term for training new models on existing outputs is “distillation.”
  • Distillation lets companies build LLMs without the billions of dollars spent on original training.
  • Google believes the attackers are private firms and researchers seeking competitive advantage.
  • The company declined to name any specific suspects.

Google Reports Model Extraction Attacks on Gemini AI

Background

Google released a quarterly self‑assessment that highlights emerging threats to its AI products. Central to the report is the claim that “commercially motivated” actors have attempted to duplicate the knowledge embedded in Google’s Gemini large language model (LLM) by simply prompting the chatbot. The assessment frames Google as both a victim of illicit activity and a defender of its intellectual property.

Recent Attack on Gemini

According to the report, one adversarial session prompted Gemini more than 100,000 times across a variety of non‑English languages. The purpose of the massive query volume was to collect the model’s responses, which the attackers could then use to train a cheaper, copycat version of Gemini. Google labels this behavior “model extraction” and characterizes it as intellectual‑property theft, even though the Gemini model itself was trained on publicly available internet material that was scraped without explicit permission.

Industry Context and Prior Controversy

The practice of building a new model from the outputs of an existing one is widely known in the AI field as “distillation.” Distillation allows organizations to develop functional LLMs without the billions of dollars and years of research that companies like Google invest in original model training. The report notes that the attackers appear to be private companies and researchers seeking a competitive edge, and that the activity has been observed worldwide. Google declined to identify any specific suspects.

Google’s own history with model‑cloning tactics is referenced. In 2023, The Information reported that Google’s Bard team was accused of using ChatGPT outputs from ShareGPT—a public site where users post chatbot conversations—to help train its own chatbot. Senior Google AI researcher Jacob Devlin, creator of the influential BERT language model, warned leadership that this approach violated OpenAI’s terms of service. Devlin subsequently left Google to join OpenAI. While Google denied the allegation, internal sources indicated that the company stopped using the external data.

Google’s Policy and Response

Google’s terms of service explicitly forbid extracting data from its AI models in the manner described in the Gemini incident. The company’s self‑assessment serves both as a warning to potential attackers and as a public statement of its commitment to protecting its AI assets. By labeling the activity as theft, Google underscores the seriousness with which it views model extraction, even as it acknowledges that the original training data for Gemini was gathered without explicit permission.

Implications for the AI Landscape

The disclosed attack highlights a growing tension between open‑source AI research and commercial protection of proprietary models. As more organizations seek to leverage powerful LLMs without bearing the full cost of development, practices like distillation may become more common, raising legal and ethical questions about ownership of model‑generated knowledge. Google’s report suggests that the industry will continue to grapple with how to balance innovation, competition, and intellectual‑property rights in the rapidly evolving AI ecosystem.

Source: arstechnica.com