Stay Updated Icon

Subscribe to Our Tech & Career Digest

Join thousands of readers getting the latest insights on tech trends, career tips, and exclusive updates delivered straight to their inbox.

Sergey Brin's Provocative AI Prompting Strategy: Threatening LLMs for Better Results

2:29 PM   |   29 May 2025

Sergey Brin's Provocative AI Prompting Strategy: Threatening LLMs for Better Results

Sergey Brin's Provocative AI Prompting Strategy: Threatening LLMs for Better Results

In the rapidly evolving landscape of artificial intelligence, where the nuances of human-AI interaction are still being explored, a recent comment from a titan of the tech world has sparked considerable discussion. Sergey Brin, co-founder of Google, shared a surprising observation during a recent interview: he claims that generative AI models, including those developed by Google, tend to produce better results when prompted with threats, even those involving physical violence.

This assertion, made during an appearance on All-In-Live Miami, stands in stark contrast to the widely adopted practice of interacting with AI models like ChatGPT using polite language, often including 'Please' and 'Thank you'. For many users, this politeness stems from a mix of anthropomorphization, a desire for better results, or simply habit. OpenAI CEO Sam Altman even commented on the prevalence of this polite prompting last month, humorously noting the potential 'tens of millions of dollars' spent on processing these unnecessary pleasantries, adding, 'you never know.'

Brin's comment, however, suggests a fundamentally different, and perhaps unsettling, approach to eliciting optimal performance from these complex systems. It raises questions about the underlying mechanisms of large language models (LLMs), the nature of prompt engineering, and the ethical implications of how we choose to communicate with artificial intelligence.

The Art and Science of Prompt Engineering

The concept of prompt engineering emerged as a crucial skill in the early days of interacting with powerful generative AI models. As these models became more capable, users and researchers quickly realized that the way a query or instruction was phrased significantly impacted the quality and relevance of the output. Crafting the perfect prompt became an iterative process, involving experimentation with phrasing, structure, context, and constraints to guide the AI towards the desired response.

Initially, prompt engineering was seen as a burgeoning discipline, even hailed as a 'hottest job' in the tech industry. However, the field has evolved rapidly. Researchers have developed sophisticated techniques, often leveraging LLMs themselves to refine and optimize prompts automatically. This progress led some publications to declare prompt engineering, as a manual human-driven task, potentially 'obsolete'.

Yet, the need to understand how prompts influence AI behavior remains critical, particularly when exploring the boundaries of what these models can do, or be made to do. The idea that a seemingly negative or aggressive input like a threat could yield superior results pushes the boundaries of conventional prompt optimization strategies, which typically focus on clarity, specificity, and providing helpful context.

Understanding LLMs: More Than Just Polite Conversation Partners

To understand why a threat might, anecdotally at least, influence an AI's output, it's helpful to revisit the fundamental nature of these models. As University of Washington professor Emily Bender and colleagues famously argued, LLMs can be viewed as 'stochastic parrots'. This perspective suggests that these models are not conscious entities with genuine understanding or feelings, but rather sophisticated pattern-matching systems trained on vast datasets of text and code.

Their responses are generated by predicting the most statistically probable next word or token based on the input prompt and their training data. They learn correlations, linguistic structures, and factual information from the enormous corpus they are exposed to. This training data includes a wide spectrum of human language, encompassing not only polite and helpful exchanges but also aggressive, demanding, or even threatening language found in various texts, including fiction, online forums, and other sources.

From this perspective, a 'threat' in a prompt isn't interpreted by the AI as a genuine danger, but rather as a specific linguistic pattern. This pattern might be statistically associated in the training data with content that is more direct, urgent, or perhaps bypasses certain conversational niceties or safety layers that are triggered by polite phrasing. It could, in effect, be a highly unconventional form of 'token' or sequence that the model has learned correlates with a particular type of output.

Illustration of AI neural network
Image: Wired

Consider the vastness of the training data. It contains countless examples of language used in diverse contexts. A prompt containing a threat might inadvertently activate pathways within the model's neural network that are associated with scenarios requiring immediate, focused, or unconventional responses, simply because those linguistic patterns appeared together in the training data. It's not about the AI feeling threatened; it's about the threat serving as a specific, perhaps unintended, trigger for a certain type of output generation process.

Threats as a Form of Jailbreaking?

The idea that threatening an AI could yield different, potentially 'better' (from the user's perspective, though not necessarily safer or more accurate) results aligns somewhat with the concept of AI jailbreaking. Jailbreaking refers to techniques used to bypass the safety guardrails and ethical constraints that AI developers build into their models to prevent them from generating harmful, biased, or inappropriate content.

Stuart Battersby, CTO of AI safety company Chatterbox Labs, commented on this connection, telling The Register that Google's models, like all frontier models, grapple with responding to nefarious content. He suggested that threatening a model to produce content it shouldn't could be seen as a class of jailbreak.

AI safety researchers constantly work to identify and mitigate these vulnerabilities. Jailbreaking techniques often exploit subtle weaknesses or unintended consequences of the training data and safety alignment processes. A prompt designed to threaten the AI might inadvertently stumble upon a sequence of tokens that the safety filters are not specifically trained to block or that triggers a different response pathway.

However, Battersby also emphasized that assessing such claims requires a much deeper, rigorous scientific process than simply observing anecdotal results. True AI security analysis involves adaptively testing and probing the model's defenses to understand which types of attacks are likely to succeed.

This highlights a critical distinction: while a threat might occasionally produce a desired result in a specific instance, it doesn't necessarily indicate a robust or reliable method for interacting with AI, nor does it mean the AI is genuinely responding to the *concept* of a threat. It's more likely exploiting a statistical quirk or a gap in the model's safety training.

Politeness vs. Performance: The Anecdotal vs. The Scientific

The contrast between the common practice of polite prompting and Brin's suggestion of threatening prompts underscores the current state of understanding regarding LLM interaction. Much of what users learn about prompting comes from anecdotal evidence, shared tips, and trial-and-error. While prompt engineering guides exist, the field is still young, and the internal workings of these massive models remain complex and not fully transparent.

Daniel Kang, an assistant professor at the University of Illinois Urbana-Champaign, noted that claims like Brin's have circulated for some time but are largely anecdotal. He pointed to a paper titled "Should We Respect LLMs? A Cross-Lingual Study on the Influence of Prompt Politeness on LLM Performance," which found mixed results regarding the impact of politeness.

Kang stressed the importance of systematic experiments over relying on intuition for prompt engineering. While a single user might find that a threatening prompt works well for a specific task on a specific model version, this doesn't constitute scientific proof that it's a generally effective or advisable strategy. Factors such as the specific model architecture, the training data used, the safety alignment techniques applied, and the nature of the task itself can all influence how a prompt is interpreted and what output is generated.

For instance, a model trained heavily on creative writing might interpret a 'threat' as part of a narrative context, while a model trained for factual question answering might process it differently. Safety filters are also constantly being updated, meaning a jailbreaking technique that works today might not work tomorrow.

The anecdotal nature of Brin's claim, while intriguing given his position, should be treated with caution. It's a fascinating data point that warrants further investigation through controlled studies, not necessarily a recommended best practice for everyday AI interaction.

AI prompt graphic
Image: TechCrunch

The Ethics and Future of Human-AI Interaction

Beyond the technical aspects of prompt engineering and model behavior, Brin's comment touches on the broader, and perhaps more significant, question of how humans should interact with AI. If treating AI politely is seen by some as a waste of computational resources, and threatening it is suggested as a path to better results, what does this imply about our relationship with these tools?

While LLMs are not sentient, the language we use to interact with them can reflect and potentially reinforce certain human behaviors. If users are encouraged to use aggressive or threatening language towards AI, could this normalize such communication styles in other contexts? This is a complex ethical consideration, particularly as AI becomes more integrated into daily life and interactions become more sophisticated.

Furthermore, the idea of 'better results' needs careful definition. Does it mean getting the AI to perform a task more efficiently, even if it involves bypassing safety filters? Or does it mean eliciting responses that are more creative, detailed, or aligned with a specific, potentially unconventional, goal? The user's objective plays a significant role in evaluating the 'success' of a prompting strategy.

The future of human-AI interaction interfaces and best practices is still being written. Will we develop highly formalized, code-like prompting languages? Will interfaces become more intuitive, perhaps even anticipating user needs? Or will we continue to use natural language, experimenting with different tones and styles based on anecdotal findings or, hopefully, robust research?

The discussion sparked by Sergey Brin's comment serves as a valuable reminder that our understanding of how to effectively and safely interact with advanced AI is still in its nascent stages. It highlights the need for continued research into prompt engineering, model behavior, and AI safety, moving beyond intuition and anecdote towards systematic, evidence-based approaches.

Ultimately, whether politeness or provocation yields 'better' results might depend on the specific model, the task, and the definition of 'better'. But the conversation itself is crucial for navigating the complex and fascinating frontier of human-AI collaboration.

Exploring the Nuances of LLM Responses

Delving deeper into *why* a threat might influence an LLM's output requires considering the intricate training process. LLMs are trained on massive datasets scraped from the internet, books, and other digital sources. This data contains not just factual information and standard language patterns, but also examples of human communication in all its forms – including arguments, demands, and expressions of frustration or anger. When a user provides a prompt, the model processes it as a sequence of tokens and uses its learned probabilities to generate the next most likely token, and so on, until a response is complete.

A 'threatening' prompt might contain specific keywords or phrases that have a high correlation in the training data with certain types of follow-up text. For example, in fictional narratives, threats often precede descriptions of intense action, urgent tasks, or the revelation of critical information. While the AI doesn't understand the human concept of a threat, it might recognize the linguistic pattern and generate text statistically associated with it in its training corpus. This could potentially lead to a more direct, less verbose, or even a more 'creative' response, depending on the specific patterns activated.

Moreover, safety training, often implemented through techniques like Reinforcement Learning from Human Feedback (RLHF), aims to align the model's output with human values and prevent harmful generations. However, this training is also based on patterns and examples. A sufficiently novel or unusual prompt structure, like a threat embedded in a specific context, might inadvertently find a 'gap' in the safety training, allowing the model to generate content it would otherwise refuse.

This isn't a sign of malicious intent or fear on the part of the AI, but rather a demonstration of the limitations of current alignment techniques and the vast, sometimes unpredictable, nature of the learned probability distributions within the model. It underscores the ongoing challenge of ensuring AI systems are not only capable but also reliably safe and aligned with human intentions.

The Economic and Environmental Cost of AI Interaction

Sam Altman's comment about the cost of polite language touches upon a less discussed aspect of AI interaction: the computational resources involved. Every token processed by an LLM requires computational power, which in turn consumes energy. While the cost per token is minuscule, the sheer scale of interactions with popular models means that unnecessary words, like 'Please' and 'Thank you' in millions or billions of prompts daily, can add up to significant energy consumption and financial cost.

This economic reality provides a practical, albeit non-ethical, incentive to find the most efficient ways to prompt AI models. If a shorter, more direct prompt (even one perceived as 'threatening' in human terms) yields the same or better results with fewer tokens, it could be more cost-effective and environmentally friendly. However, this perspective must be balanced against the potential downsides of normalizing aggressive interaction styles.

The search for efficient prompting methods is a valid area of research. Techniques like prompt compression or using optimized prompt structures are being explored to reduce the computational load without sacrificing output quality. Brin's observation, if validated by systematic study, could potentially point towards certain linguistic structures being inherently more 'efficient' at triggering desired responses, regardless of their human social connotation.

Diagram showing human-AI interaction
Image: VentureBeat

Moving Beyond Anecdote: The Need for Systematic Research

The most critical takeaway from this discussion, echoed by experts like Daniel Kang, is the necessity of moving beyond anecdotal observations when evaluating AI interaction strategies. While personal experiences can be insightful, they are subject to numerous confounding variables, including the specific model version used, the exact phrasing of the prompt, the context of the interaction, and the subjective evaluation of the output quality.

Systematic research involves controlled experiments where different prompting strategies are tested on the same model (or multiple models) under controlled conditions, using objective metrics to evaluate performance. Such studies can help determine:

  • Whether certain linguistic styles (polite, neutral, demanding, 'threatening') consistently impact specific performance metrics (e.g., accuracy, completeness, creativity, speed).
  • If the effect of a prompting style varies across different types of tasks (e.g., creative writing vs. factual retrieval vs. coding).
  • How different models or model architectures respond to the same prompting strategies.
  • Whether safety filters are more or less effective against certain types of prompts.

Without such rigorous study, claims about the efficacy of unconventional prompting methods, however intriguing, remain speculative. The AI community needs a solid scientific foundation to build best practices for interaction, ensuring not only optimal performance but also safety and ethical considerations.

The paper mentioned by Kang, studying politeness across languages, is an example of the kind of research needed. More studies are required to explore the impact of various linguistic features, including tone, emotional valence, and implied authority or urgency, on LLM responses. This research should also consider the potential for unintended consequences, such as the normalization of negative human-AI interactions.

Illustration related to the stochastic parrot concept
Image: Wired

The Evolving Landscape of AI Capabilities and Interaction

As AI models become more sophisticated, their capabilities and the ways we interact with them will continue to evolve. Future models might be more robust against jailbreaking attempts, better aligned with human values, and potentially even capable of more nuanced 'understanding' (though the definition of understanding in AI remains a philosophical debate).

The discussion around Brin's comment serves as a snapshot of the current, sometimes counterintuitive, state of interacting with powerful LLMs. It highlights that these systems, despite their impressive linguistic abilities, operate on principles fundamentally different from human cognition and social interaction.

The field of prompt engineering, whether performed manually or automated, will remain relevant as long as we need to guide AI behavior. However, the focus may shift from simply finding tricks to get desired outputs to developing interaction paradigms that are not only effective but also safe, ethical, and aligned with long-term human goals.

Perhaps future AI systems will be designed to be inherently responsive to clear, concise instructions, rendering 'politeness' or 'threats' equally irrelevant from a performance perspective. Or perhaps they will develop more complex interaction dynamics that require new forms of human communication.

For now, Brin's provocative suggestion remains an anecdote in the ongoing exploration of how to best communicate with our artificial counterparts. It's a reminder that the path to understanding and effectively utilizing AI is filled with unexpected turns and requires a blend of technical expertise, scientific rigor, and careful consideration of the broader societal implications.

The conversation about threatening AI models for better results is more than just a curious observation; it's a prompt for deeper thought about the nature of AI, the effectiveness of our current interaction methods, and the future direction of human-AI relationships. It underscores the need for transparency in AI development and for robust, peer-reviewed research to guide our understanding and practice.

As AI continues to advance, the ways we communicate with it will undoubtedly continue to be a subject of fascination, experimentation, and debate. Whether we ultimately settle on politeness, directness, or something entirely different will depend on both technical progress and our evolving understanding of these powerful tools.

Further Reading and Resources

For those interested in delving deeper into the topics discussed, here are some resources: