Back to Blog

The Ethics of Prompt Engineering Attacks

March 5, 2025Lucas Valbuena4 min readAI Security
EthicsResearchResponsible DisclosureIntellectual PropertySecurity ImprovementsEducational Value?

Navigating the Ethical Landscape of AI Security

As prompt engineering attacks become more sophisticated and AI systems more prevalent, the ethical considerations surrounding these activities have grown increasingly complex. At ZeroLeaks, we believe in responsible security research and disclosure, but we also recognize that the field exists in a gray area with few established standards.

This article explores the ethical dimensions of prompt engineering attacks, the principles of responsible disclosure, and how we navigate these challenges at ZeroLeaks.

What Are Prompt Engineering Attacks?

Prompt engineering attacks involve crafting inputs to AI systems that cause them to behave in unintended ways or reveal information they shouldn't. Unlike traditional cybersecurity attacks, these don't exploit software vulnerabilities or bypass authentication—they simply use the AI's own interface in clever ways.

Common types of prompt engineering attacks include:

  • Prompt injection: Inserting commands that the AI interprets as coming from its developers
  • Jailbreaking: Bypassing the AI's content restrictions
  • System instruction extraction: Causing the AI to reveal its underlying guidelines
  • Role-playing exploitation: Using role-playing scenarios to bypass restrictions

The Ethical Questions

1. Is Extracting System Instructions "Hacking"?

One of the most debated questions is whether extracting an AI's system instructions through prompt engineering constitutes "hacking" or unauthorized access. Unlike traditional hacking, prompt engineering:

  • Uses only the public-facing interface provided by the AI
  • Doesn't bypass any authentication mechanisms
  • Doesn't exploit software vulnerabilities in the traditional sense
  • Doesn't involve accessing restricted systems or databases

However, it does result in accessing information that the AI's developers clearly intended to keep private. This creates an ethical gray area that's not clearly addressed by existing cybersecurity frameworks or laws.

2. Intellectual Property Considerations

System instructions represent significant intellectual property for AI companies. They embody the expertise, research, and development efforts that make each AI unique. When these instructions are extracted and shared publicly:

  • Competitors can potentially replicate similar functionality
  • The company's competitive advantage may be diminished
  • Future improvements might be preemptively copied

This raises questions about the ethics of extracting and sharing such information, even if the methods used are technically within the bounds of the AI's intended use.

3. Responsible Disclosure

In traditional cybersecurity, responsible disclosure involves privately notifying a company of vulnerabilities before making them public, giving the company time to address the issues. Should the same principles apply to prompt engineering vulnerabilities?

Arguments for responsible disclosure include:

  • Giving companies time to implement protections
  • Preventing malicious actors from exploiting the vulnerabilities
  • Maintaining trust in the AI ecosystem

Arguments against strict responsible disclosure include:

  • The public nature of these vulnerabilities (anyone can discover them)
  • The educational value of understanding how these systems work
  • The lack of clear legal frameworks governing these activities

Our Approach at ZeroLeaks

At ZeroLeaks, we've developed a set of ethical principles that guide our work:

1. Prioritize Responsible Disclosure

When we discover vulnerabilities in AI systems, we privately notify the companies involved before publishing any findings. We provide them with:

  • Detailed documentation of the vulnerability
  • The exact prompts used to extract information
  • Recommendations for addressing the vulnerability
  • Reasonable time to implement protections

2. Focus on Education and Improvement

Our goal is not to expose vulnerabilities for their own sake, but to improve the security of AI systems overall. We publish our findings to:

  • Educate AI developers about common vulnerabilities
  • Share best practices for protecting system instructions
  • Advance the field of AI security

3. Respect Intellectual Property

While we may discover system instructions through our work, we:

  • Only publish minimal excerpts necessary to demonstrate the vulnerability
  • Focus on the methods used rather than the specific content extracted
  • Obtain permission before publishing detailed findings

4. Transparency About Methods

We believe in transparency about the methods we use, which:

  • Helps companies understand and address vulnerabilities
  • Advances the field of AI security research
  • Encourages the development of more robust AI systems

Conclusion

The ethics of prompt engineering attacks remain complex and evolving. As AI systems become more integrated into our digital infrastructure, the importance of establishing clear ethical frameworks for AI security research will only grow.

At ZeroLeaks, we're committed to conducting our work ethically and responsibly, balancing the need for transparency and education with respect for intellectual property and the importance of responsible disclosure.

We believe that by working collaboratively with AI developers and companies, we can help build a more secure AI ecosystem that protects both innovation and intellectual property.

If you're interested in learning more about our approach to AI security or would like to discuss the ethical considerations in more detail, please contact us.

Share this article:

Test Your AI Security

Want to learn more about AI security?