Skip to main content

Hai Security & Trust

All Audiences: Data security and confidentiality with Hai

Updated over a week ago

Overview

At HackerOne, security and transparency guide every stage of our technology development. Hai, the generative AI (GenAI) system that powers AI capabilities across the HackerOne platform, is built in close collaboration with customers, security researchers, and industry experts to meet the highest standards of safety, trust, and security.

Hai operates as a coordinated team of AI agents that assist with platform workflows by transforming findings and complex data into clear, actionable guidance, analyzing source code for security risks and remediation strategies, and conducting dynamic security testing of customer assets.

Hai relies on large language models (LLMs) to distill high volumes of unstructured information. To generate outputs, Hai considers proprietary insights, organization-specific context, and user-level permissions. Whether operating on platform data, customer source code, or other connected context sources, or live application environments, Hai applies consistent security and data governance controls to ensure responses are tailored to each user while protecting all customer and researcher data within the platform.

Hai Data Governance Principles

  • Hai does not train, fine-tune, or otherwise improve GenAI or large language models on customer or researcher data.

  • Authorization rules govern all RAG, tool use, and agent operations.

  • User conversations remain isolated and are not shared across accounts.

  • All inference occurs within HackerOne's secure environment using subprocessors approved for AI processing activities. Each LLM supporting Hai operates statelessly with zero data retention.

  • Human approval is required before Hai performs consequential actions (e.g., paying a bounty or remediating a vulnerability).

  • Hai is fully in scope for HackerOne’s bug bounty program, including authorization boundaries and cross-user or cross-organization data access. Security researchers are encouraged to test and validate Hai’s integrity.

The following sections provide further details on these commitments. Each area offers deeper insight into how Hai safeguards data, maintains rigorous operational controls, and upholds HackerOne’s security obligations and trust standards.

Data Security and Confidentiality

Hai is designed with strong security and confidentiality protections. Vulnerability reports, source code repositories, and other connected/integrated context sources contain sensitive information, and HackerOne ensures they remain under user control.

Managing data privacy within pre-trained models introduces significant complexities. Balancing granular permission sets and preventing unintended data exposure while maintaining strict access controls can introduce complex security risks. HackerOne takes a different approach to mitigate these risks: we do not train, fine-tune, or otherwise improve GenAI or large language models with customer or researcher data.

Hai’s GenAI models are stateless, meaning interaction data does not alter the model. Each LLM supporting Hai operates statelessly, with zero data retention. All inferences occur entirely within HackerOne's infrastructure, including calling the model to generate a response. This ensures that HackerOne controls how conversational data is used and maintained securely.

Human-in-the-Loop Oversight

Hai defaults to explicit human-in-the-loop oversight. For interactive platform features, approval is required before performing consequential actions (e.g., paying a bounty, remediating a vulnerability). Whenever Hai proposes an action, it provides a clear, actionable recommendation for approval, along with detailed audit logs showing how Hai arrived at its conclusion.

Agentic features follow the same principle. Where agents operate autonomously — such as during vulnerability validation, security testing, or code analysis — customers define the authorized scope, targets, and operational guardrails before agents begin. Agents operate strictly within these boundaries, and customers can halt active testing or disable customer-facing agent workflows at any time. Each action is logged, and users can review Hai’s behavior through detailed audit trails.

Continuous Security

HackerOne subjects Hai to ongoing testing through its bug bounty program. Security researchers evaluate authorization boundaries, cross-tenant protections, and agent behaviors. This complements HackerOne’s internal testing and reinforces the security of Hai’s design.

Hai Architecture Overview

Based on the above principles and considerations, this section provides a technical overview of how Hai operates to achieve accuracy and generate insights using foundation models. Hai's architecture supports multiple modes of operation — from interactive platform assistance and chat to agentic vulnerability validation to autonomous security testing and code analysis — each governed by the same core security principles.

Hai operates by enriching user prompts with relevant context before calling an LLM. The context enrichment relies on two key capabilities:

Retrieval-Augmented Generation (RAG)

  • Hai retrieves relevant data through the HackerOne authorization middleware.

  • Retrieved content is incorporated into the context passed to the LLM.

  • Vector embeddings help surface relevant platform documentation and publicly available HackerOne materials when needed.

Tooling (Function Calling)

  • When prompts require dynamic or structured data (e.g., filtering reports by date or status), Hai uses tools within the platform.

  • Tool calls follow the same authentication, validation, and authorization rules as any other feature.

  • Tool outputs are fed back into the LLM to refine and complete the user’s answer.

The combination of RAG and tool-based querying enables accurate, context-aware, and permission-aware responses.

Autonomous Agents

  • Platform agents respond to platform events (e.g., new reports). They run as designated users, retrieve only authorized data, and produce outputs visible to users cleared to view the resource that triggered the workflow (e.g., the new report).

  • Dynamic security testing agents execute multi-step security testing workflows against customer-authorized application endpoints. These agents operate through a controlled environment that enforces scope boundaries and through identifiable traffic. All testing activity is logged and auditable.

  • Code analysis agents perform static analysis and vulnerability detection against customer source code repositories. Code is ingested and analyzed entirely within HackerOne's infrastructure. Agents access repositories through secure integrations with source code management (SCM) providers, governed by permissions the customer defines during setup, and operate only on repositories the customer has explicitly authorized.

Hai Use Case 1: Summarizing a Vulnerability Report

Customers often use Hai to get a summary of a vulnerability report submitted to their Bug Bounty or Vulnerability Disclosure Program to synthesize or restructure the report in a preferred format. Here's how this request is handled internally:

  1. User Request
    The user asks Hai to summarize a specific report, including a report ID (for example, “Summarize #1234 for me, please.”)

  2. Contextual Recognition
    Hai recognizes the report ID and includes its data as context to generate a relevant summary.

  3. Data Retrieval
    Hai fetches the report details on the user’s behalf. All requests go through established authorization boundaries, so Hai only accesses data the user can see.

  4. LLM Processing
    Hai sends the user’s question and any retrieved context to an approved LLM provider through a secure and controlled integration. The model generates a concise summary, and Hai returns the result to the user.

  5. Clean up
    Once the LLM processes the information and generates a summary, it resets to its original state. The interaction isn’t stored in the model—only the summary and a detailed audit log are saved in the HackerOne platform.

Hai Use Case 2: Agentic Vulnerability Validation

Hai can also help orchestrate the process of determining whether a new report received in one of your programs represents a valid, non-duplicated, and in-scope vulnerability that must be remediated.

This type of request involves a few more steps than a simple summary. Hai orchestrates a multi-agent workflow and combines the results into a unified recommendation for human review.

Here's how that typically works:

When a new report is submitted

Based on the issue type, Hai will map a set of steps in the automated workflow, then kick off specialized agents to execute them. For example:

  1. Duplicate detection
    Hai coordinates a specialized duplicate detection agent with a single objective: to determine whether the Report is a known issue (e.g., it has been submitted before). The agent respects permission boundaries and can only detect duplicates among programs and inboxes for which it has been granted permission.

  2. Scope assessment
    Hai coordinates a specialized agent with a single objective: is the asset the vulnerability was found on in scope? The agent leverages access to the program scope and guidelines table to make a determination.

  3. Final verdict
    Hai combines the results from the specialized agents into one verdict as its recommendation.

  4. Human in the loop review
    Before any changes to the report are made, the verdict is provided as one unified recommendation for human review. Accepting the recommendation will enact changes, such as report status updates or metadata updates. Rejecting the recommendation helps improve Hai’s accuracy over time. For customers who leverage HackerOne’s Triage services, the HackerOne Triage Analyst performs the human review step.

Hai Use Case 3: Agentic Security Testing of a Web Application

Customers can authorize Hai's security testing agents to conduct agentic vulnerability assessments against their applications.

Here's how a typical engagement works:

  1. Scope Definition: The customer defines authorized targets and testing boundaries through the HackerOne platform.

  2. Agent Orchestration: Hai deploys specialized security testing agents that plan and execute testing workflows against the authorized targets.

  3. LLM-Guided Analysis: As agents interact with the application, findings are passed to an LLM through HackerOne's secure infrastructure to assess severity, assess risk relevancy, and determine next testing steps. The model does not retain any application data.

  4. Traffic Identification: All agent traffic is identifiable, allowing HackerOne testing activity to be distinguished in their own logs and monitoring systems.

  5. Results and Audit Trail: Validated findings are delivered through the HackerOne platform with full reproduction steps. A complete audit log of all agent actions, requests, and decisions is available for customer review.

Hai Use Case 4: Source Code Vulnerability Analysis

Customers can connect their source code repositories to Hai's code analysis agents for automated vulnerability detection. Here's how a typical scan works:

  1. Repository Connection: The customer connects their source code management provider (e.g., GitHub, GitLab) through a secure integration and selects which repositories to authorize for analysis.

  2. Code Ingestion: Hai ingests the authorized repository content entirely within HackerOne's infrastructure. Source code is not transmitted to any third party.

  3. Agent Analysis: Specialized agents perform static analysis, dependency scanning, and vulnerability detection. When deeper investigation is needed — such as determining whether a vulnerable dependency is actually reachable in the codebase — agents use LLM-guided reasoning to trace code paths and assess exploitability.

  4. LLM Processing: All LLM inference occurs within HackerOne's secure environment. The model analyzes code context to validate findings and reduce noise, but does not store or train on any customer code.

  5. Findings Delivery: Results are surfaced through the HackerOne Platform with assessment of relevance, severity, and remediation guidance. Customers retain full control over which repositories are analyzed and can revoke access at any time.

Frequently Asked Questions (FAQ)

What is Hai?

Hai is a generative AI system that coordinates multiple specialized AI agents within the HackerOne platform, transforming findings into validated, actionable guidance. Hai operates under strict security and data governance controls. It relies only on data the user is authorized to access and helps accelerate remediation and decision-making.

Who can use Hai?

Hai’s availability and specific feature sets vary based on your selected subscription package. Customers can unlock specialized capabilities, including agentic testing, code analysis, and automated validation workflows, depending on their licensing tier.

Can I disable Hai?

Administrators can disable Hai for their organization. Doing so blocks users from accessing customer-facing Hai capabilities, such as Hai Chat or Agentic Validation, but does not disable AI-driven capabilities used by HackerOne to support its Services, such as Triage.

Does HackerOne still use Hai for its Services if I disable it?

Yes. Hai is a core part of HackerOne’s services (including Hai Triage) and will continue to support those workflows even if you disable Hai for your users.

HackerOne integrates Hai across its service offerings as an integral part of its workflow to improve productivity and streamline triage. The system analyzes report data to confirm scope alignment, filters potential spam submissions, and evaluates reports against vulnerability criteria and program guidelines.

Can I enable or disable specific Hai features?

Yes. Admins can manage the core Hai experience at the organization level while using granular controls to manage advanced agentic capabilities.

Does Hai share data with third-party providers?

Hai uses pre-trained LLMs from private, trusted providers (see subprocessors approved for AI processing activities here). When Hai sends data to an LLM for inference, it does so through secure, governed integrations. These providers process the data only to generate the requested output and do not retain it. Technical and contractual guardrails prevent our approved model providers from using input or output data for their own model improvements or training.

Does Hai use my data for GenAI training purposes?

As outlined above, Hai is a GenAI tool, meaning the agent uses your data to provide an answer using the pre-trained LLM. However, Hai does not train, fine-tune, or otherwise improve GenAI or large language models with customer or researcher data.

What kind of data do you process with Hai?

AI agents perform best with sufficient context. We use a progressive context model, where agents start with minimal context and gain access to additional context as needed and as authorized.

The system is modular, allowing agents to operate with limited input and expand context during task execution. The LLMs do not retain any data after the task is completed. All access is handled just-in-time (JIT) and can be audited through agent logs and traces.

The categories of data processed by AI include:

  • Vulnerability reports and metadata

  • Live application behavior (e.g., when performing reconnaissance or attempting exploits)

  • Context from connectors when granted (e.g., APIs, webhooks, or platform integrations), such as

    • Source code

    • Technical product or architectural documentation

    • Past incidents or issue tracker tickets

    • Infrastructure or cloud configuration data

    • Organizational hierarchy

How does Hai keep my data secure?

Hai ensures that your data remains under your control and is not shared outside HackerOne without your consent. HackerOne is ISO 27001, SOC 2, and FedRAMP certified, and GDPR compliant. Hai adheres to all existing security and compliance protocols applicable across the HackerOne platform. This includes strict authentication and authorization controls, governed integrations with vetted model providers, and safeguards that prevent customer and researcher data from being used to train or fine-tune third-party models. All inference requests are handled in a way that ensures data is processed securely, not retained by model providers, and not exposed across users or organizations.

Hai is subject to all of our existing high-level security and compliance protocols, which include:

  1. Role-based access controls (RBAC)

  2. System hardening

  3. Regular patching and maintenance

  4. Robust logging

  5. At least annual AI Red Teaming & Penetration Testing

But don't just take our word for it. We invite you and any third-party researchers to validate our controls by using Hai in our bug bounty program.

Does Hai provide explainable outputs and audit logs?

Yes. Hai maintains a record of all interactions through its conversation history feature. Questions and answers are stored on the HackerOne platform, allowing users with appropriate permissions to access historical conversations. This serves as an audit trail, capturing inputs and the corresponding outputs generated. Additionally, agent-based interactions (such as agentic validation or agentic security testing) are tracked via an agent log or agent trace, which provides a detailed account of what happened and why.

When generating responses to a user’s prompt, Hai uses only data from the program or programs that the user has permission to access. Hai adds context to help users understand the basis for its responses and ensures traceability for how conclusions are reached.

How is data stored and retained?

Data is stored and retained in accordance with HackerOne’s standard data retention policies, privacy policies, and terms. This data is safeguarded through our established security measures. Interactions are only stored on the platform to allow users with the proper permissions to access and view historical conversations related to their program.

How does HackerOne limit unintended GenAI outcomes (toxicity, hallucinations, bias)?

Our carefully selected model and AI inference providers, such as Anthropic and Amazon, manage change and release processes to identify, reduce, mitigate, and manage toxic outputs, hallucinations, and biases. The model/system cards provided by Amazon Titan, Anthropic Claude Opus 4.6, Anthropic Claude Sonnet 4.6, and Anthropic Claude Haiku 4.5 provide more information on these risks.

How does HackerOne train staff on AI regulatory requirements?

HackerOne employees receive training on AI best practices and regulatory obligations through our annual privacy training program. Additionally, in compliance with the EU AI Act, HackerOne meets AI Literacy training requirements for employees.

What is HackerOne's approach to AI risk tolerance and impact measurement?

HackerOne defines reasonable risk tolerances for AI systems, which are informed by laws, regulations, best practices, and industry standards. HackerOne also establishes policies to define mechanisms for measuring or understanding an AI system's potential impacts, e.g., via regular impact assessments at key stages in the AI lifecycle connected to system impacts and the frequency of system updates.

How does HackerOne govern the use of AI tools by Community Members?

Security researchers have quickly built, adapted, and improved the usage of emerging AI-based technologies across the platform. We call these AI-powered hacking tools Hackbots.

  • Community Members are creative, innovative, and independent service providers who may use assistive tools, like Hackbots, when participating in a customer’s program. Given the rise in the use of Hackbots, HackerOne has updated its guidelines for Community Members to reflect this innovation:

  • All Hackbots must operate within the published vulnerability disclosure guidelines of the programs they interact with and comply with HackerOne's Code of Conduct and Disclosure Guidelines.

  • AI tools are not allowed to operate fully autonomously. Our 'hacker-in-the-loop' model requires human experts to investigate, validate, and confirm all potential vulnerabilities before submitting them to any Vulnerability Disclosure or Bug Bounty Program.

  • Hackbot operators are fully responsible for their AI tools and must exercise due diligence to ensure compliance with platform rules and program guidelines.

  • Human operators using Hackbots qualify for applicable Rewards, just as if vulnerabilities were discovered through traditional means.

What is the Report Assistant Agent, and how does it support researchers?

The Report Assistant Agent on-platform AI capability helps Community Members improve clarity, completeness, and reproducibility when writing vulnerability reports. It works only with content provided by the researcher, does not submit reports autonomously, and preserves the researcher’s full responsibility and ownership for each submission. All outputs remain subject to platform rules, program guidelines, and human-in-the-loop requirements.

How does HackerOne ensure the responsible use of the Hai and researcher-operated AI tools?

HackerOne encourages careful human oversight and adherence to established disclosure practices. These guardrails support responsible innovation while protecting customers, researchers, and the broader ecosystem, and HackerOne continues to collaborate with the community to refine these practices as AI-accelerated hacking evolves.

How does HackerOne ensure AI providers meet your security requirements?

HackerOne applies the same rigorous security controls regardless of which AI provider processes a request. Every provider interaction flows through our authorization middleware, encryption layer, and audit logging. Your data receives identical protection regardless of which approved provider processes the request. The engine may differ, but the security envelope does not.

HackerOne maintains full transparency about AI providers through our published subprocessors list. Before any new provider can process customer data, they are formally added to this list. Customers subscribed to subprocessor notifications receive a 30-day notice of any additions, giving you visibility and time to review changes. This is the same process we use for all data processors—AI providers receive no special treatment or exemptions.

What if HackerOne changes its position on training LLMs with confidential customer data?

We understand and respect the significance of the confidential information you entrust to us. We do not currently train LLMs using confidential customer data. However, this is a fast-moving environment. If we ever consider doing so, we would only do so with customer permission.

How does HackerOne use AI beyond GenAI?

HackerOne leverages AI technologies beyond GenAI, including machine learning (ML) models and automation tools. These technologies enable systems to automatically learn and identify patterns, and to provide advanced capabilities such as predictive analytics, automation, personalization, and anomaly detection. For many years, HackerOne has used ML models and automation tools to analyze data, identify patterns, and improve accuracy in tasks such as vulnerability classification.

Did this answer your question?