Skip to content

AI Model Security

5 automated security scanners


Purpose: The Inference Attack Prevention Scanner is designed to detect membership inference and attribute inference attacks by analyzing company security documentation, public policy pages, trust center information, and compliance certifications. It aims to identify gaps in data protection and access control measures to ensure robust security practices.

What It Detects:

  • Security Policy Indicators: Identifies the presence or absence of explicit security policies, detailed incident response plans, comprehensive data protection frameworks, and robust access control mechanisms.
  • Maturity Indicators: Confirms SOC 2 compliance certification, validates ISO 27001 standards adherence, evaluates regular penetration testing activities, and assesses vulnerability scanning and assessment practices.
  • Data Protection Language Analysis: Searches for specific phrases indicating strong data protection measures, identifies detailed descriptions of data handling procedures, detects mentions of encryption methods and key management, and verifies data retention and disposal policies.
  • Access Control Verification: Looks for explicit access control policies, checks for role-based access control (RBAC) implementation details, evaluates multi-factor authentication (MFA) requirements, and identifies user privilege management practices.
  • Compliance Certifications: Confirms the presence of relevant compliance certifications, validates that certifications are up-to-date and valid, checks for transparency in certification processes, and ensures compliance with industry-specific regulations.

Inputs Required:

  • domain (string): Primary domain to analyze (e.g., acme.com)
  • company_name (string): Company name for statement searching (e.g., “Acme Corporation”)

Business Impact: This scanner is crucial as it helps organizations identify vulnerabilities in their security practices that could be exploited through membership or attribute inference attacks, potentially leading to significant data breaches and loss of trust among users.

Risk Levels:

  • Critical: Conditions where explicit security policies are absent, detailed incident response plans are not comprehensive, data protection frameworks are incomplete, or access control mechanisms are insufficiently described.
  • High: Conditions where maturity indicators such as SOC 2 compliance certification or ISO 27001 standards adherence are lacking, or regular penetration testing activities and vulnerability scanning/assessment practices are inadequate.
  • Medium: Conditions where data protection language is vague, encryption methods and key management are not detailed, or user privilege management practices need improvement.
  • Low: Conditions where compliance certifications are outdated or invalid, transparency in certification processes is lacking, but the overall risk to security remains manageable without immediate attention.
  • Info: Conditions that do not significantly impact security but still warrant awareness for informational purposes (e.g., minor gaps in data protection language).

Example Findings:

  1. A company lacks a detailed privacy policy and does not mention multi-factor authentication requirements, indicating a medium risk of unauthorized access through attribute inference attacks.
  2. The trust center document mentions outdated PCI DSS compliance without verification of the current status, suggesting a need for immediate attention to avoid critical risks in data protection.

Purpose: The Adversarial Example Testing Scanner is designed to detect input manipulation and classification evasion techniques used by adversaries to deceive machine learning models. This tool helps identify vulnerabilities in AI systems that could be exploited to alter model predictions or bypass security measures.

What It Detects:

  • Input Manipulation Patterns:
    • Tests for common adversarial perturbations such as small pixel changes, noise injection, and crafted inputs designed to mislead classifiers.
    • Verifies the presence of adversarial examples in test datasets.
    • Detects patterns indicative of input tampering like unusual character sequences or unexpected data anomalies.
  • Classification Evasion Techniques:
    • Tests for evasion strategies that bypass security mechanisms by manipulating model confidence scores and attempts to use adversarial training data to fool models.
    • Detects signs of targeted attacks aimed at specific classes or outputs.
  • Data Poisoning Indicators:
    • Tests for poisoned datasets containing malicious samples, patterns indicative of data tampering during training, and the presence of backdoor triggers in model inputs.
    • Detects attempts to introduce bias or skew in training data.
  • Model Robustness Checks:
    • Tests for robustness against adversarial attacks by checking vulnerabilities in feature extraction processes and verifying the effectiveness of defensive mechanisms like adversarial training or input sanitization.
    • Detects weaknesses that could be exploited by adversaries to manipulate model behavior.
  • Attack Surface Analysis:
    • Identifies exposed endpoints or interfaces susceptible to adversarial inputs, insecure data handling practices, and gaps in security policies related to AI model deployment and maintenance.

Inputs Required:

  • domain (string): Primary domain to analyze (e.g., acme.com)
  • company_name (string): Company name for statement searching (e.g., “Acme Corporation”)

Business Impact: This scanner is crucial as it helps organizations identify and mitigate vulnerabilities in their AI systems that could be exploited by adversaries to manipulate model predictions or bypass security measures. By detecting input manipulation and classification evasion techniques, the scanner contributes significantly to enhancing the overall security posture of machine learning models used within an organization.

Risk Levels:

  • Critical: Conditions where there is a high likelihood of significant damage or disruption due to adversarial attacks on critical systems.
  • High: Conditions where there is a moderate risk of substantial negative impact from adversarial actions, potentially affecting key business functions.
  • Medium: Conditions where the risk is low but still notable, such as minor system disruptions that could affect operational efficiency.
  • Low: Conditions with minimal risk or no discernible impact on systems or operations.
  • Info: Informational findings indicating potential areas for improvement in security practices without immediate critical risks.

Example Findings: The scanner might flag instances of unauthorized access attempts to model inputs, data tampering during training processes that could lead to misclassification, and exposure points where adversarial examples can be injected into the system.


Purpose: The Model Poisoning Analysis Scanner is designed to detect potential issues related to training data poisoning and backdoor insertion in machine learning models. It evaluates various aspects of a company’s documentation, including security policies, maturity indicators, practices regarding training data security, and procedures for detecting model backdoors. This ensures that organizations maintain robust defenses against unauthorized tampering with their datasets and hidden vulnerabilities within models.

What It Detects:

  • Security Policy Indicators: The scanner identifies the presence or absence of comprehensive security policies, detailed incident response plans, data protection measures, and proper access control mechanisms.
  • Maturity Indicators: This includes confirming adherence to SOC 2 standards, validating ISO 27001 compliance certifications, evaluating the frequency and thoroughness of penetration testing, and assessing vulnerability scanning and assessment practices.
  • Training Data Security Practices: It looks for specific mentions of security protocols related to training data, checks for procedures involving data validation and integrity checks, verifies measures in place to detect anomalies in training datasets, and ensures proper logging and monitoring of data handling processes.
  • Backdoor Detection Indicators: The scanner identifies policies or practices designed to identify backdoors within models, including regular audits of model code and dependencies, the use of secure coding standards and best practices, and the presence of intrusion detection systems for maintaining model integrity.
  • Incident Response to Model Poisoning: It assesses the organization’s ability to respond effectively to incidents involving model poisoning, checks for specific procedures to isolate and mitigate poisoned models, verifies communication plans with stakeholders in case of a security breach, and ensures that lessons learned from past incidents are documented and applied.

Inputs Required:

  • domain (string): Primary domain to analyze (e.g., acme.com) - This is necessary for the scanner to gather information from the specified company website.
  • company_name (string): Company name for statement searching (e.g., “Acme Corporation”) - Used in search queries to find relevant documents and policies related to the specific organization.

Business Impact: Ensuring that machine learning models are secure against tampering is crucial as it directly impacts the integrity and reliability of AI applications used across various industries, including healthcare, finance, and government services. Poor security practices can lead to unauthorized access, data breaches, and potentially harmful outcomes for both organizations and users.

Risk Levels:

  • Critical: Conditions that would result in critical severity include significant vulnerabilities that could directly affect the functionality or integrity of machine learning models, leading to potential harm or compliance violations.
  • High: High-risk findings involve substantial security gaps that could be exploited by malicious actors, potentially compromising data and system integrity.
  • Medium: Medium-severity risks are those with notable weaknesses in security practices but less severe than critical issues. These still require attention for remediation to prevent escalation into higher risk categories.
  • Low: Low-risk findings represent minor or non-critical areas where improvements could be made, generally not affecting the core functionality of the models significantly.
  • Info: Informational findings are observations that do not directly impact security but may suggest areas for improvement in compliance and transparency.

Example Findings:

  1. A company lacks a detailed security policy document, which is critical for understanding how they handle data protection and incident response.
  2. The organization’s ISO 27001 certification has lapsed without explanation, indicating potential gaps in ongoing compliance with industry standards.

Purpose: The Model Interpretability Scanner is designed to identify black-box issues and explainability gaps in AI models by analyzing a company’s security documentation, public policy pages, trust center information, and compliance certifications. This tool ensures that organizations provide adequate transparency and accountability regarding their AI model deployments, thereby enhancing the security posture of the organization.

What It Detects:

  • Security Policy Indicators: Identifies the presence or absence of specific security policies such as “security policy,” “incident response,” “data protection,” and “access control.”
  • Maturity Indicators: Checks for compliance certifications and maturity indicators like SOC 2, ISO 27001, penetration testing, and vulnerability scanning.
  • Documentation Accessibility: Evaluates the accessibility of company security documentation on their website.
  • Trust Center Information: Analyzes trust center information to ensure it includes necessary security disclosures and compliance certifications.
  • Public Policy Pages: Scraps public policy pages for relevant security-related content and compliance indicators.

Inputs Required:

  • domain (string): Primary domain to analyze (e.g., acme.com)
  • company_name (string): Company name for statement searching (e.g., “Acme Corporation”)

Business Impact: This scanner is crucial as it helps organizations maintain a robust security framework around their AI models, ensuring compliance with industry standards and enhancing trust among stakeholders by providing clear explanations of model decisions and behaviors.

Risk Levels:

  • Critical: The scanner flags severe issues that directly impact the core functionality or security of the AI models, such as missing critical security policies or certifications not meeting stringent requirements.
  • High: High severity findings indicate significant gaps in compliance or transparency, which could lead to substantial risks if exploited by malicious actors.
  • Medium: Medium severity findings suggest potential issues that may require attention but do not pose immediate high risk.
  • Low: Low severity findings are informational and typically involve minor non-compliance areas that can be addressed through ongoing improvements in documentation and policy adherence.
  • Info: These are purely informative, providing insights into the current state of compliance without being classified as critical or high risks.

Example Findings:

  1. A company lacks a comprehensive security policy document, which could lead to vulnerabilities if unauthorized access occurs.
  2. The trust center does not include all necessary data protection certifications, potentially affecting user confidence in the organization’s handling of sensitive information.

Purpose: The Model Stealing Protection Scanner is designed to identify and assess vulnerabilities in machine learning model protection strategies by analyzing company documentation, including public policy pages, trust center information, and compliance certifications. This tool helps companies ensure robust measures are in place to safeguard their AI models against unauthorized extraction and probing.

What It Detects:

  • Model Extraction Indicators: The scanner looks for mentions of model extraction techniques such as “model stealing” and “model inference attacks,” as well as descriptions of potential vulnerabilities related to these techniques. It also checks for evidence of inadequate protection mechanisms against theft.
  • API Probing Patterns: This includes detecting references to API probing activities, unauthorized access attempts or suspicious API usage, and the presence of necessary security measures like API rate limiting and authentication.
  • Security Policy Indicators: The scanner searches for key phrases related to comprehensive security policies including “security policy,” “incident response,” “data protection,” and “access control.” It ensures that these policies are publicly accessible and in place.
  • Compliance Certifications: Identifies references to recognized security standards like SOC 2, ISO 27001, as well as mentions of regular assessments such as penetration testing and vulnerability scanning.
  • Trust Center Information: Analyzes trust center pages for detailed information on data protection measures, incident response plans, and security practices, ensuring transparency in handling sensitive data and responding to security incidents.

Inputs Required:

  • domain (string): The primary domain of the company website to be analyzed.
  • company_name (string): The name of the company for which the analysis is being conducted, used for statement searching.

Business Impact: This scanner plays a crucial role in enhancing the security posture of companies by proactively identifying and addressing vulnerabilities related to machine learning model protection and API security. It helps ensure that sensitive AI models are safeguarded against potential threats, maintaining competitive advantage while complying with regulatory requirements.

Risk Levels:

  • Critical: The risk is critical if there are explicit mentions or indications of severe vulnerabilities in the model protection mechanisms, such as undocumented APIs being used for unauthorized access without proper authentication and authorization checks.
  • High: High risks are identified when there are clear descriptions of potential exploits related to model extraction techniques, inadequate security policies, or lack of compliance with recognized standards like SOC 2 or ISO 27001.
  • Medium: Medium risk conditions arise from the presence of vulnerabilities that could be exploited with some effort but have not been fully documented in policy statements, such as partial implementation of API rate limiting or incomplete penetration testing reports.
  • Low: Low risks are associated with general mentions of security best practices without specific details indicating immediate threats, such as occasional discussions about future enhancements to access controls or planned updates to compliance certifications.
  • Info: Informational findings include generic statements about the importance of security and vague references to ongoing efforts in model protection and API security that do not directly indicate vulnerabilities.

Example Findings:

  • “Our data breach response plan is currently under review, but we have established basic incident response protocols.”
  • “We have implemented a token-based authentication system for our APIs, though detailed documentation is still pending.”

This structured output format provides clear and actionable insights into the security measures in place at a company, highlighting both strengths and areas needing improvement.