Security and Privacy (TRAPS)

Aisera's proprietary security system protects your applications and data against attacks

Aisera provides a proprietary security system that uses a Trusted, Responsible, Auditable, Private, and Secure (TRAPS) framework to protect your applications against attacks.

Prompt Injection Attacks

Protecting against prompt injection attacks is crucial to ensure the security and reliability of interactions with Al models. Prompt injection attacks occur when malicious or inappropriate content is inserted into the user's input, attempting to manipulate the AI into generating harmful or undesirable outputs.

Aisera uses the following strategies to protect against prompt injection attacks:

Input Validation and Sanitization

Aisera has implemented input validation and sanitization mechanisms through a Content Moderation Service to automatically detect and filter out or sanitize any potentially harmful input. This involves checking for specific patterns, characters, or keywords that might indicate an attack. The validation happens both on the ingress gateway (WAF) but also inside our service itself.

Predefined Prompts

Instead of allowing free-form inputs, our pipelines utilize system prompts that have been validated at the factory. This restricts the input to a predefined context, and minimizes the risk of malicious injections.

Model Poisoning Attacks

Protecting Al models against model poisoning attacks is crucial to ensure the integrity and reliability of the model's predictions. Model poisoning involves introducing malicious data during the training process with the goal of compromising the model's performance or behavior.

Aisera uses the following strategies to protect from model poisoning attacks:

Data Sanitization and Validation

Aisera has implemented data validation and sanitization to detect and filter out potentially malicious or poisoned data before it enters the training process. Aisera also uses anomaly detection techniques to identify abnormal patterns or outliers in the training data that might indicate poisoning attempts.

Data Audit Trail

Aisera maintains an audit trail of data used during training to trace the origin of data points and identify potential poisoning sources.

Data Curation/Human in the loop

Any data that will be eligible for the model training phase must go through a human curation process where users with specific roles in the organization review the data to be admitted into the various Al models.

Input Validation at Inference

For most Al models, Aisera has implemented input validation and filtering mechanisms at inference time to detect and reject potentially poisoned inputs.

Regularization and Weight Clipping

Aisera also applies regularization techniques to the model's training process to reduce the impact of individual data points or small subsets. This includes the use of weight clipping to bound the model's weights and prevent extreme values due to potential malicious data.

Model Verification

Aisera's system allows for regular testing of a model's performance on a separate validation dataset to identify any sudden drops in accuracy or unexpected behavior. The Al lens capability allows for testing inputs and outputs and validating bot behavior.

Ensemble Models

In many cases, Aisera uses ensemble techniques by combining predictions from multiple models. This makes it harder for attackers to consistently manipulate predictions.

Continuous Monitoring:

Our platform continuously monitors the model's performance and behavior in production to detect any signs of compromised predictions.

Secure Model Deployment

Aisera deploys models in secure environments with proper access controls and monitors incoming queries for unusual patterns.

Adversarial Input Attacks

Protecting Al models against adversarial attacks is a critical aspect of ensuring their security and robustness. Adversarial attacks involve manipulating input data in subtle ways to cause Al models to make incorrect predictions or generate unintended outputs.

Aisera uses the following safeguards to protect Al models against adversarial attacks:

Feature Transformation

Aisera applies input preprocessing and feature transformations to input data to reduce the effectiveness of adversarial attacks. These transformations can help filter out noise or malicious perturbations. In these feature transformations, Aisera's Personally Identifiable Information anonymization layer plays a significant role in limiting these attacks.

Regularization

Using Aisera's Knowledge graph and robust ontologies, Aisera applies regularization techniques during model training to encourage the model to generalize better and resist overfitting to adversarial perturbations.

Ensemble Models

In many cases, Aisera uses ensemble techniques by combining predictions from multiple models. Adversarial attacks may have a harder time finding consistent vulnerabilities across different models.

Input Validation

As mentioned earlier, Aisera has implemented input validation and sanitization mechanisms to filter out or correct potentially adversarial inputs. This can involve checking for anomalies or patterns commonly associated with adversarial attacks.

Out-of-Distribution Detection

In most cases, Aisera's models are given certain boundaries under which they can operate. These include specific topics/domains that need to be registered. As a result, they can detect when the input data is out-of-distribution or deviates significantly from the training data. This can help identify potential adversarial inputs.

Secure Deployment

Aisera has implemented proper security measures for deploying Al models, such as access controls, model versioning, and monitoring of incoming requests for unusual patterns.

Last updated

Was this helpful?