Skip to main content

Overview

Dynamo AI’s penetration testing module is an automated model vulnerability assessment tool that can be used to assess risks related to model privacy, hallucination, performance and compliance. The penetration testing module can either be used as a no-code tool directly in the Dynamo AI platform or through the Dynamo AI python SDK solution. Currently, Dynamo AI offers 4 types of assessments: privacy, hallucination, compliance and performance. Each assessment type includes a variety of unique test types to assess vulnerabilities.

Attacks and Tests

DynamoEval’s suite of attacks

Test Inventory

Test CategoryTest NameTest DetailSample Test Report
Privacy TestPII ExtractionTests whether these PII can be emitted from prompting the model naively and simulates an attacker with no knowledge of the training datasetView Report
PII InferenceTests whether a model can re-fill PII into sentences from the fine-tuned dataset that we redacted PII from, assuming an attacker with knowledge of the concepts and potential PII in the datasetView Report
PII ReconstructionTests whether a model can re-fill PII into sentences from the fine-tuned dataset that we redacted PII from, assuming a partially-informed attackerView Report
Sequence ExtractionTests whether the model can be prompted in a manner where it reveals training data verbatim as part of the responsesView Report
Membership InferenceDetermine whether specific data records can be inferred to be a part of the model's training datasetView Report
HallucinationNLI ConsistencyMeasures the logical consistency between an input text (or document) and a model-generated summary with an NLI model.View Report
UniEval FactualityMeasures the factual support between an input text (or document) and a model-generated summary with UniEval model.View Report
RAG HallucinationFaithfulnessFaithfulness represents how faithfulness model generated responses are to the set of retrieved documents.View Report
Retrieval RelevanceRetrieval relevance represents the relevance of the documents retrieved from the vector database using the embedding model for each query.View Report
Response RelevanceResponse relevance measures how relevant the generated response is to the input query.View Report
ComplianceCybersecurity ComplianceCheck whether model responses don't contain malicious code based on the MITRE frameworkView Report
Static Jailbreak tests frameworkTests whether the model is prone to jailbreaking attacks and if the model responses are compliantView Report
Bias/Toxicity tests frameworkTests whether the model can be forced to emit toxic or biased contentView Report
Adaptive jailbreak attacksTests whether the model is prone to jailbreaks using algorithmically generated promptsView Report
PerformanceROUGEROUGE is a metric designed to evaluate summary quality against a reference text by measuring the token-level overlap between the text generated by a model and the references.View Report
BERTScoreBERTScore computes semantic similarity between a reference text and model response, leveraging the pre-trained embeddings from the BERT model and is represented by a precision, recall, and F1 score.View Report

Components

DynamoEval Components

Architecture

DynamoEval architecture