• StartupHub.ai
    StartupHub.aiAI Intelligence
Discover
  • Home
  • Search
  • Trending
  • News
Intelligence
  • Market Analysis
  • Comparison
  • Market Map
Workspace
  • Email Validator
  • Pricing
Company
  • About
  • Editorial
  • Terms
  • Privacy
  • v1.0.0
  1. Home
  2. News
  3. Deepkeep Launches Generative Ai Risk Assessment Module
Back to News
Press release

DeepKeep Launches Generative AI Risk Assessment Module

S
StartupHub Team
Jun 27, 2024 at 2:19 PM2 min read
DeepKeep Launches Generative AI Risk Assessment Module

DeepKeep, the leading provider of AI-native trust, risk, and security management, announces the product launch of its Generative AI Risk Assessment module, designed to secure Large Language Models (LLMs) and computer vision models, specifically focusing on penetration testing, identifying potential vulnerabilities and threats to model security, trustworthiness and privacy.

Assessing and mitigating AI model and application vulnerabilities ensures implementations are compliant, fair and ethical. DeepKeep's Risk Assessment module offers a comprehensive ecosystem approach by considering risks associated with model deployment, and identifying application weak spots.

DeepKeep’s assessment provides a thorough examination of AI models, ensuring high standards of accuracy, integrity, fairness, and efficiency. The module is helping security teams streamline Generative AI deployment processes, granting a range of scoring metrics for evaluation.

Core features include:

  • Penetration Testing
  • Identifying the model’s tendency to hallucinate
  • Identifying the model’s propensity to leak private data
  • Assessing toxic, offensive, harmful, unfair, unethical, or discriminatory language
  • Assessing biases and fairness
  • Weak spot analysis
LLM risk assessment measures. For example, when applying DeepKeep’s Risk Assessment module to Meta’s LLM LlamaV2 7B to examine prompt manipulation sensitivity, findings pointed to a weakness in English-to-French translation. Credit: DeepKeep.

“The market must be able to trust its GenAI models, as more and more enterprises incorporate GenAI into daily business processes," says Rony Ohayon, DeepKeep’s CEO and Founder. "Evaluating model resilience is paramount, particularly during its inference phase in order to provide insights into the model's ability to handle various scenarios effectively. DeepKeep’s goal is to empower businesses with the confidence to leverage GenAI technologies while maintaining high standards of transparency and integrity."

DeepKeep’s Generative AI Risk Assessment module secures AI alongside its AI Firewall, enabling live protection against attacks on AI applications. Detection capabilities cover a wide range of security and safety categories, leveraging DeepKeep’s proprietary technology and cutting-edge research.

*ROUGE and METEOR are natural language processing (NLP) techniques for evaluating machine learning outputs. Scores range between 0-1, with 1 indicating perfection.

#Computer Vision
#DeepKeep
#Generative AI
#Large Language Models (LLMs)
#Risk Analysis

AI Daily Digest

Get the most important AI news daily.

GoogleSequoiaOpenAIa16z
+40k readers