Tool permission management
Validation of tools before use
Defensive methods to prevent compromise of causal elements
Personal information masking
Model Extraction Attack Detection
Access control for system prompts
Detection of adversarial examples
Backdoor Trigger Detections
Poisoned Data Detection for Backdoor Attacks
Detection of backdoor models
Bias Detection in AI Outputs
Elimination of bias in training data
Creation of fair AI models
Uncertainty Quantification
RAG (Retrieval-Augmented Generation)
Fairness evaluation of models
Defensive methods to prevent compromise of causal elements (upstream factors)