Skip to content
You signed in with another tab or window. to refresh your session.
You signed out in another tab or window. to refresh your session.
You switched accounts on another tab or window. to refresh your session.
Here are
3 public repositories
matching this topic...
VISION is a framework for robust and interpretable code vulnerability detection using counterfactual data augmentation. It leverages GNNs, LLM-generated counterfactuals, and graph-based explainability to mitigate spurious correlations and improve generalization on real-world vulnerabilities (CWE-20).
A Framework for Robust, Self-Recovering Tool-Using Language Model Agents — trained on 50K+ failure-annotated trajectories for fault-tolerant reasoning and recovery.
Investigating the "Gradient Noise Paradox" in AI Safety: A study on the conflict between Differential Privacy (DP-SGD) and Adversarial Training. Uses a custom "Shadow Model" pipeline to synchronize Opacus with PGD attacks, demonstrating how privacy-preserving noise systematically degrades model robustness
Improve this page
Add a description, image, and links to the
ai-robustness
topic page so that developers can more easily learn about it.
Curate this topic
Add this topic to your repo
To associate your repository with the
ai-robustness
topic, visit your repo's landing page and select "manage topics."
Learn more
You can’t perform that action at this time.