Repository navigation

#

ml-safety

PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to adversarial prompt attacks. 🏆 Best Paper Awards @ NeurIPS ML Safety Workshop 2022

Python
361
1 年前

Aligning AI With Shared Human Values (ICLR 2021)

Python
284
2 年前
Python
266
4 年前

ImageNet-R(endition) and DeepAugment (ICCV 2021)

Python
264
4 年前

Repo for "Benchmarking Robustness of 3D Point Cloud Recognition against Common Corruptions" https://arxiv.org/abs/2201.12296

Python
207
2 年前

The Combined Anomalous Object Segmentation (CAOS) Benchmark

Python
157
2 年前

Pre-Training Buys Better Robustness and Uncertainty Estimates (ICML 2019)

Python
100
3 年前

[ICML 2019] ME-Net: Towards Effective Adversarial Robustness with Matrix Estimation

Python
54
1 个月前

Jiminy Cricket Environment (NeurIPS 2021)

ZAP
25
3 年前

Predicting Out-of-Distribution Error with the Projection Norm

Python
17
3 年前

Code for the attack multiplicative filter attack MUFIA, from the paper "Frequency-based vulnerability analysis of deep learning models against image corruptions".

Python
3
2 年前

AAAI 2025 Tutorial on Machine Learning Safety

3
2 个月前

This repository contains the project for the Advanced AI course @CentraleSupélec

Jupyter Notebook
2
3 年前