SAFR AI Lab @ Harvard
Pinned Loading
Repositories
- pandora-llm Public
An extensible library for evaluating data extraction attacks and MIAs against LLMs. Developed as part of the pandora llm paper, linked below.
safr-ai-lab/pandora-llm’s past year of commit activity - survey-llm Public
A survey of privacy problems in Large Language Models (LLMs). Contains summary of the corresponding paper along with relevant code
safr-ai-lab/survey-llm’s past year of commit activity - xai-disparity Public
safr-ai-lab/xai-disparity’s past year of commit activity - In-Context-Unlearning Public
"In-Context Unlearning: Language Models as Few Shot Unlearners". Martin Pawelczyk, Seth Neel* and Himabindu Lakkaraju*; arXiv preprint: arXiv:2310.07579; 2023.
safr-ai-lab/In-Context-Unlearning’s past year of commit activity - CounterfactualDistanceAttack Public
"On the Privacy Risks of Algorithmic Recourse". Martin Pawelczyk, Himabindu Lakkaraju* and Seth Neel*. In International Conference on Artificial Intelligence and Statistics (AISTATS), PMLR, 2023.
safr-ai-lab/CounterfactualDistanceAttack’s past year of commit activity - GerryFair Public Forked from sethneel/GerryFair
Package implementing methods developed in "Preventing Fairness Gerrymandering" [ICML '18], "Rich Subgroup Fairness for Machine Learning" [ FAT* '19]. active development fork @algowatchupenn
safr-ai-lab/GerryFair’s past year of commit activity
People
This organization has no public members. You must be a member to see who’s a part of this organization.
Top languages
Loading…
Most used topics
Loading…