[知识编辑] [ACL 2024] An Easy-to-use Knowledge Editing Framework for LLMs.
-
Updated
Jul 31, 2024 - Jupyter Notebook
[知识编辑] [ACL 2024] An Easy-to-use Knowledge Editing Framework for LLMs.
[知识编辑] Must-read Papers on Knowledge Editing for Large Language Models.
EMNLP'23 survey: a curation of awesome papers and resources on refreshing large language models (LLMs) without expensive retraining.
Official code repo for "Editing Implicit Assumptions in Text-to-Image Diffusion Models"
[知识编辑] [知识回路] Knowledge Circuits in Pretrained Transformers
Code for "Learning to Edit: Aligning LLMs with Knowledge Editing (ACL 2024)"
[ICLR 2024] Unveiling the Pitfalls of Knowledge Editing for Large Language Models
[ACL 2024] An Easy-to-use Hallucination Detection Framework for LLMs.
[知识编辑] [知识遗忘] To Forget or Not? Towards Practical Knowledge Unlearning for Large Language Models
Official codes for COLING 2024 paper "Robust and Scalable Model Editing for Large Language Models": https://arxiv.org/abs/2403.17431v1
Stable Knowledge Editing in Large Language Models
Debiasing Stereotyped Language Models via Model Editing
An Automated Framework to Construct Datasets for Assessing Knowledge Editing or Multi-Hop Reasoning Capability of Language Models.
MLaKE: Multilingual Knowledge Editing Benchmark for Large Language Models
Add a description, image, and links to the knowledge-editing topic page so that developers can more easily learn about it.
To associate your repository with the knowledge-editing topic, visit your repo's landing page and select "manage topics."