msakarvadia/memorization

Localizing Memorized Sequences in Language Models

36
/ 100
Emerging

This project helps AI developers and researchers prevent large language models (LLMs) from inadvertently revealing private or sensitive training data during inference. It takes a pre-trained language model and identifies and removes memorized sequences without significantly impacting the model's overall performance. AI engineers concerned with data privacy and security in their LLM deployments are the primary users.

Use this if you need to precisely remove specific memorized information from a language model's weights to enhance data privacy and mitigate risks of sensitive data exposure.

Not ideal if you are looking for general model fine-tuning or regularization methods that are not specifically focused on pinpointing and removing memorized data.

AI-safety data-privacy LLM-security model-unlearning AI-ethics
No Package No Dependents
Maintenance 6 / 25
Adoption 6 / 25
Maturity 16 / 25
Community 8 / 25

How are scores calculated?

Stars

20

Forks

2

Language

Jupyter Notebook

License

MIT

Last pushed

Oct 15, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/msakarvadia/memorization"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.