Misplaced Pages

Machine unlearning: Difference between revisions

Article snapshot taken from Wikipedia with creative commons attribution-sharealike license. Give it a read and then ask your questions in the chat. We can research this topic together.
Browse history interactively← Previous editContent deleted Content addedVisualWikitext
Revision as of 09:39, 10 December 2024 editDancingPhilosopher (talk | contribs)Extended confirmed users5,601 edits History← Previous edit Latest revision as of 14:15, 25 December 2024 edit undoCielquiparle (talk | contribs)Autopatrolled, Extended confirmed users, New page reviewers34,231 edits +refTag: Visual edit 
(11 intermediate revisions by 4 users not shown)
Line 1: Line 1:
{{unreferenced|date=December 2024}} {{Short description|Field of study in artificial intelligence}}{{More sources|date=December 2024}}

'''Machine unlearning''' is a branch of machine learning focused on removing specific undesired element, such as private data, outdated information, copyrighted material, harmful content, dangerous abilities, or misinformation, without needing to rebuild models from the ground up. '''Machine unlearning''' is a branch of ] focused on removing specific undesired element, such as private data, outdated information, copyrighted material, harmful content, dangerous abilities, or misinformation, without needing to rebuild models from the ground up.

Large language models, like the ones powering ], may be asked not just to remove specific elements but also to unlearn a "concept," "fact," or "knowledge," which aren't easily linked to specific examples. New terms such as "model editing," "concept editing," and "knowledge unlearning" have emerged to describe this process.<ref name="Liu_2024">{{Cite web |title=Machine Unlearning in 2024 |url=https://ai.stanford.edu/~kzliu/blog/unlearning |archive-url=http://web.archive.org/web/20241213234527/https://ai.stanford.edu/~kzliu/blog/unlearning |archive-date=2024-12-13 |access-date=2024-12-24 |website=Ken Ziyu Liu - Stanford Computer Science |language=en-US}}</ref>

== History == == History ==
Early research efforts were largely motivated by Article 17 of the ], the European Union's privacy regulation commonly known as the "right to be forgotten" (RTBF), introduced in 2014. RTBF was not designed with machine learning in mind. In 2014, policymakers couldn’t foresee the complexity of deep learning’s data-computation mix, making data erasure challenging. This challenge later spurred research into “data deletion” and “machine unlearning.” Early research efforts were largely motivated by Article 17 of the ], the European Union's privacy regulation commonly known as the "right to be forgotten" (RTBF), introduced in 2014.<ref>{{Cite journal |last=Hine |first=E. |last2=Novelli |first2=C. |last3=Taddeo |first3=M. |date=2024 |title=Supporting Trustworthy AI Through Machine Unlearning |journal=Science Engineering & Ethics |volume=30 |issue=43 |doi=10.1007/s11948-024-00500-5}}</ref>


==Present==
Following the deployment of large language models, unlearning is driven by more than just user privacy. The focus has shifted from training small networks on face images to large models trained on harmful content that may need to be "erased."
The GDPR did not anticipate that the development of ]s would make data erasure a complex task. This issue has since led to research on "machine unlearning," with a growing focus on removing copyrighted material, harmful content, dangerous capabilities, and misinformation. Just as early experiences in humans shape later ones, some concepts are more fundamental and harder to unlearn. A piece of knowledge may be so deeply embedded in the model’s knowledge graph that unlearning it could cause internal contradictions, requiring adjustments to other parts of the graph to resolve them.{{Citation needed|date=December 2024}}


== References == == References ==
{{Reflist}} {{Reflist}}
] ]
{{improve categories|date=December 2024}}

Latest revision as of 14:15, 25 December 2024

Field of study in artificial intelligence
This article needs additional citations for verification. Please help improve this article by adding citations to reliable sources. Unsourced material may be challenged and removed.
Find sources: "Machine unlearning" – news · newspapers · books · scholar · JSTOR (December 2024) (Learn how and when to remove this message)

Machine unlearning is a branch of machine learning focused on removing specific undesired element, such as private data, outdated information, copyrighted material, harmful content, dangerous abilities, or misinformation, without needing to rebuild models from the ground up.

Large language models, like the ones powering ChatGPT, may be asked not just to remove specific elements but also to unlearn a "concept," "fact," or "knowledge," which aren't easily linked to specific examples. New terms such as "model editing," "concept editing," and "knowledge unlearning" have emerged to describe this process.

History

Early research efforts were largely motivated by Article 17 of the GDPR, the European Union's privacy regulation commonly known as the "right to be forgotten" (RTBF), introduced in 2014.

Present

The GDPR did not anticipate that the development of large language models would make data erasure a complex task. This issue has since led to research on "machine unlearning," with a growing focus on removing copyrighted material, harmful content, dangerous capabilities, and misinformation. Just as early experiences in humans shape later ones, some concepts are more fundamental and harder to unlearn. A piece of knowledge may be so deeply embedded in the model’s knowledge graph that unlearning it could cause internal contradictions, requiring adjustments to other parts of the graph to resolve them.

References

  1. "Machine Unlearning in 2024". Ken Ziyu Liu - Stanford Computer Science. Archived from the original on 2024-12-13. Retrieved 2024-12-24.
  2. Hine, E.; Novelli, C.; Taddeo, M. (2024). "Supporting Trustworthy AI Through Machine Unlearning". Science Engineering & Ethics. 30 (43). doi:10.1007/s11948-024-00500-5.
Category: