item 1 out of 1
- Author
- Year
- 2021
- Title
- HedgeCut: Maintaining Randomised Trees for Low-Latency Machine Unlearning
- Event
- 2021 International Conference on the Management of Data
- Book/source title
- SIGMOD '21
- Book/source subtitle
- proceedings of the 2021 International Conference on the Management of Data : June 20 -25, 2021, virtual event, China
- Pages (from-to)
- 1545–1557
- Publisher
- New York, NY: Association for Computing Machinery
- ISBN (electronic)
- 9781450383431
- Document type
- Conference contribution
- Faculty
- Faculty of Science (FNWI)
- Institute
- Informatics Institute (IVI)
- Abstract
-
Software systems that learn from user data with machine learning (ML) have become ubiquitous over the last years. Recent law such as the "General Data Protection Regulation" (GDPR) requires organisations that process personal data to delete user data upon request (enacting the "right to be forgotten"). However, this regulation does not only require the deletion of user data from databases, but also applies to ML models that have been learned from the stored data. We therefore argue that ML applications should offer users to unlearn their data from trained models in a timely manner. We explore how fast this unlearning can be done under the constraints imposed by real world deployments, and introduce the problem of low-latency machine unlearning: maintaining a deployed ML model in-place under the removal of a small fraction of training samples without retraining.We propose HedgeCut, a classification model based on an ensemble of randomised decision trees, which is designed to answer unlearning requests with low latency. We detail how to efficiently implement HedgeCut with vectorised operators for decision tree learning. We conduct an experimental evaluation on five privacy-sensitive datasets, where we find that HedgeCut can unlearn training samples with a latency of around 100 microseconds and answers up to 36,000 prediction requests per second, while providing a training time and predictive accuracy similar to widely used implementations of tree-based ML models such as Random Forests.
- URL
- go to publisher's site
- Language
- English
- Note
- With supplementary material
- Persistent Identifier
- https://hdl.handle.net/11245.1/f3cce3c7-cf8b-4ffe-8bac-a2ae57298f0a
- Downloads
-

3448016.3457239(Final published version)
- Supplementary materials
Disclaimer/Complaints regulations
If you believe that digital publication of certain material infringes any of your rights or (privacy) interests, please let the Library know, stating your reasons. In case of a legitimate complaint, the Library will make the material inaccessible and/or remove it from the website. Please Ask the Library, or send a letter to: Library of the University of Amsterdam, Secretariat, Singel 425, 1012 WP Amsterdam, The Netherlands. You will be contacted as soon as possible.

UvA