Show simple item record

dc.contributor.author
Dimitrov, Dimitar Iliev
dc.contributor.author
Balunović, Mislav
dc.contributor.author
Konstantinov, Nikola
dc.contributor.author
Vechev, Martin
dc.date.accessioned
2023-01-19T13:40:55Z
dc.date.available
2023-01-19T10:31:20Z
dc.date.available
2023-01-19T13:40:55Z
dc.date.issued
2022-11
dc.identifier.issn
2835-8856
dc.identifier.uri
http://hdl.handle.net/20.500.11850/593553
dc.identifier.doi
10.3929/ethz-b-000593553
dc.description.abstract
Recent attacks have shown that user data can be recovered from FedSGD updates, thus breaking privacy. However, these attacks are of limited practical relevance as federated learning typically uses the FedAvg algorithm. Compared to FedSGD, recovering data from FedAvg updates is much harder as: (i) the updates are computed at unobserved intermediate network weights, (ii) a large number of batches are used, and (iii) labels and network weights vary simultaneously across client steps. In this work, we propose a new optimization-based attack which successfully attacks FedAvg by addressing the above challenges. First, we solve the optimization problem using automatic differentiation that forces a simulation of the client's update that generates the unobserved parameters for the recovered labels and inputs to match the received client update. Second, we address the large number of batches by relating images from different epochs with a permutation invariant prior. Third, we recover the labels by estimating the parameters of existing FedSGD attacks at every FedAvg step. On the popular FEMNIST dataset, we demonstrate that on average we successfully recover >45% of the client's images from realistic FedAvg updates computed on 10 local epochs of 10 batches each with 5 images, compared to only <10% using the baseline. Our findings show many real-world federated learning implementations based on FedAvg are vulnerable.
en_US
dc.format
application/pdf
en_US
dc.language.iso
en
en_US
dc.publisher
OpenReview
en_US
dc.rights.uri
http://creativecommons.org/licenses/by/4.0/
dc.subject
Federated learning
en_US
dc.subject
Gradient leakage
en_US
dc.title
Data Leakage in Federated Averaging
en_US
dc.type
Journal Article
dc.rights.license
Creative Commons Attribution 4.0 International
ethz.journal.title
Transactions on Machine Learning Research
ethz.size
24 p.
en_US
ethz.version.deposit
publishedVersion
en_US
ethz.publication.place
s.l.
en_US
ethz.publication.status
published
en_US
ethz.leitzahl
ETH Zürich::00002 - ETH Zürich::00012 - Lehre und Forschung::00007 - Departemente::02150 - Dep. Informatik / Dep. of Computer Science::02664 - Inst. f. Programmiersprachen u. -systeme / Inst. Programming Languages and Systems::03948 - Vechev, Martin / Vechev, Martin
en_US
ethz.leitzahl
ETH Zürich::00002 - ETH Zürich::00012 - Lehre und Forschung::00007 - Departemente::02150 - Dep. Informatik / Dep. of Computer Science::02219 - ETH AI Center / ETH AI Center
ethz.leitzahl.certified
ETH Zürich::00002 - ETH Zürich::00012 - Lehre und Forschung::00007 - Departemente::02150 - Dep. Informatik / Dep. of Computer Science::02664 - Inst. f. Programmiersprachen u. -systeme / Inst. Programming Languages and Systems::03948 - Vechev, Martin / Vechev, Martin
en_US
ethz.identifier.url
https://openreview.net/forum?id=e7A0B99zJf
ethz.date.deposited
2023-01-19T10:31:20Z
ethz.source
FORM
ethz.eth
yes
en_US
ethz.availability
Open access
en_US
ethz.rosetta.installDate
2023-01-19T13:40:57Z
ethz.rosetta.lastUpdated
2024-02-02T19:29:35Z
ethz.rosetta.versionExported
true
ethz.COinS
ctx_ver=Z39.88-2004&amp;rft_val_fmt=info:ofi/fmt:kev:mtx:journal&amp;rft.atitle=Data%20Leakage%20in%20Federated%20Averaging&amp;rft.jtitle=Transactions%20on%20Machine%20Learning%20Research&amp;rft.date=2022-11&amp;rft.issn=2835-8856&amp;rft.au=Dimitrov,%20Dimitar%20Iliev&amp;Balunovi%C4%87,%20Mislav&amp;Konstantinov,%20Nikola&amp;Vechev,%20Martin&amp;rft.genre=article&amp;
 Search print copy at ETH Library

Files in this item

Thumbnail

Publication type

Show simple item record