Show simple item record

dc.contributor.author
Barmpas, Konstantinos
dc.contributor.supervisor
Hofmann, Thomas
dc.contributor.supervisor
Roth, Kevin
dc.contributor.supervisor
Kilcher, Yannic
dc.contributor.supervisor
Haber, David
dc.date.accessioned
2022-02-09T11:13:01Z
dc.date.available
2022-02-09T10:33:41Z
dc.date.available
2022-02-09T11:13:01Z
dc.date.issued
2020
dc.identifier.uri
http://hdl.handle.net/20.500.11850/531617
dc.identifier.doi
10.3929/ethz-b-000531617
dc.description.abstract
Over the last decades, complex deep neural networks have revolutionized Artificial Intelligence (AI) research. These models can now achieve impressive performances on various complex tasks like recognition, detection and image semantic segmentation, achieving accuracy close to, or even better, than human perception. However, these neural networks require to be both deep and complex and this complexity constitutes a danger for the safety verification (certification) and interpretability of a neural network model. This project explores the certification properties of complex neural networks by taking them into ”shallow waters”. First, a detailed investigation of efficient model distillation techniques is conducted. Then, using the shallow models trained with these distillation methods, several of their properties are further explored, among them adversarial robustness and their performance under parameter reduction procedures. Finally, by combining network’s convex relaxation with model compression, the certification area of shallow student models (derived from either normally or robustly trained teacher networks) is researched. Through all of these experimental results, it is empirically demonstrated and proved that model distillation leads to shallow models with larger certification areas than their equivalent complex teacher networks. Therefore, based on this thesis evidence, shallow distillated networks constitute a possible solution to the safety and interpretability issues that modern complex Artificial Intelligence (AI) models face.
en_US
dc.format
application/pdf
en_US
dc.language.iso
en
en_US
dc.publisher
ETH Zurich
en_US
dc.rights.uri
http://rightsstatements.org/page/InC-NC/1.0/
dc.subject
Machine learning
en_US
dc.subject
Artificial intelligence
en_US
dc.subject
Model Distillation
en_US
dc.title
Certifying Properties of Deep Networks by Taking them into Shallow Waters
en_US
dc.type
Master Thesis
dc.rights.license
In Copyright - Non-Commercial Use Permitted
ethz.size
100 p.
en_US
ethz.publication.place
Zurich
en_US
ethz.publication.status
published
en_US
ethz.leitzahl
ETH Zürich::00002 - ETH Zürich::00012 - Lehre und Forschung::00007 - Departemente::02150 - Dep. Informatik / Dep. of Computer Science::02661 - Institut für Maschinelles Lernen / Institute for Machine Learning::09462 - Hofmann, Thomas / Hofmann, Thomas
en_US
ethz.date.deposited
2022-02-09T10:33:48Z
ethz.source
FORM
ethz.eth
yes
en_US
ethz.availability
Open access
en_US
ethz.rosetta.installDate
2022-02-09T11:13:14Z
ethz.rosetta.lastUpdated
2022-03-29T18:44:13Z
ethz.rosetta.versionExported
true
ethz.COinS
ctx_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.atitle=Certifying%20Properties%20of%20Deep%20Networks%20by%20Taking%20them%20into%20Shallow%20Waters&rft.date=2020&rft.au=Barmpas,%20Konstantinos&rft.genre=unknown&rft.btitle=Certifying%20Properties%20of%20Deep%20Networks%20by%20Taking%20them%20into%20Shallow%20Waters
 Search print copy at ETH Library

Files in this item

Thumbnail

Publication type

Show simple item record