University of Leicester
Browse
LaTeX_DDAD_For_IJMLC(1).pdf (10.41 MB)

Dual discriminator adversarial distillation for data-free model compression

Download (10.41 MB)
journal contribution
posted on 2024-04-12, 08:18 authored by Haoran Zhao, Xin Sun, Junyu Dong, Hui Yu, Huiyu Zhou

Knowledge distillation has been widely used to produce portable and efficient neural networks which can be well applied on edge devices for computer vision tasks. However, almost all top-performing knowledge distillation methods need to access the original training data, which usually has a huge size and is often unavailable. To tackle this problem, we propose a novel data-free approach in this paper, named Dual Discriminator Adversarial Distillation (DDAD) to distill a neural network without any training data or meta-data. To be specific, we use a generator to create samples through dual discriminator adversarial distillation, which mimics the original training data. The generator not only uses the pre-trained teacher's intrinsic statistics in existing batch normalization layers but also obtains the maximum discrepancy from the student model. Then the generated samples are used to train the compact student network under the supervision of the teacher. The proposed method obtains an efficient student network which closely approximates its teacher network, despite using no original training data. Extensive experiments are conducted to to demonstrate the effectiveness of the proposed approach on CIFAR-10, CIFAR-100 and Caltech101 datasets for classification tasks. Moreover, we extend our method to semantic segmentation tasks on several public datasets such as CamVid and NYUv2. All experiments show that our method outperforms all baselines for data-free knowledge distillation.

Funding

We thank supports of the National Natural Science Foundation of China (No. 61971388, U1706218, L1824025), and Key Natural Science Foundation of Shandong Province (No. ZR2018ZB0852).

History

Author affiliation

School of Computing and Mathematical Sciences, University of Leicester

Version

  • AM (Accepted Manuscript)

Published in

International Journal of Machine Learning and Cybernetics

Volume

13

Pagination

1213–1230

Publisher

Springer Verlag

Copyright date

2021

Available date

2024-04-12

Language

en

Usage metrics

    University of Leicester Publications

    Categories

    No categories selected

    Keywords

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC