Paper Abstract and Keywords |
Presentation |
2022-12-15 14:25
A DNN compression method based on output error of activation functions Koji Kamma, Toshikazu Wada (Wakayama Univ.) PRMU2022-38 |
Abstract |
(in Japanese) |
(See Japanese page) |
(in English) |
Deep Neural Networks (DNNs) are dominant in the field of machine learning. However, because DNN models have large computational complexity, implementation of DNN models on resource-limited equipment is challenging. Therefore, techniques for compressing DNN models without degrading their accuracy is desired. Pruning is one such technique that re- moves redundant neurons (or channels). In this paper, we present Pruning with Output Error Minimization (POEM), a method that performs not only pruning but also reconstruction to compensate the error caused by pruning. The strength of POEM lies in its reconstruction to minimize the output error of the activation function, whereas the previous methods minimize the error before the activation function. The experiments with well-known DNN models (VGG-16, ResNet-18, MobileNet) and image recognition datasets (ImageNet, CUB-200-2011) were conducted. The results show that POEM significantly outperformed the previous methods in maintaining the accuracy of the compressed models. |
Keyword |
(in Japanese) |
(See Japanese page) |
(in English) |
pruning / reconstruction / activation function / / / / / |
Reference Info. |
IEICE Tech. Rep., vol. 122, no. 314, PRMU2022-38, pp. 34-39, Dec. 2022. |
Paper # |
PRMU2022-38 |
Date of Issue |
2022-12-08 (PRMU) |
ISSN |
Online edition: ISSN 2432-6380 |
Copyright and reproduction |
All rights are reserved and no part of this publication may be reproduced or transmitted in any form or by any means, electronic or mechanical, including photocopy, recording, or any information storage and retrieval system, without permission in writing from the publisher. Notwithstanding, instructors are permitted to photocopy isolated articles for noncommercial classroom use without fee. (License No.: 10GA0019/12GB0052/13GB0056/17GB0034/18GB0034) |
Download PDF |
PRMU2022-38 |
Conference Information |
Committee |
PRMU |
Conference Date |
2022-12-15 - 2022-12-16 |
Place (in Japanese) |
(See Japanese page) |
Place (in English) |
Toyama International Conference Center |
Topics (in Japanese) |
(See Japanese page) |
Topics (in English) |
|
Paper Information |
Registration To |
PRMU |
Conference Code |
2022-12-PRMU |
Language |
Japanese |
Title (in Japanese) |
(See Japanese page) |
Sub Title (in Japanese) |
(See Japanese page) |
Title (in English) |
A DNN compression method based on output error of activation functions |
Sub Title (in English) |
|
Keyword(1) |
pruning |
Keyword(2) |
reconstruction |
Keyword(3) |
activation function |
Keyword(4) |
|
Keyword(5) |
|
Keyword(6) |
|
Keyword(7) |
|
Keyword(8) |
|
1st Author's Name |
Koji Kamma |
1st Author's Affiliation |
Wakayama University (Wakayama Univ.) |
2nd Author's Name |
Toshikazu Wada |
2nd Author's Affiliation |
Wakayama University (Wakayama Univ.) |
3rd Author's Name |
|
3rd Author's Affiliation |
() |
4th Author's Name |
|
4th Author's Affiliation |
() |
5th Author's Name |
|
5th Author's Affiliation |
() |
6th Author's Name |
|
6th Author's Affiliation |
() |
7th Author's Name |
|
7th Author's Affiliation |
() |
8th Author's Name |
|
8th Author's Affiliation |
() |
9th Author's Name |
|
9th Author's Affiliation |
() |
10th Author's Name |
|
10th Author's Affiliation |
() |
11th Author's Name |
|
11th Author's Affiliation |
() |
12th Author's Name |
|
12th Author's Affiliation |
() |
13th Author's Name |
|
13th Author's Affiliation |
() |
14th Author's Name |
|
14th Author's Affiliation |
() |
15th Author's Name |
|
15th Author's Affiliation |
() |
16th Author's Name |
|
16th Author's Affiliation |
() |
17th Author's Name |
|
17th Author's Affiliation |
() |
18th Author's Name |
|
18th Author's Affiliation |
() |
19th Author's Name |
|
19th Author's Affiliation |
() |
20th Author's Name |
|
20th Author's Affiliation |
() |
Speaker |
Author-1 |
Date Time |
2022-12-15 14:25:00 |
Presentation Time |
15 minutes |
Registration for |
PRMU |
Paper # |
PRMU2022-38 |
Volume (vol) |
vol.122 |
Number (no) |
no.314 |
Page |
pp.34-39 |
#Pages |
6 |
Date of Issue |
2022-12-08 (PRMU) |
|