Hostname: page-component-89b8bd64d-72crv Total loading time: 0 Render date: 2026-05-06T21:23:52.750Z Has data issue: false hasContentIssue false

A protection method of trained CNN model with a secret key from unauthorized access

Published online by Cambridge University Press:  09 July 2021

AprilPyone Maungmaung*
Affiliation:
Tokyo Metropolitan University, Tokyo, Japan
*
Corresponding author: H. Kiya Email: kiya@tmu.ac.jp

Abstract

In this paper, we propose a novel method for protecting convolutional neural network models with a secret key set so that unauthorized users without the correct key set cannot access trained models. The method enables us to protect not only from copyright infringement but also the functionality of a model from unauthorized access without any noticeable overhead. We introduce three block-wise transformations with a secret key set to generate learnable transformed images: pixel shuffling, negative/positive transformation, and format-preserving Feistel-based encryption. Protected models are trained by using transformed images. The results of experiments with the CIFAR and ImageNet datasets show that the performance of a protected model was close to that of non-protected models when the key set was correct, while the accuracy severely dropped when an incorrect key set was given. The protected model was also demonstrated to be robust against various attacks. Compared with the state-of-the-art model protection with passports, the proposed method does not have any additional layers in the network, and therefore, there is no overhead during training and inference processes.

Information

Type
Original Paper
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
Copyright © The Author(s), 2021. Published by Cambridge University Press in association with Asia Pacific Signal and Information Processing Association
Figure 0

Fig. 1. Overview of image classification with proposed model protection method.

Figure 1

Fig. 2. Process of block-wise transformation.

Figure 2

Fig. 3. Example of block-wise transformed images ($M = 4$) with key set $K$. (a) Original. (b) SHF. (c) NP. (d) FFX. (e) SHF + NP. (f) SHF + FFX. (g) SHF + NP + FFX.

Figure 3

Table 1. Accuracy (%) of protected models and baseline model for three datasets

Figure 4

Table 2. Accuracy (%) of protected models ($M = 4$) under use of estimated key set $K'$

Figure 5

Table 3. Accuracy (%) of protected models under fine-tuning attacks with incorrect key and small dataset

Figure 6

Table 4. Accuracy (%) of protected models under fine-tuning attacks with new dataset (CIFAR-100 to CIFAR-10) for fixed CNN and fine-tuned CNN

Figure 7

Table 5. Comparison of proposed protected model NP and state-of-the-art passport-protected model in terms of classification accuracy (%) for CIFAR datasets

Figure 8

Fig. 4. Horizontal, vertical, and diagonal correlation test results for plain image and images transformed by SHF, NP, and FFX ($M = 4$). (a) represents horizontal, vertical, and diagonal correlation distribution of plain image, (b) represents that of image transformed by SHF, (c) represents that of image transformed by NP, and (d) represents that of image transformed by FFX.

Figure 9

Table 6. Key sensitivity of various transformations with $M = 4$ and $8$