Abstract

Since Kocher (CRYPTO’96) proposed timing attack, side channel analysis (SCA) has shown great potential to break cryptosystems via physical leakage. Recently, deep learning techniques are widely used in SCA and show equivalent and even better performance compared to traditional methods. However, it remains unknown why and when deep learning techniques are effective and efficient for SCA. Masure et al. (IACR TCHES 2020(1):348–375) illustrated that deep learning paradigm is suitable for evaluating implementations against SCA from a worst-case scenario point of view, yet their work is limited to balanced data and a specific loss function. Besides, deep learning metrics are not consistent with side channel metrics. In most cases, they are deceptive in foreseeing the feasibility and complexity of mounting a successful attack, especially for imbalanced data. To mitigate the gap between deep learning metrics and side channel metrics, we propose a novel Cross Entropy Ratio (CER) metric to evaluate the performance of deep learning models for SCA. CER is closely related to traditional side channel metrics Guessing Entropy (GE) and Success Rate (SR) and fits to deep learning scenario. Besides, we show that it works stably while deep learning metrics such as accuracy becomes rather unreliable when the training data tends to be imbalanced. However, estimating CER can be done as easy as natural metrics in deep learning algorithms with low computational complexity. Furthermore, we adapt CER metric to a new kind of loss function, namely CER loss function, designed specifically for deep learning in side channel scenario. In this way, we link directly the SCA objective to deep learning optimization. Our experiments on several datasets show that, for SCA with imbalanced data, CER loss function outperforms Cross Entropy loss function in various conditions.

Highlights

  • Side channel analysis has received a significant amount of attention since it was first proposed in [Koc96]

  • The structures of the Multi-Layer Perceptron (MLP) models used in our experiments are similar to the ones proposed in [PSB+18] and we vary the epoch and batch size to explore whether Cross Entropy Ratio (CER) reflects Guessing Entropy (GE)/Success Rate (SR) under different conditions

  • Using MLP models with CER loss function we can attain about 100% SR with 1000 traces when epoch = 100 and batch size = 200, or epoch = 200 and batch size = 100 or 200, while we can get no more than 50% SR with CE loss function

Read more

Summary

Introduction

Side channel analysis has received a significant amount of attention since it was first proposed in [Koc96]. It exploits weakness of physical implementations instead of the algorithms, bringing a serious threat to various electronic devices. Throughout the paper, we use calligraphic letters X to denote sets and upper-case letter X to denote random variables over X. We use lower case letters x to denote a sample or realization of X. We use E[X] to denote the expectation of the random variable X and the condition will be specified under the expectation symbol or as a subscript, e.g. E and EX.

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call