Abstract
Background: Complex structures such as a high proportion of power electronic equipment
has brought new challenges to the safe and stable operation of new-type power system, increasing
the possibility of the system being attacked, especially the more complex Advanced Persistent
Threat (APT). This kind of attack has a long duration and strong concealment.
Objective: Traditional detection methods target a relatively single attack mode, and the time span of
APT processed is relatively short. None of them can effectively capture the long-term correlation in the
attack, and the detection rate is low. These methods can’t meet the safety requirements of the new-type
power system. In order to solve this problem, this paper proposes an improved transformer model
called STSA-transformer algorithm, and applies it to the detection of APT in new-type power systems.
Methods: In the STSA-transformer model, the network traffic collected from the power system is
first converted into a sequence of feature vectors, and the location information and local feature of
the sequence, is extracted by combining position encoding with convolutional embedding operations,
and then global characteristics of attack sequences is captured using the multi-head selfattention
mechanism of the transformer encoder, the higher-frequency features of the attention are
extracted through the self-learning threshold operation, combined with the PowerNorm algorithm to
standardize the samples, and finally classify the network traffic of the APT.
Results: After multiple rounds of training on the model, the expected effect can be achieved and
applied to the APT detection of a new-type power system.
Conclusion: The experimental results show that the proposed STSA-transformer algorithm has better
detection accuracy and lower detection false-alarm rate than traditional deep learning algorithms
and machine learning algorithms.
Keywords:
Network security, new-type power system, advanced persistent threat, self-attention mechanism, STSAtransformer, false-alarm rate.
Graphical Abstract
[1]
"Xi Jinping presided over the ninth meeting of the Central Finance
and Economics Committee", People’s Daily, 2021, New York on
May 9 2022
[2]
State Grid Corporation of China Enterprise Standard Summary Table (2003.1.1-2020.12)., Beijing: State Grid Corporation of China, 2021.
[7]
S. Siddiqui, M.S. Khan, K. Ferens, and W. Kinsner, "Detecting advanced persistent threats using fractal dimension based machine learning classification", In In Sixth ACM Conference on Data and Application Security and Privacy, New York: USA, 2016, pp. 64-69
[9]
L. Haibo, "Advanced persistent threat detection based on generative adversarial networks and long short-term memory", Comput. Sci., vol. 47, no. 1, pp. 281-286, 2019.
[11]
L. He, L. Xin, N. Yin, and L. Chao, "APT attack detection method combining dynamic behavior and static features", Computer Engineering and Application,, pp. 1-13.
[12]
R. Liang, G. Yue, and X. Zhao, "APT attack detection method on traceability graph based on sequence feature extraction", Chinese Science: Information Science, vol. 52, no. 08, pp. 1463-1480, 2022.
[17]
A. Vaswani, N. Shazeer, and N. Parmar, "Attention is all you need", ArXiv, 2301.07583v1, pp. 2999-3007,, 2003.
[18]
W. Zaremba, "Recurrent Neural Network Regularization", ArXiv, 1409.2329v5, , 2014.
[20]
H. Zhou, S. Zhang, and J. Peng, "Informer: Beyond efficient transformer for long sequence time-series forecasting", Proceedings of AAAI, vol. 35, no. 12, pp. 11106-11115, 2021.
[21]
S. Shen, Z. Yao, and A. Gholami, "Powernorm: Rethinking batch normalization in transformers", Proceedings of the 37th International Conference on Machine Learning, pp. 8741-8751, 2020.
[23]
K. He, "Deep Residual Learning for Image Recognition", arXiv, 1512.03385,, 2015.
[25]
P. Zhou, W. Shi, and J. Tian, "Attention-based bidirectional long short-term memory networks for relation classification", Deployable
Machine Learning for Security Defense. Communications in
Computer and Information Science, vol. 1271. Springer, Cham, 2020.
[26]
J. Chung, C. Gulcehre, and K.H. Cho, "Empirical evaluation of gated recurrent neural networks on sequence modeling", ArXiv, 1412.3555,, 2014.