TY - GEN
T1 - SupConFL
T2 - 14th Asia-Pacific Symposium on Internetware, Internetware 2023
AU - Chen, Wei
AU - Chen, Wu
AU - Liu, Jiamou
AU - Zhao, Kaiqi
AU - Zhang, Mingyue
N1 - Publisher Copyright:
© 2023 ACM.
PY - 2023/8/4
Y1 - 2023/8/4
N2 - Recent years have seen a growing interest in deep learning-based approaches to localize faults in software. However, existing methods have not reached a satisfying level of accuracy. The main reason is that the feature extraction of faulty code elements is insufficient. Namely, these deep learning-based methods will learn some features that are not relevant to fault localization, and thus ignore the features related to fault localization. We propose SupConFL, a new framework for statement-level fault localization. Our framework combines the statement-level abstract syntax tree with the statement sequence, and adopt controllable attention-based LSTM to locate the faulty elements. The training is done through contrastive learning between the faulty code and its fixed version. By comparing the faulty code with the fixed code, the model can learn richer features of the faulty code elements. Our experiments on Defects4j-1.2.0 dataset show that our method outperforms the current state-of-the-art. Specifically, SupConFL improves Top-1 score by 7.96% in comparison with the current state-of-the-art. In addition, our method has also achieved good results in cross-project experiments.
AB - Recent years have seen a growing interest in deep learning-based approaches to localize faults in software. However, existing methods have not reached a satisfying level of accuracy. The main reason is that the feature extraction of faulty code elements is insufficient. Namely, these deep learning-based methods will learn some features that are not relevant to fault localization, and thus ignore the features related to fault localization. We propose SupConFL, a new framework for statement-level fault localization. Our framework combines the statement-level abstract syntax tree with the statement sequence, and adopt controllable attention-based LSTM to locate the faulty elements. The training is done through contrastive learning between the faulty code and its fixed version. By comparing the faulty code with the fixed code, the model can learn richer features of the faulty code elements. Our experiments on Defects4j-1.2.0 dataset show that our method outperforms the current state-of-the-art. Specifically, SupConFL improves Top-1 score by 7.96% in comparison with the current state-of-the-art. In addition, our method has also achieved good results in cross-project experiments.
KW - code representation
KW - contrastive learning
KW - fault localization
UR - https://www.scopus.com/pages/publications/85175728528
U2 - 10.1145/3609437.3609441
DO - 10.1145/3609437.3609441
M3 - 会议稿件
AN - SCOPUS:85175728528
T3 - ACM International Conference Proceeding Series
SP - 44
EP - 54
BT - 14th Asia-Pacific Symposium on Internetware, Internetware 2023 - Proceedings
PB - Association for Computing Machinery
Y2 - 4 August 2023 through 6 August 2023
ER -