• 대한전기학회
Mobile QR Code QR CODE : The Transactions of the Korean Institute of Electrical Engineers
  • COPE
  • kcse
  • 한국과학기술단체총연합회
  • 한국학술지인용색인
  • Scopus
  • crossref
  • orcid

  1. (Dept. of IT Applied Engineering, Jeonbuk National University, Jeonju, Rebublic of Korea.)
  2. (Dept. of Energy engineering, Jeonbuk National University, Jeonju, Rebublic of Korea.)



VGG16, Hybrid Model, LSTM, Dense Model, short-circuit traces

1. Introduction

Fire disasters pose a serious threat across various sectors, from residential environments to industrial facilities [1]. Among them, electrical fires resulting from short circuits are particularly hazardous, endangering both infrastructure and human life.

Investigating the cause of electric fires and determining their presence at fire scenes is essential for both prevention and post-incident analysis [4]. Electrical fires are typically categorized based on ignition sources, such as electrical overload, compression damage, tracking, insulation breakdown, and short circuits. Each cause leaves specific physical evidence, and examining these traces allows experts to confirm whether an electric fire occurred [5][6].

Timely and accurate detection of these events is crucial to minimizing damage and improving emergency response effectiveness. However, conventional inspection methods depend heavily on manual visual analysis, which can be inconsistent, time-consuming, and prone to human error. With the increasing availability of labeled image datasets and rapid advancements in deep learning, automated classification systems have emerged as a promising alternative to traditional methods.

In particular, melting marks on wires or metal components caused by the heat of electric fires are key indicators, but these are often described using various non-standardized terms such as "melting marks" or "short-circuit marks" [7][8]. To bring clarity, this paper defines primary short-circuit marks as traces directly linked to fire ignition and secondary short-circuit marks as those resulting indirectly from the fire. This classification helps make the investigation process more intuitive [9].

This paper proposes a deep learning-based approach utilizing a hybrid VGG16 framework that combines CNN capabilities with either Dense layers or LSTM units. The VGG16 + Dense model emphasizes simplicity and faster training, while the VGG16 + LSTM model is designed to capture temporal and spatial features more effectively.

Using transfer learning, the pre-trained VGG16 model was fine-tuned by unfreezing selected layers. Training data for both models consisted of labeled primary and secondary short-circuit marks, generated under controlled laboratory conditions.

This paper presents a comparative evaluation of two convolutional neural network (CNN)-based architectures: VGG16 combined with Dense layers and VGG16 combined with Long Short-Term Memory (LSTM), for the classification of electric fire short-circuit trace images [2][3]. By analyzing and comparing the performance of these hybrid architectures, the paper aims to identify an optimal solution that balances classification accuracy with computational efficiency, addressing the challenges posed by traditional approaches.

2. Proposed Work

2.1 Research background on model

The Visual Geometry Group at Oxford University developed the VGG architecture, which emphasizes deeper networks with smaller (3×3) convolution filters for enhanced image classification performance. VGG16, in particular, is known for its efficient architecture, comprising 13 convolutional layers and 3 fully connected layers, along with pooling and SoftMax layers [10].

In this paper, we apply the VGG16 backbone in two hybrid deep learning frameworks for classifying electric fire short-circuit trace images: VGG16 combined with Dense layers (VGG16-Dense) and VGG16 combined with Long Short-Term Memory layers (VGG16-LSTM). The VGG16-Dense model adds fully connected layers after the convolutional feature extractor to learn complex spatial patterns, offering a straightforward structure and faster training.

In contrast, the VGG16-LSTM model incorporates an LSTM layer to capture sequential dependencies within the spatial features extracted from the CNN, which is beneficial for distinguishing fine grained differences in trace patterns. This architecture is particularly effective in identifying primary and secondary short-circuit marks, crucial for determining the root cause of electric fires.

Both models leverage transfer learning from pre-trained VGG16 weights on ImageNet. For VGG16-LSTM, the last four convolutional layers are unfrozen for fine-tuning, allowing better adaptation to the specific dataset, while VGG16-Dense maintains a frozen base for simplicity and efficiency.

Overall, the paper highlights the effectiveness of hybrid CNN-based architectures specifically those integrating Dense and LSTM layers with VGG16 in enhancing classification accuracy for electric fire short-circuit detection. This comparative analysis demonstrates the importance of architectural choice in safety-critical applications such as fire investigation.

2.2. Dataset preparation

In this paper, a well-structured dataset was constructed to classify electric fire short-circuit traces, consisting of two distinct categories: primary short-circuit traces and secondary short-circuit traces. Each category included 867 labeled images, summing up to a total of 1,734 samples. These images were collected from controlled laboratory conditions and field investigations to ensure diversity and authenticity in appearance, shape, and texture of electric fire evidence.

The experimental data were generated in the laboratory using specialized equipment. Wires were fixed to the anode and moved with a transfer device to create contact and induce short circuits for sample fabrication. Two copper wire types (1.5SQ and 4SQ) were tested at 25℃ and 900℃, the latter chosen since fire temperatures typically reach 800–900℃. In the 900℃ tests, wires were either heated or exposed to flames for two minutes before short-circuiting. The primary short-circuit traces were produced by continuously applying current until a short circuit occurred, while the secondary short-circuit traces were formed by applying a flame to induce the short circuit. The thermal trace was obtained by applying a flame without current until a short circuit appeared.

To prepare the dataset for model training and evaluation, a train-test split was performed to ensure balanced representation across classes. Specifically, 348 images comprising 174 from each class were allocated to the test set to facilitate a fair and unbiased assessment of model performance. The remaining 1,386 images were designated for training, providing a substantial dataset for learning meaningful and discriminative features. To improve the model's generalization capability and mitigate overfitting, the training dataset underwent extensive data augmentation. This included a range of image transformations such as rotation, scaling, horizontal and vertical flipping, zooming, shifting, and brightness adjustment. These techniques significantly enriched the diversity of the training samples without the need for additional data collection, thereby enhancing the robustness and accuracy of the deep learning models.

The raw images varied widely in resolution and dimension, which posed a challenge for input compatibility with deep learning architectures. As VGG16 requires fixed size inputs, a preprocessing pipeline was implemented using Python. Images were first standardized to 500×375 pixels, and subsequently resized to 224×224 pixels, which matches the input requirement of the VGG16 backbone. This resizing ensured compatibility with pre-trained weights and preserved essential visual features. The resizing process was confirmed using IPython.display to verify visual integrity post-processing.

To further enrich the dataset and improve the model's generalization capacity, data augmentation was applied using the ImageDataGenerator class from TensorFlow Keras. This step artificially increases the size and diversity of the training set without additional data collection. A total of 693 images per class (1,386 images) from the training set were augmented across multiple transformations that simulate real world distortions and variabilities. The following augmentation techniques were applied using the ImageDataGenerator class from TensorFlow Keras.

표 1. 적용된 증강 기법

Table 1. The applied augmentation techniques

Technique Value/Range Purpose
Rescaling 1/255 Normalize pixel values to [0, 1]
Rotation ±30 degrees Allow to learn rotational invariance
Width Shift 20% of width Tolerate horizontal displacements
Height Shift 20% of height Tolerate vertical displacements
Zoom ±30% Handle variations in image size and scale
Horizontal Flip True Learn from mirrored images
Vertical Flip True Increase robustness to vertical variations
Brightness Adjustment Range [0.7, 1.3] Adapt to different lighting conditions

Data augmentation was applied throughout training to improve model generalization. For the VGG16 + Dense model, trained over 25 epochs, this resulted in about 34,650 augmented images. For the VGG16 + LSTM model, trained over 35 epochs, approximately 48,510 augmented images were generated. This strategy helped both models learn from more diverse data and improved their classification performance.

These preprocessing and augmentation steps formed a critical foundation for the deep learning framework, enabling the hybrid models to achieve high accuracy in detecting and distinguishing between primary and secondary electric fire short-circuit traces.

2.3. The two propsoed model architectures

In this paper, we utilized the VGG16 model as the backbone of two hybrid deep learning architectures VGG16 + Dense and VGG16 + LSTM to classify primary and secondary short-circuit trace images. Deep learning networks typically require extensive high-quality annotated image data for effective training. However, in fire investigation scenarios, short-circuit traces are often difficult to collect and label due to their random and irregular appearances in real cases.

To address the challenge of limited data, we employed transfer learning by leveraging the pre-trained weights of the VGG16 model trained on the ImageNet dataset. While ImageNet primarily contains natural scene images unrelated to electric fire traces, the large scale and diverse data distribution helps VGG16 extract generalized low-level features. These features were then fine-tuned on our specialized dataset, enhancing the learning efficiency and reducing overfitting.

The two hybrid architectures differ in their structure and objectives. The VGG16 + Dense model employs a frozen or partially trainable VGG16 convolutional base, followed by a Flatten layer and several Dense layers for classification. It focuses on extracting and processing spatial features in a straightforward feedforward manner.

In contrast, the VGG16 + LSTM model takes the output of the convolutional layers, reshapes the feature maps into sequences, and processes them through an LSTM layer. This allows the model to capture spatial dependencies across the feature sequences, improving its ability to recognize complex patterns in the trace images.

A notable distinction between the two models lies in the learning rate settings. The VGG16 + Dense model uses a learning rate of 1e-4, which allows for faster convergence due to its relatively simple architecture and fewer trainable parameters. Meanwhile, the VGG16 + LSTM model uses a lower learning rate of 1e-5. This is because the LSTM layer introduces more complexity and sensitivity to weight updates, requiring smaller learning steps to ensure stable and effective training. A higher learning rate in the LSTM-based model could result in unstable gradients or poor convergence, particularly when learning long range spatial dependencies.

The key architectural and performance differences between the two models are summarized below:

표 2. 제안된 두 모델의 구조 및 성능 비교

Table 2. The key architectural and performance differences between the two proposed models

Aspect VGG16-Dense VGG16-LSTM
Feature extractor Fully frozen Last 4 layers fine-tuned
After CNN Flatten Reshape + LSTM
Type of classifier Fully connected layers Sequential(memory-based) layer
Learning rate 1e-4 (faster) 1e-5 (slower)
Complexity Lower Higher
Strength Faster training, simpler Can capture spatial dependencies

3. Result and Experimental Analysis of the Two Proposed Models

All experiments and model training were conducted on a 64-bit Windows 11 machine equipped with an AMD Ryzen 9 5900X 12-Core Processor 3.70 GHz, 32 GB of RAM, and running on a x64-based architecture. The models were implemented using Python, leveraging the Keras library with TensorFlow backend. The training process was executed in an Anaconda environment to take advantage of hardware acceleration. Although no dedicated GPU was specified in the system details, the CPU provided sufficient performance for moderate-scale deep learning experiments.

The dataset was categorized into two classes: primary short-circuit marks and secondary molten marks. These were further divided into training and testing sets. To improve model generalization and robustness, data augmentation techniques including image rotation, shifting, zooming, flipping, and brightness adjustment were applied to the training images.

The paper implemented two transfer learning-based hybrid models: VGG16 + Dense and VGG16 + LSTM. These architectures utilized the VGG16 convolutional base pre-trained on ImageNet, followed by custom classification layers suited for the specific characteristics of electric short-circuit image data. The sigmoid activation function was used in the final layer for binary classification, and Adam optimizer was employed for efficient gradient-based optimization.

Model performance was evaluated using accuracy and loss metrics, with separate calculations for training and validation datasets using standard evaluation formulas. The performance metrics are as follows:

그림 1. 학습 및 검증 정확도 (VGG16+Dense)

Fig. 1. Training and validation accuracy (VGG16+Dense)

../../Resources/kiee/KIEE.2026.75.1.217/fig1.png

그림 2. 학습 및 검증 손실 (VGG16+Dense)

Fig. 2. Training and validation loss(VGG16+Dense)

../../Resources/kiee/KIEE.2026.75.1.217/fig2.png

그림 3. 혼동 행렬 (VGG16+Dense)

Fig. 3. Confusion matrix (VGG16+Dense)

../../Resources/kiee/KIEE.2026.75.1.217/fig3.png

To optimize training and avoid overfitting, the early stopping technique was applied to both models. Although the training process for each model was initially set to run for 50 epochs, the early stopping callback automatically halted training when the validation loss showed no improvement over a defined patience period. As a result, the VGG16 + Dense model stopped at epoch 25, and the VGG16 + LSTM model stopped at epoch 35, preserving the best weights and preventing unnecessary training. This approach not only improved model generalization but also reduced computational time and resource usage. While both models achieved strong results, the VGG16 + LSTM consistently outperformed the Dense variant across all metrics.

그림 4. 학습 및 검증 정확도 (VGG16+LSTM)

Fig. 4. Training and validation accuracy (VGG16+LSTM)

../../Resources/kiee/KIEE.2026.75.1.217/fig4.png

그림 5. 학습 및 검증 손실 (VGG16+LSTM)

Fig. 5. Training and validation loss (VGG16+LSTM)

../../Resources/kiee/KIEE.2026.75.1.217/fig5.png

그림 6. 혼동 행렬 (VGG16+LSTM)

Fig. 6. Confusion matrix (VGG16+LSTM)

../../Resources/kiee/KIEE.2026.75.1.217/fig6.png

표 3. 제안된 VGG16+LSTM 모델의 분류 성능

Table 3. Classification Performance of the Proposed VGG16+LSTM Model

Class Precision Recall F1-Score
Burn marks 0.98 0.99 0.99
Melt Marks 0.99 0.98 0.99

To evaluate the effectiveness of the two proposed models, VGG16 + Dense and VGG16 + LSTM, their performance on the validation set was measured using accuracy and loss metrics. These metrics provide insight into how well each model generalizes to unseen data. Validation accuracy indicates the proportion of correct predictions, while validation loss reflects the model's confidence and prediction error. The results, summarized in Table 3, demonstrate the differences in learning behavior and performance between the two architectures.

표 4. 제안된 모델들의 검증 성능 비교

Table 4. Validation performance comparison of the proposed models

Metric VGG16+Dense VGG16+LSTM
val_accuracy 95.11% 98.85%
val_loss 0.1532 0.0365

4. Conclusion

In this paper, the VGG16–Dense hybrid model and the VGG16–LSTM hybrid model are proposed for analyzing the causes of electrical fires, and the performance of these two models is comparatively evaluated.

In conclusion, both the VGG16 + Dense and VGG16 + LSTM models proved effective for classifying electric fire short-circuit images, leveraging the power of transfer learning and deep neural networks. The VGG16 + Dense model offers faster training and simpler architecture, making it suitable for deployment in resource-constrained environments. On the other hand, the VGG16 + LSTM model, though more complex, achieved superior performance by capturing spatial dependencies in the data more effectively.

Performance metrics validate these differences: the VGG16 + Dense model achieved a validation accuracy of 95.11% with a validation loss of 0.1532, whereas the VGG16 + LSTM model reached an impressive 98.85% accuracy with a significantly lower validation loss of 0.0365. These results highlight that incorporating LSTM layers can significantly enhance the model's ability to learn complex spatial patterns in short-circuit image traces.

Overall, the paper confirms that combining convolutional and sequential deep learning techniques can significantly improve the classification of electric short-circuit marks, offering practical solutions for intelligent fire detection and forensic analysis systems.

Acknowledgements

This work was supported by the Korea Institute of Energy Technology Evaluation and Planning (KETEP) grant funded by the Korea government (MCEE) RS-2022-KP002707, Jeonbuk Regional Energy Cluster Training of human resources.

References

1 
J. Jeong, S. Bae, H. Yoon, J.-H. Kim, 2025, A novel approach for UV signal-based fire detection and classification leveraging contrastive learning, Expert Systems with Applications, Vol. 237, pp. 127120DOI
2 
S. Kharroubi, S. Boubaker, 2023, Assessment of machine and deep learning approaches for fault diagnosis in photovoltaic systems using infrared thermography, Remote Sensing, Vol. 15, No. 6, pp. 1686DOI
3 
Y. Bekele, S. Rani, D. B. Kanagachidambaresan, M. Moghaddam, S. Licea, C. Munoz, A. Ortiz, B. S. Tadese, 2025, Hybrid CNN-LSTM detection of cyber-attacks on MPPT controllers in PV systems, pp. 1-6DOI
4 
S. H. Jung, S. M. Kim, S. H. Lee, 2020, Electrical Arc Detection using Convolutional Neural Network, J. Broadcast Eng., Vol. 25, No. 4, pp. 569-575DOI
5 
J. Park, J. H. Kang, E. P. Lee, Y. H. Ko, S. B. Bang, 2021, New Approach to Distinguish Copper Molten Marks Based on Quantitative Microstructure Analysis Using Electron Backscatter Diffraction, Fire Technol., Vol. 57, No. 4, pp. 1667-1682DOI
6 
L. Wang, H. W. Yao, Z. X. Wu, D. Liang, 2014, Research on evolution of electrical fire melted marks based on digital image processing technology, pp. 793-797DOI
7 
V. Babrauskas, 2004, Arc beads from fires: Can ‘cause’ beads be distinguished from ‘victim’ beads by physical or chemical testing, J. Fire Prot. Eng., Vol. 14, No. 2, pp. 125-147DOI
8 
A. Ayala, B. Fernandes, F. Cruz, D. Macêdo, A. L. I. Oliveira, C. Zanchettin, 2020, KutralNet: A Portable Deep Learning Model for Fire RecognitionDOI
9 
E. X. Li, N. Xu, N. Li, H. Ma, J. B. Shi, M. Zhao, 2018, Analysis of Primary Short-Circuited Melted Mark Evidence Identification in Electrical Fire Case, Appl. Mech. Mater., Vol. 876, pp. 161-165DOI
10 
K. Chatfield, K. Simonyan, A. Vedaldi, A. Zisserman, 2014, Return of the devil in the details: Delving deep into convolutional nets, pp. 1-11DOI

저자소개

하디 (Nazari Mohammad Hadi)
../../Resources/kiee/KIEE.2026.75.1.217/au1.png

He is currently enrolled in a Master degree program in the dept. of IT Applied System Engineering at Jeonbuk National University. His main research interests are deep learning and energy optimization., Artificial Intelligence, Machine learning modeling and computer software developments.

방준호 (Junho Bang)
../../Resources/kiee/KIEE.2026.75.1.217/au2.png

He received B.S., M.S. and Ph.D. degrees in Department of Electrical Engineering from Jeonbuk National University, in 1989, 1991 and 1996 respectively. He was a research engineer with LG Semiconductor from 1997 to 1998. He is currently working as a professor in Division of Convergence Technology Engineering and Department of Energy/Conversion Engineering of Graduate School, Jeonbuk National University, Jeonju, Rebublic of Korea. His main research interests include IT convergency systemdesign.

최철영 (Chul-Young Choi)
../../Resources/kiee/KIEE.2026.75.1.217/au3.png

He completed his Master’s degree in the Department of Energy Engineering from March 2020 to August 2022, and he began his Ph.D. program in September 2022. His main research interests include energy systems and offshore wind.

선로빈 (Robin Sun)
../../Resources/kiee/KIEE.2026.75.1.217/au4.png

He completed his Master’s degree in the Department of Energy Storage Conversion Engineering in February 2022, and he began his Ph.D. program in the Department of IT Applied System Engineering in August 2022.

김든찬 (Deunchan Kim)
../../Resources/kiee/KIEE.2026.75.1.217/au5.png

He is currently enrolled in a Ph.D. program at Jeonbuk National University with a master's degree in the dept. of IT Applied Systems Engineering. His main research interests are fault diagnosis and anomaly detection for power distribution facilities.