Open Access
Issue
Wuhan Univ. J. Nat. Sci.
Volume 31, Number 1, February 2026
Page(s) 35 - 44
DOI https://doi.org/10.1051/wujns/2026311035
Published online 06 March 2026

© Wuhan University 2026

Licence Creative CommonsThis is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

0 Introduction

In recent years, Distributed Acoustic Sensing (DAS) technology has rapidly emerged in industrial and scientific fields due to its wide coverage, high sensitivity, and excellent resistance to electromagnetic interference [1]. As a cutting-edge technology based on fiber-optic communication, DAS uses optical fibers as sensing media to achieve real-time monitoring over large areas through distributed sensing. It has demonstrated unique advantages in applications such as pipeline leak detection[2], seismic activity monitoring[3], bridge and tunnel health monitoring[4], and border security protection[5].

However, the signals acquired by DAS systems are often accompanied by significant noise, complex spatiotemporal characteristics, and imbalanced multi-class distributions. These issues pose severe challenges to the robustness and generalization capability of signal classification algorithms. Traditional signal classification methods can be broadly categorized into two types: the first relies on feature engineering-based machine learning approaches (such as Support Vector Machines, Random Forests, and k-Nearest Neighbors), which often depend on manually designed feature extraction processes. The performance of these methods heavily relies on the quality of the extracted features and struggles to fully exploit deep signal information[6]. The second category utilizes end-to-end deep learning for feature extraction. Models such as Convolutional Neural Networks (CNNs) and Long Short-Term Memory (LSTM) networks fall into this category and perform exceptionally well in image and sequential signal classification by automatically learning data features[7]. In addition, several recent studies have introduced Transformer-based or hybrid lightweight architectures into DAS signal analysis. For example, Li et al[8] proposed a CNN-LSTM-Transformer hybrid network for vibration event recognition in distributed fiber sensing systems, demonstrating enhanced modeling of long-range temporal dependencies. Similarly, Duan et al[9] developed a self-supervised Transformer-based Distributed Acoustic Sensing-Masked Autoencoder framework that achieves high-performance representation learning under limited labeled data. These advances further highlight the growing role of Transformer-style architectures in DAS classification tasks.

Despite the apparent advantages of traditional deep learning methods in automatic feature extraction and high-precision classification, they generally assume that the input data have a fixed and regular structure and require large amounts of labeled data to achieve ideal performance. DAS signals often contain latent non-Euclidean topological relationships, which traditional models struggle to comprehensively capture. Moreover, the high cost of collecting and annotating large-scale datasets further limits the practical application of these conventional methods.

In contrast, in the field of image classification, to address the aforementioned issues, the Inception-ResNet network architecture has been proposed[10]. The Inception module captures multi-scale signal features, enabling flexible handling of diverse patterns. Complementarily, the residual connections in ResNet mitigate the vanishing gradient problem and significantly enhance deep feature extraction. The organic integration of the two not only makes Inception-ResNet more expressive and precise in feature representation but also considerably reduces dependency on large training datasets, maintaining high classification accuracy even with limited data resources.

Extensive experimental results have shown that models based on Inception-ResNet perform excellently in tasks such as object detection and image classification[11]. Even with significantly reduced training data, they achieve stable and efficient classification outcomes. Several improved versions now exist, including Inception-ResNet-v1 and Inception-ResNet-v2. Notably, empirical studies have confirmed that Inception-ResNet-v2 demonstrates a superior ability to capture complex feature representations[12]. Boasting a deeper, wider architecture and more parallel branches than v1, Inception-ResNet-v2 provides a stronger capacity for learning multi-scale features—a crucial characteristic for modeling the noisy and complex patterns in DAS signals. This enhanced representational capacity, despite its higher computational cost, is particularly well-suited for fiber optic signal classification under data-limited conditions.

Against this backdrop, this paper proposes applying a modified Inception-ResNet model to DAS signal classification in distributed fiber-optic sensing. To better suit the characteristics of optical fiber signal data, the architecture is adapted by drastically reducing the network depth—retaining only the Stem, one Inception-ResNet-A, and one Reduction-A module while removing all Inception-B and Inception-C stages—to form a lightweight multi-branch feature extractor and validated through experiments on public datasets. To enhance interpretability, the proposed model is further equipped with an explainable mechanism that highlights critical signal patterns contributing to classification decisions. Consequently, the proposed model achieves significant advantages in multi-class classification tasks, with an average classification accuracy of 92.6%, and outperforms traditional methods in terms of training data requirements.

The main contributions of this paper are as follows:

(1) A lightweight Inception-ResNet architecture is designed for DAS intrusion recognition, achieving high classification accuracy even under limited training data. By reducing the network depth while retaining multi-branch feature extraction, the proposed model significantly alleviates dependence on large-scale labeled datasets and improves learning efficiency in small-sample scenarios.

(2) The model effectively handles high-noise, complex-pattern DAS signals through multi-scale feature extraction and residual-enhanced deep representation learning. This design captures latent multi-level vibration structures and high-dimensional discriminative patterns, thereby improving robustness and generalization performance under challenging sensing conditions.

(3) The improved Inception-ResNet framework is tailored to the non-Euclidean temporal-spatial characteristics of distributed fiber sensing signals and provides an interpretable decision process. By integrating a Gradient-weighted Class Activation Mapping-based visualization mechanism, the model highlights class-specific high-contribution regions in vibration signals, offering transparent and trustworthy explanations of classification results and supporting practical deployment in real DOFS applications.

The structure of this paper is as follows: Section 1 elaborates on the architecture and implementation of the proposed model; Section 2 introduces the experimental setup and provides an analysis and in-depth discussion of the results; finally, Section 3 summarizes the research and presents prospects for future work.

1 Inception-ResNet-Based Vibration Intrusion Recognition Framework

1.1 Distributed Optical Fiber Vibration Sensing System

A distributed acoustic sensing (DAS) system based on Φ-OTDR technology is used to monitor ground vibrations around gas pipelines, as illustrated in Fig.1. The system employs a narrow-linewidth continuous laser, which is first pulse-modulated by an acousto-optic modulator (AOM), with the modulation controlled by a signal from the laser driver. The resulting laser pulses are then amplified using an erbium-doped fiber amplifier (EDFA), passed through a filter to remove noise, and injected into the sensing fiber via a fiber coupler.

Thumbnail: Fig.1 Refer to the following caption and surrounding text. Fig.1 Structure of distributed optical fiber vibration sensing system (DVS)

When the sensing fiber experiences external vibrations, part of the backward Rayleigh-scattered light is reflected back along the fiber to the system. This scattered light is amplified again by the EDFA and ultimately detected by a photodetector, where it is converted into an electrical signal for event detection and analysis.

The intensity of the Rayleigh backscattered light from the disturbance source at distance LiMathematical equation along the sensing fiber can be expressed as the vector sum of the amplitudes and phases of the scattering centers within that region:

E b ( L i ) = E 0 e - α ( i - 1 ) Δ L k = 1 M α k i e j φ k i ,                     ( 1 ) Mathematical equation

where the backscattered field Eb(Li)Mathematical equation in Eq. (1) originates from the coherent superposition of contributions from MMathematical equation scattering centers within the i-th segment of optical fiber (with length LMathematical equation). The incident optical field E0Mathematical equation, before reaching the i-th segment, undergoes attenuation described by e-α(i-1)ΔLMathematical equation , where αMathematical equation is the attenuation coefficient of the optical fiber, and (i-1)ΔLMathematical equation represents the propagation distance of the light. Each k-th scattering center is characterized by its scattering amplitude αkiMathematical equation and phase φkiMathematical equation , where the phase is highly sensitive to changes in the round-trip optical path and can thus respond to external disturbances such as strain and temperature.

1.2 Inception-ResNet-Based Model Framework

1.2.1 Inception module

The original Inception structure v1, also known as GoogLeNet, was proposed by Szegedy et al in 2014[13]. As shown in Fig.2, the Inception module—the core of GoogLeNet—uses multiple layers with different filter sizes (1×1, 3×3, and 5×5) and pooling operations in parallel. This architecture allows the network to extract features at multiple scales, capturing complex representations. To reduce computational cost and overfitting while preserving representational power, 1×1 convolutions are applied before the 3×3 and 5×5 convolutions, preventing an explosion in the number of parameters.

Thumbnail: Fig.2 Refer to the following caption and surrounding text. Fig.2 Inception module with dimension reductions

1.2.2 Residual connections

ResNet (Residual Network), proposed by He et al in 2015[14], introduces residual connections in convolutional operations, as shown in Fig.3. Unlike the Inception structure, residual connections directly add the input and the output of the convolution layer under the same channel dimension. This identity mapping enables direct information flow, improving training efficiency.

Thumbnail: Fig.3 Refer to the following caption and surrounding text. Fig.3 Residual connections

1.2.3 Inception-ResNet framework for distributed optical fiber sensing

For a closed-set multi-class classification problem, consider a 1D vibration signal dataset D={(x1, y1),, (xn, yn)}Mathematical equation, containing samples from K known classes. Each sample xiRDMathematical equation represents a vibration signal, with size depending on the sampling rate and duration of the analog-to-digital conversion. The label yi {1, 2,, K}Mathematical equation denotes the class.

As shown in Fig. 4, the proposed framework processes input training samples xi with a shape of 10 000×12×1. It employs an improved Inception-ResNet-V2 architecture, the overall structure of which can be clearly divided into two parts: feature extraction and classification. The feature extraction part is composed of the Stem module, the Inception-ResNet modules, and the Reduction modules, which are detailed below.

Thumbnail: Fig.4 Refer to the following caption and surrounding text. Fig.4 Inception-ResNet main framework (Stem, Inception-ResNet, Reduction, GAP, Softmax)

1) Stem module: Serving as the entry point of the network, it is responsible for rapid preliminary feature extraction and dimensionality reduction of the input raw vibration signals (with a shape of 10 000×12×1).

The structure of the Stem module and input shape changes are illustrated in Fig. 5. It performs six steps: three convolutions, max pooling, two more convolutions, another max pooling, followed by multiscale feature extraction and fusion. All convolutions include batch normalization and activation layers to accelerate convergence and reduce gradient vanishing. The convolution kernel sizes in the length dimension are kept constant, with multiscale filters of sizes 1, 3, and 5. After two rounds of downsampling, the width remains unchanged.

Thumbnail: Fig.5 Refer to the following caption and surrounding text. Fig.5 The stem structure includes the input layer, convolutional layers (Conv-1-11), Max pool-1 and Max pool-2, Avg pool-1, and filter concatenation layers

The input dimension is 10 000×12×1Mathematical equation. Layer parameters and output dimensions are annotated in the figure.

2) Inception-ResNet module: This is the core of the network. By incorporating multi-scale convolutions (inspired by the Inception concept shown in Fig.2) and residual connections (whose basic unit is shown in Fig.3), it achieves deep fusion and efficient propagation of features.

The Inception-ResNet structure is shown in Fig. 6, containing four branches. The fourth branch uses residual connections instead of simple filter concatenation. The outputs are merged and passed through a 1×1 convolution and multiplied by a scale factor before being added to the input. This scaled residual connection helps improve gradient flow and training stability.

Thumbnail: Fig.6 Refer to the following caption and surrounding text. Fig.6 The Inception-ResNet structure comprises ReLU activation, convolutional layers (Conv-12–18), and residual addition

Note:   Following ReLU activation, the input is split into three branches, processed through Conv-12-17, concatenated via Conv-18, and combined with the original input using a residual connection. Layer parameters and output dimensions are annotated in the figure.

3) Reduction module (its structure is detailed in Fig.7): At specific stages of the network, this module is responsible for compressing the spatial dimensions of the feature maps while increasing the number of channels, thereby extracting more abstract features.

Thumbnail: Fig.7 Refer to the following caption and surrounding text. Fig.7 The reduction structure consists of Max pool-3, convolutional layers (Conv-19–22), a filter concatenation layer, and ReLU activation

Layer parameters (kernel size, stride, and padding) and output dimensions are annotated in the figure.

As illustrated in Fig. 7, the reduction module has three branches and plays a crucial role in downsampling the feature maps' spatial dimensions, reducing memory and computational cost, and increasing channel numbers. By compressing spatial information, it enables the network to learn more abstract and richer features in deeper layers.

Following the aforementioned feature extraction process, the classification part consists of a global average pooling (GAP) layer, a Dropout layer, and a Softmax layer. The final output shape of the network is 1×1×6, representing the predicted probability for each class.

The original Inception-ResNet-V2 was designed for the 1 000-class classification task of ImageNet and is extremely deep. However, the dataset used in this study contains only 6 classes. Therefore, we significantly reduced the network depth to prevent overfitting and improve efficiency. The proposed network is derived from the Inception-ResNet-V2 architecture, but a depth-reduced design was adopted to better match the characteristics of fiber-optic intrusion signals. While the original Inception-ResNet-V2 contains five Inception-A modules, ten Inception-B modules, five Inception-C modules, and two reduction blocks, our model retains only the Stem block, a single Inception-ResNet-A module, and the Reduction-A block. All Inception-B and Inception-C stages were removed, which significantly decreases the number of layers and parameters while preserving the multi-branch feature aggregation mechanism that is essential for representing distributed vibration patterns.

Our proposed model learns a mapping function f(x, θ)Mathematical equation through forward propagation. The Stem layer is defined as:

X 0 = S t e m ( x ) ,                                  ( 2 ) Mathematical equation

where xMathematical equation is the input tensor representing the fiber vibration signal, producing the initial feature map X0Mathematical equation. The Inception-ResNet layer is defined as:

F 1 ( X ) = C o n v 1 × 1 ( C o n c a t ( Y 1 , Y 2 , Y 3 , Y 4 ) ) ,       ( 3 ) Mathematical equation

X 1 = R e L U ( X 0 + s c a l e × F 1 ( X 0 ) ) ,               ( 4 ) Mathematical equation

where YiMathematical equation are the outputs from different branches that capture local, medium, and global features. The Reduction layer is then defined as:

X 2 = R e d u c t i o n ( X 1 ) ,                          ( 5 ) Mathematical equation

which downsamples the feature map to obtain X2Mathematical equation with more channels. After classification, the final output is:

Y = f ( x , θ ) = s o f t m a x ( W T × G A P ( X 2 ) + b ) ,         ( 6 ) Mathematical equation

where θ represents the model parameters, which specifically include the weight matrix W and the bias vector b of the final fully-connected layer, x is the sampled vibration signal, and GAP denotes global average pooling.

1.2.4 Categorical cross-entropy loss

Since this task is a multi-class classification problem, we adopt the classification cross-entropy loss as the loss function of the model. This loss function can effectively measure the difference between the predicted probability distribution output by the model and the true label, thereby guiding the model to learn more discriminative feature representations during the training process. The formula for classification cross-entropy loss is defined as follows:

L o s s = - 1 N i = 1 N i = 1 K y i , k l o g   ( y ^ i , k ) ,                 ( 7 ) Mathematical equation

where NMathematical equation represents the number of samples in a batch, KMathematical equation represents the total number of categories (in this article,K=6Mathematical equation), yi,kMathematical equation is the true label (one-hot encoding) of the i-th sample on the k-th class; y^i,kMathematical equation is the predicted probability of the model that the i-th sample belongs to the k-th category.

1.3 Grad-CAM-Based Interpretability Analysis

To improve interpretability, a Grad-CAM-based mechanism is incorporated to visualize class-discriminative regions in the vibration signals. Let

AkR H×WMathematical equation denote the k-thMathematical equation feature map (with height HMathematical equation and width WMathematical equation) extracted from a chosen convolutional layer,

ycMathematical equation represent the pre-softmax class score for category cMathematical equation, and

Z=H×WMathematical equation be the spatial normalization factor.

(1) Compute gradients of class score w.r.t. feature maps

The importance of each channel kMathematical equation to class cMathematical equation is measured by the gradient of ycMathematical equation with respect to each spatial location in AkMathematical equation:

y c A i j k , i = 1 , , H ,   j = 1 , , W .                  ( 8 ) Mathematical equation

(2) Compute channel-wise weights via global average pooling of gradients

The channel importance weight akcMathematical equation is obtained by averaging the gradients over all spatial locations:

a k c = 1 Z i = 1 H j = 1 W y c A i j k .                            ( 9 ) Mathematical equation

This operation reflects the overall contribution of feature map AkMathematical equation to the activation of class cMathematical equation.

(3) Generate the Grad-CAM heatmap

The class-specific heatmap LGrad-CAMcMathematical equation is computed as the weighted linear combination of all feature maps, followed by a ReLU nonlinearity to retain only positive contributions:

L G r a d - C A M c ( i , j ) = R e L U ( k = 1 K a k c A i j k ) .                ( 10 ) Mathematical equation

Here, KMathematical equation is the total number of channels in the chosen convolutional layer.

(4) Upsampling and overlay

The heatmap is then normalized and upsampled from size H×WMathematical equation to match the original signal representation (e.g., 10 000×12), using bilinear interpolation:

    L ^ G r a d - C A M c = U p s a m p l e ( L G r a d - C A M c ,   s i z e = i n p u t   d i m e n s i o n )                        ( 11 ) Mathematical equation

Finally, the upsampled heatmap is superimposed onto the corresponding signal segment to highlight the regions with the greatest influence on the classification results.

This detailed Grad-CAM formulation enables transparent analysis of how the network distinguishes between different intrusion events by revealing class-specific high-contribution regions in the vibration signals.

2 Experimental Results and Analysis

2.1 Dataset Construction and Signal Processing

In this experiment, we utilized a publicly available dataset[15], which contains a large number of samples across six categories: background, dig, knock, water, shake, and walk. In the initial experiment, 500 samples were randomly selected from each category, resulting in a total of 3 000 samples.

As illustrated in Fig. 8, in order to better adapt to the Inception-ResNet network, preprocessing was applied to the fiber optic sensing vibration signals. To enhance input consistency and accelerate convergence, each raw vibration signal was first normalized using Min-Max scaling:

x n o r m = x - m i n   ( x ) m a x   ( x ) - m i n   ( x ) .                       ( 12 ) Mathematical equation

Thumbnail: Fig. 8 Refer to the following caption and surrounding text. Fig. 8 Grayscale representation

The resulting values were subsequently scaled to the grayscale range of [0,255]Mathematical equation and converted to 8-bit integer format. This preprocessing generates grayscale signal images that are compatible with the Inception-ResNet network. This 2D representation enables the Inception-ResNet network to exploit multi-scale spatial convolutions (1×1, 3×3, 5×5), allowing it to capture local transients, medium-scale oscillations, and long-range temporal patterns simultaneously. Compared with a 1D CNN, which relies solely on linear temporal kernels, the 2D encoding provides richer receptive-field diversity and improves the separability of vibration event patterns. This design choice was empirically validated in our experiments, where the 2D grayscale representation yielded superior classification accuracy under the same network depth.

2.2 Classification Performance Testing

The purpose of this experiment is to evaluate whether the proposed Inception-ResNet network model can achieve the expected results with a significantly reduced training dataset and no additional anomalous signal inputs. The results are compared with those from other deep learning models based on classical convolutional neural networks.

First, 300 samples from each category (60% of the total dataset) were randomly selected as the training set, with the remaining 40% used as the test set. During model training, we used a batch size of 16 and adopted the cosine annealing learning rate schedule[16], starting from 0.02 and decaying gradually to 0 over 200 epochs.

To more accurately reflect the model's performance, accuracy was used as an evaluation metrics, defined as follows:

A c c u r a c y = T P + T N T P + F P + F N + T N ,                ( 13 ) Mathematical equation

P r i c i s i o n = T P T P + F P ,                        ( 14 ) Mathematical equation

R e c a l l = T P T P + F N ,                           ( 15 ) Mathematical equation

F 1   S c o r e   = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n   +   R e c a l l ,               ( 16 ) Mathematical equation

where TP denotes true positives, FP false positives, TN true negatives, and FN false negatives.

As shown in Fig. 9 and Fig. 10, the evaluation metrics and loss decrease curve respectively demonstrate that after around 100 training epochs, the accuracy exceeded 90% and stabilized thereafter, reaching as high as 92.6%. A detailed analysis of Fig. 11 reveals that the proposed model achieves consistently high classification accuracy across all six categories. The learned feature representations exhibit high class separability, reflecting the model's strong discriminative capability. Nevertheless, certain confusion patterns can still be observed. A small proportion of background samples are misclassified as walk, which is likely due to their similar low-intensity vibration amplitudes and ambient noise levels under both conditions. More noticeable confusion occurs between water and shake, as both classes produce oscillatory waveforms with overlapping frequency ranges and comparable temporal continuity, leading to partial overlap in the learned feature space. In contrast, knock signals are characterized by short-duration, high-amplitude transients that create distinctive temporal sparsity, allowing the model to capture them robustly and achieve a 100% recall. The relatively lower recall of the walk class (0.84) indicates greater intra-class variability, which may arise from irregular human step intervals or signal attenuation along the sensing fiber.

Thumbnail: Fig.9 Refer to the following caption and surrounding text. Fig.9 The change trend of accuracy

Thumbnail: Fig.10 Refer to the following caption and surrounding text. Fig.10 The change trend of loss with epochs

Thumbnail: Fig.11 Refer to the following caption and surrounding text. Fig.11 Confusion matrix

To analyze the Grad-CAM heatmaps (as shown in Fig. 12), we leverage the principle that warm colors (red and yellow) indicate regions with high contribution to the model's classification. The results demonstrate that the model captures semantically meaningful features, with each class's attention regions corresponding to visually distinct patterns in the original signals. Moreover, the attention regions vary across classes, suggesting that the model relies on unique combinations of spatial layout, texture, and amplitude distribution to differentiate between classes. These distinct attention patterns demonstrate the model's interpretability and confirm that it bases its decisions on semantically relevant features, which underpins its reliable classification performance.

Thumbnail: Fig.12 Refer to the following caption and surrounding text. Fig.12 Visualization of model attention on fiber-optic vibration signals using Grad-CAM

A two-dimensional feature output layer was added at the end of the network to visualize the distribution of different event types in 2D space, as shown in Fig. 13. To comprehensively evaluate the performance of the proposed model, we compared it with several representative baseline methods under the same experimental setup. These baselines include a one-dimensional convolutional neural network (CNN), VGG-16, ResNet, and a traditional Support Vector Machine (SVM). All models were trained and tested using the identical dataset split (60% training, 40% testing) and the same preprocessing pipeline. The classification accuracies of all models are summarized in Table 1.

Thumbnail: Fig.13 Refer to the following caption and surrounding text. Fig.13 Test set data distribution

Table 1 summarizes the classification performance of all models under identical experimental conditions. It clearly highlights the superior performance of our proposed model compared to other feature extraction networks, particularly under significantly reduced training data. A quantitative comparison shows that our model achieves the highest accuracy of 92.6%, which represents improvements of 14.0, 6.8, 4.8, and 2.4 percentage points over the SVM, CNN, VGG-16, and ResNet baselines, respectively.

Table 1

Accuracy under different models and methods (unit:%)

3 Conclusion

This paper proposes a feature extraction network based on Inception-Net-V2 and designs a model framework suitable for DVS (Distributed Vibration Sensing) signal classification. Using a publicly available dataset as experimental support, the results demonstrate that the proposed model performs exceptionally well in closed-set classification, even with a significantly reduced training set, greatly enhancing the model's recognition capability. The model improves effectiveness and robustness in real-world applications and presents a promising direction for future research on small-sample models.

Future work will explore integrating attention mechanisms or Transformer modules into the Inception-ResNet architecture to further enhance feature extraction. In addition, research will focus on developing lightweight model structures to meet real-time online monitoring or edge computing needs while minimizing data usage.

References

  1. He Z Y, Liu Q W. Optical fiber distributed acoustic sensors: A review[J]. Journal of Lightwave Technology, 2021, 39(12): 3671-3686. [Google Scholar]
  2. Fu S C, Zhang D, Peng Y, et al. A simulation of gas pipeline leakage monitoring based on distributed acoustic sensing[J]. Measurement Science and Technology, 2022, 33(9): 095108. [Google Scholar]
  3. Fernández-Ruiz M R, Soto M A, Williams E F, et al. Distributed acoustic sensing for seismic activity monitoring[J]. APL Photonics, 2020, 5(3): 030901. [Google Scholar]
  4. Wu Y Y, Gao L, Chai J, et al. Overview of health-monitoring technology for long-distance transportation pipeline and progress in DAS technology application[J]. Sensors, 2024, 24(2): 413. [Google Scholar]
  5. Zhong R J, Chen S K, Liu Z Q, et al. Intrusion identification method for φ-OTDR systems based on few-shot and multi-task learning[C]//2024 IEEE 6th Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC). New York: IEEE, 2024: 1353-1360. [Google Scholar]
  6. Kandamali D F, Cao X M, Tian M L, et al. Machine learning methods for identification and classification of events in φ-OTDR systems: A review[J]. Applied Optics, 2022, 61(11): 2975-2997. [Google Scholar]
  7. Wang M, Feng H, Qi D Z, et al. φ-OTDR pattern recognition based on CNN-LSTM[J]. Optik, 2023, 272: 170380. [Google Scholar]
  8. Li J, Wang L Q, Liu J, et al. Distributed fiber optic vibration sensing event recognition method based on CNN-LSTM-transformer net[J]. Wuhan University Journal of Natural Sciences, 2025, 30(4): 321-333. [Google Scholar]
  9. Duan J Y, Chen J G, He Z Y. DAS-MAE: A self-supervised pre-training framework for universal and high-performance representation learning of distributed fiber-optic acoustic sensing[EB/OL]. [2025-06-05]. arXiv:2506.04552. [Google Scholar]
  10. Szegedy C, Ioffe S, Vanhoucke V, et al. Inception-v4, inception-ResNet and the impact of residual connections on learning[J]. Proceedings of the AAAI Conference on Artificial Intelligence, 2017, 31(1): 4278-4284. [Google Scholar]
  11. Paul S, Das R, Malviya V, et al. Comparative analysis of transfer learning and few-shot learning with CNN architectures for chest X-ray classification under data constraints[J]. Informatica, 2025, 49(4): 1-12. [Google Scholar]
  12. Umapathi K, Vanithamani R, Priya L, et al. Effective skin cancer classification by modified and optimized inception-ResNet-V2 model[J]. Scientific Reports, 2025, 15(1): 41150. [Google Scholar]
  13. Szegedy C, Liu W, Jia Y Q, et al. Going deeper with convolutions[C]//2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). New York: IEEE, 2015: 1-9. [Google Scholar]
  14. He K M, Zhang X Y, Ren S Q, et al. Deep residual learning for image recognition[C]//2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). New York: IEEE, 2016: 770-778. [Google Scholar]
  15. Cao X M, Su Y S, Jin Z Y, et al. An open dataset of φ-OTDR events with two classification models as baselines[J]. Results in Optics, 2023, 10: 100372. [Google Scholar]
  16. Loshchilov I, Hutter F. SGDR: Stochastic gradient descent with warm restarts[EB/OL]. [2024-10-05]. arXiv:1608.03983. [Google Scholar]

All Tables

Table 1

Accuracy under different models and methods (unit:%)

All Figures

Thumbnail: Fig.1 Refer to the following caption and surrounding text. Fig.1 Structure of distributed optical fiber vibration sensing system (DVS)
In the text
Thumbnail: Fig.2 Refer to the following caption and surrounding text. Fig.2 Inception module with dimension reductions
In the text
Thumbnail: Fig.3 Refer to the following caption and surrounding text. Fig.3 Residual connections
In the text
Thumbnail: Fig.4 Refer to the following caption and surrounding text. Fig.4 Inception-ResNet main framework (Stem, Inception-ResNet, Reduction, GAP, Softmax)
In the text
Thumbnail: Fig.5 Refer to the following caption and surrounding text. Fig.5 The stem structure includes the input layer, convolutional layers (Conv-1-11), Max pool-1 and Max pool-2, Avg pool-1, and filter concatenation layers

The input dimension is 10 000×12×1Mathematical equation. Layer parameters and output dimensions are annotated in the figure.

In the text
Thumbnail: Fig.6 Refer to the following caption and surrounding text. Fig.6 The Inception-ResNet structure comprises ReLU activation, convolutional layers (Conv-12–18), and residual addition

Note:   Following ReLU activation, the input is split into three branches, processed through Conv-12-17, concatenated via Conv-18, and combined with the original input using a residual connection. Layer parameters and output dimensions are annotated in the figure.

In the text
Thumbnail: Fig.7 Refer to the following caption and surrounding text. Fig.7 The reduction structure consists of Max pool-3, convolutional layers (Conv-19–22), a filter concatenation layer, and ReLU activation

Layer parameters (kernel size, stride, and padding) and output dimensions are annotated in the figure.

In the text
Thumbnail: Fig. 8 Refer to the following caption and surrounding text. Fig. 8 Grayscale representation
In the text
Thumbnail: Fig.9 Refer to the following caption and surrounding text. Fig.9 The change trend of accuracy
In the text
Thumbnail: Fig.10 Refer to the following caption and surrounding text. Fig.10 The change trend of loss with epochs
In the text
Thumbnail: Fig.11 Refer to the following caption and surrounding text. Fig.11 Confusion matrix
In the text
Thumbnail: Fig.12 Refer to the following caption and surrounding text. Fig.12 Visualization of model attention on fiber-optic vibration signals using Grad-CAM
In the text
Thumbnail: Fig.13 Refer to the following caption and surrounding text. Fig.13 Test set data distribution
In the text

Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.

Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.

Initial download of the metrics may take a while.