banner

Blog

Dec 01, 2023

A deep learning approach for detecting drill bit failures from a small sound dataset

Scientific Reports volume 12, Article number: 9623 (2022) Cite this article

1905 Accesses

4 Citations

3 Altmetric

Metrics details

Monitoring the conditions of machines is vital in the manufacturing industry. Early detection of faulty components in machines for stopping and repairing the failed components can minimize the downtime of the machine. In this article, we present a method for detecting failures in drill machines using drill sounds in Valmet AB, a company in Sundsvall, Sweden that supplies equipment and processes for the production of pulp, paper, and biofuels. The drill dataset includes two classes: anomalous sounds and normal sounds. Detecting drill failure effectively remains a challenge due to the following reasons. The waveform of drill sound is complex and short for detection. Furthermore, in realistic soundscapes, both sounds and noise exist simultaneously. Besides, the balanced dataset is small to apply state-of-the-art deep learning techniques. Due to these aforementioned difficulties, sound augmentation methods were applied to increase the number of sounds in the dataset. In this study, a convolutional neural network (CNN) was combined with a long-short-term memory (LSTM) to extract features from log-Mel spectrograms and to learn global representations of two classes. A leaky rectified linear unit (Leaky ReLU) was utilized as the activation function for the proposed CNN instead of the ReLU. Moreover, an attention mechanism was deployed at the frame level after the LSTM layer to pay attention to the anomaly in sounds. As a result, the proposed method reached an overall accuracy of 92.62% to classify two classes of machine sounds on Valmet's dataset. In addition, an extensive experiment on another drilling dataset with short sounds yielded 97.47% accuracy. With multiple classes and long-duration sounds, an experiment utilizing the publicly available UrbanSound8K dataset obtains 91.45%. Extensive experiments on our dataset as well as publicly available datasets confirm the efficacy and robustness of our proposed method. For reproducing and deploying the proposed system, an open-source repository is publicly available at https://github.com/thanhtran1965/DrillFailureDetection_SciRep2022.

Drill fault detection systems are widely used in factories to prevent machine failure. The drilling machine is included 90 or 120 drill bits to drill thousand of small holes on the surface of the metal1. When drill bits break, there is a need for manual drilling, and post-production, which is resource-demanding and overall costly for the company. The maintenance technician stops the machine every 10 min to identify any broken drill bits and change them before re-initiating the drilling machine. Therefore, a fault detection system for the drilling machine is very crucial to minimize the downtime of the machine as well as the maintenance cost.

Many studies have been conducted on detecting and diagnosing drill failure in the past decade. Choi et al.2 proposed a time domain and frequency domain feature extraction method named characteristic parameters of the drill failure (CPDF). In the second step, a multilayer perceptron (MLP) was used to predict drill failure based on the drill state index threshold. This would lower the error rate. To improve the accuracy of diagnosis for drill failure, Skalle et al.3 proposed a method based on symptom detection (e.g., soft formation, cutting accumulation, local dog leg). Kumar et al.4 used vibration signals to detect and classify drill failures using three different classifier schemes: artificial neural network (ANN), support vector machines (SVM), and Bayesian classifiers. Because vibration-based signals often contain noise, several techniques were required to remove noise and separate the sources to improve fault detection accuracy.

Researchers have used sound and vibration analysis in recent years to detect and classify faults5. As a result of the development of deep learning, Convolutional Neural Networks (CNNs) were used to automatically extract features to diagnose and classify faults on machines, especially drill machines. Due to the advantage of acoustic analysis over vibrations, Glowacz6 proposed an acoustic-based fault detection method for electric impact drills and coffee grinders. These acoustic features, including the root mean square (RMS) and a method selection of amplitude using a multi-expanded filter (MSAF-17-MULTIEXPANDED-FILTER-14), were used to classify fault status by the nearest neighbor classifier. Additionally, to detect the fault in electric impact drills, it is necessary to determine the fault of the gearbox device of the drill, since the gears are the main component of the power transmission. Jing et al.7 proposed a method for detecting electric impact drill failure by using logistic regression from time-varying loudness and acoustic signals.

Recently, a number of techniques have been investigated in the field of fault detection and machine condition monitoring. Hou et al.8 used wavelet packet energy to extract features from acoustic signals, then applied a feature selection method based on the Pearson correlation coefficient to select features. The selected features were used to classify the fault status with a neural network classifier. In addition to synchronous hydraulic motors, this approach can be applied to other rotating machines as well. In another approach, Wang et al.9 proposed a multimodal method to detect bearing faults by fusing acoustic and vibration signals collected from the accelerometer and the microphone using the 1 dimensional CNN.

In recent years, deep learning has had a great deal of success in the detection and diagnosis of mechanical faults by using vibration and acoustics signals10,11,12,13,14,15,16. Besides, recent studies have demonstrated that image representations of sound signals can be used to train the deep learning architecture for sound classification tasks. Researchers have proposed a lot of image representations for sounds, such as Mel-frequency cepstral coefficients (MFCCs)17,18, spectrogram19, Mel spectrogram20. Additionally, many state-of-the-art deep learning models have been used for sound classification. Boddapati et al.19 compared the classification accuracy of AlexNet and GoogleLNet on three different feature representations of sound (spectrogram, MFCC, and cross recurrence plot). A variant of conditional neural networks, called masked conditional neural network (MCLNN) has been proposed by Medhat et al.21 for classifying sounds. Researchers have used dilated CNNs with dilated filters and leaky ReLU activation functions17,22. The effect of modulating the dilation rate in dilated CNN on sound classification was compared in Chen et al.22. Recent studies have shown that recurrent neural networks (RNN) produce excellent results for variable-length sound sequences. Wang et al.23 proposed a CNN architecture with a parallel temporal-spectral attention mechanism to capture certain frames where sound events occur and pay attention to varying frequency bands. Zhang et al.24 proposed a CNN architecture to learn spectro-temporal features and a bidirectional gated recurrent unit (Bi-GRU) with a frame-level attention mechanism for sound classification. Moreover, drilling sound analysis has been used in orthopedic surgical operations, e.g., manual drilling bone. For instance, Torun and Pazarci25 proposed an ANN-based classifier scheme to classify whether breakthrough or non-breakthrough occurred, using parametric-based Power Spectral Density Estimation. Seibold et al.26 fed log-Mel spectrograms of drilling sounds into ResNet-18 to detect drill breakthrough events and demonstrate the potential of deep learning-based acoustic sensing for surgical error prevention.

Our article proposed an approach to detect drill machine failures based on drill sounds from Valmet AB. This is a company in Sundsvall that provides processes and equipment for biofuels production. Valmet AB is currently operating multiple drilling machines to drill holes in metal materials. Most drilling fault detection studies, however, used a large, balanced dataset. Broken drill bits do not occur quite often, hence the sound of broken drill bits only accounts for a small percentage of the total. It is difficult to train advanced deep learning models on small datasets in real-world applications. In addition, the extracted features from raw sound signals are insufficient for classification because the sample duration for sounds in the dataset is around 20.83 ms and 41.67 ms. This makes it more challenging to compare our results to those of previous research in the field of drilling sounds classification. As a result, an end-to-end deep learning system faces many challenges when it comes to detecting drill faults. To overcome these difficulties, data augmentation methods were applied to generate more samples of the dataset. The augmentation methods were shifting the sound by 5 ms and increasing the volume by 2. These sounds in the augmented dataset were converted into log-Mel spectrograms. In addition, a CNN combined with an attention-based LSTM was proposed for classifying drill sounds. Feature maps were extracted from the log-Mel spectrograms using CNN, and then an LSTM layer was used to learn high-level global feature representation from extracted features. Leaky ReLU was used in CNN instead of ReLU to alleviate the potential problem that CNN stops learning when the ReLU has a value of less than zero. Leaky ReLU helps CNN continue learning when input values are negative. To focus on the important parts of drill sounds and discard the unnecessary parts, an attention layer was added after the LSTM.

The proposed architecture is described as shown in Fig. 1. Initially, audio augmentation methods were applied to original sounds to increase the number of samples in the dataset. In the next step, a small CNN architecture that includes five layers was proposed to generate features from the Mel spectrograms of sounds. Finally, these features were used as the input of the LSTM with the attention mechanism to learn high-level feature representation. The details of the layers in our proposed model are described in Table 1, where nC is the number of classes and \((S=1)\) is the stride of 1 for the convolutional layer.

The proposed methodology.

Valmet AB drills small holes in metal plates with multiple machines. There are two types of drilling machines in a factory that are 90 and 120 bits. Figure 2 shows a healthy drill bit and a broken drill bit1. In this dataset, sound from a drill machine in Sundsvall, Sweden was recorded with four AudioBox iTwo Studio microphones. For capturing drill sounds, 96 kHz was used as the sampling rate. The dataset contains 134 sounds with lengths of 20.83ms and 41.67ms in two classes (normal and anomalous).

A healthy drill bit (on the left side) and a broken drill bit (on the right side)1.

Although hyper-parameters of the model were fine-tuned to adapt to the drill fault detection task, however, the lack of sounds when the drills were broken is still a big challenge. To tackle this challenge, data augmentation methods were applied to the original sounds. Thus, the number of sounds in the dataset increased. Besides, data augmentation helps improve the generalization capability of the proposed model. There are many methods of audio augmentation such as time-stretching, pitch-shifting, volume control, noise addition, etc. It is not appropriate to apply some augmentation methods to the sounds in the dataset since they are very short at only 20.83 ms or 41.67 ms. Experimentation revealed that only time-shifting and volume control data augmentation methods are effective for the dataset.

In this article, time-shifting and volume control were applied to generate syntactical sounds. We did not add noise to the sound as an augmentation method because the sound in our dataset is very short. Noise makes it difficult to classify sounds. MATLAB provides a simple function, audioDataAugmenter, to augment the sound. It would be prudent to investigate other augmentation methods when applying the proposed method to other datasets.

A time-shifting is the process of shifting the sound backward or forward at random. The starting point of the sound was shifted by 5 ms to the right, then padded it back to its original length. Figure 3a shows the time representation of the original fault sound and augmented sound using time-shifting.

The time representation of the original fault sound and the augmented sound.

The volume was increased by multiplying the audio by a random amplitude factor. The volume gain was set as 2 dB. Using this technique, we can gain some in-variance concerning the audio input gain. The time representation of the original fault sound and augmented sound using the volume control are shown in Fig. 3b.

Recent advances in the field of image classification using CNN for multiple classes with high accuracy motivated us to investigate the ability of image representation of sounds to detect drill failures. In this paper, drill sounds were converted into log-Mel spectrograms to feed into the proposed CNN. The log-Mel spectrogram was generated as follows. Given a raw drill sound, the Mel spectrogram was computed using short-time Fourier transform (STFT) with Hamming windows of 100 ms and the hop length of 50 ms, the length of FFT was 2048, the sampling rate was 96 kHz, and the number of Mel-filter bank was 96. Since the authors in27 found that the log-scaled Mel spectrogram improves the classification accuracy compared to the Mel spectrogram. Therefore, the logarithm of the Mel spectrogram was taken as the input of the proposed CNN architecture. Figure 4 shows log-Mel spectrograms of an original anomalous sound and its augmented sound using volume control and time-shifting.

Log-Mel spectrograms of an original anomalous sound, the augmented sounds using time-shifting and volume control.

A CNN architecture was proposed for extracting features from log-Mel spectrograms. As a result, the third convolutional layer was used for extracting features instead of adding a dense layer at the end. Additionally, Leaky ReLU was used as the activation function. The experiment results show that using Leaky ReLU can improve the classification accuracy of the dataset. To learn global high-level feature representation, the extracted features were fed into LSTM with an attention mechanism.

The proposed CNN architecture consisted of three convolutional layers and two max-pooling layers, and six batch normalization layers with the Leaky ReLU activation functions. Log-Mel spectrograms were fed into the proposed CNN to extract high-level features for the classification task. Firstly, three convolutional layers with 3 × 3 filter kernel sizes were utilized. Three convolutional layers have 128, 128, and 256 feature maps, respectively. Secondly, a max-pooling layer with 2 × 4 filter kernel sizes were added after the first two convolutional layers. A pair of batch normalization (BN) layers with Leaky ReLU were added before and after the convolutional layers to normalize the features and reduce over-fitting.

The equation for ReLU is \( f(x) = max(0,x)\). When the input of the layer is negative, the ReLU is equal to zero. Consequently, gradient descents reach the value of zero and cannot converge to the local minimum. For Leaky ReLU, there is always a small slope to allow the weight update of the accumulated gradient. Therefore, although ReLU can compute faster, Leaky ReLU was used instead of the ReLU so that the layers did not stop learning when the slope of the ReLU is zero. The Leaky ReLU activation function28 is described by the Eq. (1):

where \(\alpha \) was set to 0.3 in this research.

In this article, LSTM29 was utilized to learn sequential feature maps that are extracted from the proposed CNN. The LSTM unit can be updated as in Eqs. (2)–(7):

where \(X_{t}\) is the mini-batch input; \(i_{t}\) is the input gate; \(f_{t}\) is the forget gate; \(o_{t}\) is the output gate; \(\tilde{c}_{t}\) is the input cell; \(c_{t}\) is the cell state; \(h_{t}\) is the hidden state; \(\sigma \) is the sigmoid function; \(\tau \) is the tanh function; W, U are the weight matrices; b is the bias parameter; t is the time step.

Due to different frame-level features contributing unequally to classifying event sound classes, an attention mechanism30 has been widely used in the sequence-to-sequence model. In this paper, a feed-forward attention layer31 was added after LSTM to specific points in a sequence when computing its output. Additionally, during the transition from the normal state of the drill bit to the broken state, the pitch of the audio changes. Therefore, the features extracted from the log-Mel spectrogram right at the moment the drill bit cracks will have an abnormality. The purpose of the attention layer is to focus on that anomaly. For the LSTM, the output of attention att can be defined as below:

where \(h_{t}\) denotes the \(t_{th}\) hidden output from the LSTM at time step t, T represents the total number of time steps in the input sequence, and the \(\alpha _{t}\) is the attention weight can be computed as follows:

The proposed method was evaluated on our Valmet's dataset. Besides, our proposed method was also verified on a drilling dataset in26 called Seibold's dataset, and a benchmark dataset, namely UrbanSound8K32.

Valmet's drilling dataset includes 134 sounds, divided into two categories: anomaly sounds and normal sounds. After applying time-shifting and volume control augmentation methods to 134 original sounds from two categories, the extended dataset includes 402 sounds. These sounds in the augmented dataset were converted into log-Mel spectrograms to train an end-to-end model. Around 70% of the dataset (280 log-Mel spectrograms) and 30% (122 log-Mel spectrograms) were used for training and testing, respectively. When training the model on the training set, 280 sounds were divided by the ratio of 70/30 for training and validation sets.

Our proposed method is also evaluated on the drilling dataset in26. It consists of two classes, cortical and breakthrough. In this dataset, samples were recorded at a sample rate of 44.1 kHz and a bit depth of 24 bits. Sounds in this dataset are short, just like those in our Valmet's dataset. There are 126 sounds in the cortical category and 136 sounds in the breakthrough category. In the dataset, the sounds have varying lengths but are generally shorter than one second. The breakthrough events last between 100 and 250 ms, which is shorter than sounds in the cortical category. Our proposed data augmentation approach was not applied to this dataset. The dataset is divided into 70% (88 cortical sounds and 95 breakthrough sounds) for training and 30% (38 cortical sounds and 41 breakthrough sounds) for testing. All other experiment setups are the same as those we conducted on Valmet's dataset.

UrbanSound8K32 was used to test the proposed method's effectiveness in multiple classes classification with longer sounds (under or equal to 4 seconds). There are 8732 sounds in this dataset representing urban sounds from 10 classes: air conditioner, car horn, children playing, dog bark, drilling, engine idling, gunshot, jackhammer, siren, and street music. The same experimental setups were used to train 6111 sounds (70% of the dataset) and test 2621 sounds (30%). Data augmentation methods were not applied on this dataset.

The proposed deep learning model was trained on Intel CORE i5 8th Gen with NVIDIA Graphics Card 1050Ti. Keras library33 with TensorFlow toolkit34 that are popular deep learning frameworks were used to implement and deploy the proposed method. Additionally, the Librosa library35 was used to generate log-Mel spectrograms from original drill sounds.

For hyper-parameters optimization, the Adam optimizer36 was used with a learning rate of 0.001, a batch size of 4, a momentum of 0.9, and 100 epochs. During training, categorical cross-entropy was used as the loss function \(L_{f}\) to update the network weights. It is defined as follows:

where nC is the number of classes, \(y_{n}\) is the ground truth, and \(\hat{y}_{n}\) is the predicted class probabilities for the \(n_{th}\) element of model predictions. Furthermore, to avoid over-fitting and to improve the generalized model, early stopping was applied to train the network with the patience of 5.

Table 2 shows the results of our method on the Valmet's dataset and the above-mentioned datasets. Our proposed method reached the accuracy of 92.62% and 97.47% on Valmet's dataset and Seibold's dataset, respectively. The proposed method not only performs well on small datasets with short sounds (Valmet's and Seibold's datasets), but it also obtains a high accuracy on the ten-class UrbanSound8K dataset with longer sounds. From the last column of Table 2, we can see that our model performs the best on the UrbanSound8K, compared to state-of-the-art methods. The performance of our proposed method reached 91.45%, slightly higher than other methods.

As shown in Table 3, the overall accuracy of the proposed method, CNN using the Leaky ReLU activation function combined with attention-based LSTM (CNN-LSTM-Attention-Leaky ReLU), was 92.62%. The confusion matrix for the proposed method is in Fig. 5. Table 3 shows the F1-score, precision, and recall for each class in the augmented dataset.

The confusion matrix for the proposed model (CNN-LSTM-Attention-Leaky ReLU) on the augmented dataset.

The role of the various modules is investigated through ablation experiments using our proposed method. As previously stated, our model has four key components: the CNN module, the LSTM layer, the attention mechanism, and the Leaky ReLU activation function. We analyze each component's role as we eliminate modules one by one in our ablation experiments or change the activation function. The mean accuracy of all the experiments is shown in Table 4 for comparison. Using CNN with Leaky ReLU activation function in conjunction with attention-based LSTM achieves the highest accuracy of 92.62%. This result demonstrates that the Leaky ReLU and the attention mechanism can affect the overall accuracy of the proposed method when combined with CNN and LSTM. The following experiments were conducted to validate the effectiveness of our proposed method:

In CNN architecture, we run experiments with the Leaky ReLU activation function. The experiment parameters were identical to the CNN architecture in the proposed method. However, two fully connected layers we used at the end of CNN for classification. According to Table 4, the overall accuracy for this method was only 86.89%, which is lower than the accuracy of our proposed method (92.62%). Figure 6a shows the confusion matrix for this method. Table 5 shows precision, recall, F1-score for each class.

Comparison of different methods.

This part experimented with CNN uses the Leaky ReLU activation function in conjunction with LSTM. This experiment tests whether incorporating an attention layer into the model is effective. In this method, accuracy achieved 90.16%, which is less than our proposed method (accuracy of 92.62%). It is clear that the accuracy of the model was improved by including the attention layer. In theory, with the attention layer, the LSTM is supposed to invest more computing power of that small but important part of the input, so the network enhances these parts and fades out the rest. The confusion matrix for this method is shown in Fig. 6b. Table 6 shows precision, recall, F1-score for each class.

This part experimented with CNN architecture uses the ReLU activation function in conjunction with attention-based LSTM. The confusion matrix for this method is shown in Fig. 6c. In this experiment, the model was run with ReLU activation to prove it is less effective than Leaky ReLU activation on our dataset. When using the ReLU activation function, the accuracy was 91.80%, while using Leaky ReLU, the accuracy was higher (92.62%). As Leaky ReLU has a slope of 0.3 instead of 0, CNN can train faster and avoid the ‘dying ReLU’ problem on our dataset. Table 7 shows precision, recall, F1-score for each class.

To test the efficiency of the data augmentation process, the proposed model in section 2 was run on both the original and augmented datasets. Table 8 shows precision, recall, F1-score for each class. The accuracy on the augmented dataset reached 92.62% whereas the accuracy on the original dataset only reached 88.10% (Table 9). The accuracy of our proposed method on the augmented dataset (402 sounds) is clearly higher than on the original dataset (201 sounds). The confusion matrix for our proposed method on the original dataset is shown in Fig. 6d.

The sound is too short, and the balanced dataset has too few samples, which are the two major challenges in developing a machine failure detection system for Valmet AB. To begin with, it is difficult to apply data augmentation methods to short sounds. Some modern data augmentation strategies, such as synthesizing new data using generative models, have recently attracted the interest of researchers. GAN, for example, is a common generative model used to synthesis new data from a small dataset in image processing and computer vision. The drilling sounds in Valmet's dataset, on the other hand, are far too short to be used with state-of-the-art GAN. Furthermore, a model with too many parameters may underfit a limited training dataset. When deep learning models fail to catch the underlying trends in data, this is known as underfitting. As a result, the model will make numerous inaccurate predictions. To avoid underfitting, it is required to utilize a larger dataset with longer sounds. However, due to the high costs and labor-intensive nature of capturing and identifying sounds in factories, it is not feasible to collect large and balanced datasets. On a limited dataset, our proposed method can be utilized to develop a classification model. Drill sounds may be recorded and identified using this classification model right at the factory. A skilled technician can confirm the accuracy of the recorded sounds identified by this model. These new sounds are then added to a bigger dataset. When a deep learning model is trained on a larger dataset, it can generate better results.

Our proposed method is validated on Seibold's dataset to demonstrate its effectiveness as well as to avoid bias in the specified dataset from Valmet AB. Results of the experiment indicate that the proposed method achieves greater accuracy than the baseline system using ResNet-1840 in the previous study26 on the same dataset. As shown in Table 2, the mean accuracy of our proposed method on this dataset reached 97.47%, whereas, Seibold et al.26 reached an accuracy of 91.90%. Figure 7 depicts the confusion matrix using Seibold's dataset with our proposed method, while the precision, recall, and F1-score for each class are shown in Table 10. According to these results, our proposed method is capable of efficiently processing short and small sound datasets, such as those of Valmet's and Seibold's datasets.

The confusion matrix for the proposed model (CNN-LSTM-Attention-Leaky ReLU) on Seibold's dataset.

Figure 8 depicts the confusion matrix of the proposed method on the UrbandSound8K benchmark dataset. In Table 11, precision, recall, and F1-score for each class in the UrbandSound8K are presented. As shown in Table 2, our proposed method achieves better accuracy than state-of-the-art and the latest methods on the UrbandSound8K dataset. The mean accuracy of our proposed method is 91.45%, while they are 70.90%, 89.00%, and 90.07% for Stride-Ds-2437, 1D CNN38, and AudioCLIP39, respectively. These findings suggest that our method outperforms modern techniques on multi-class datasets with longer sounds. Additionally, it validates the generalization of our proposed method, which works well not only on small and short sound datasets but also on large datasets with many classes and long sounds.

The confusion matrix for the proposed model (CNN-LSTM-Attention-Leaky ReLU) on UrbanSound8K dataset.

In this article, a deep learning model was proposed for a drill fault detection system. Besides, time-shifting and volume control augmentation methods were applied to increase the number of sounds in the small dataset. The sounds in the augmented dataset were converted into log-Mel spectrograms and were used to train the proposed CNN architecture with the Leaky ReLU activation function in conjunction with attention-based LSTM for detecting drill failure. It was found that the overall accuracy of our proposed system reached 92.62% on our Valmet's dataset. In terms of identifying broken drill bits, the accuracy of the proposed method is acceptable. This method has a huge potential to be used to diagnose faults in industrial machines. It is a non-invasive method of diagnosing machine failure using short sounds or small datasets. Moreover, this paper used both a private dataset with small and short sounds namely Seibold's dataset, and a benchmark UrbanSound8K dataset to validate the effectiveness and generalization of the proposed method. Studies show that our proposed method is more accurate than the state-of-the-art and the latest methods on both datasets. On the Seibold's and UrbandSound8K datasets, our proposed method reaches an accuracy of 97.47% and 91.45%, respectively. We are considering combining sound and images to identify drilling errors and improve drilling error detection results in the future. Aspects of having many events in the same sound, such as polyphonic sounds, which blended both anomalous drill sounds and others, will also be examined.

The datasets generated during and/or analysed during the current study are available from the corresponding author on reasonable request.

Tran, T. & Lundgren, J. Drill fault diagnosis based on the scalogram and MEL spectrogram of sound signals using artificial intelligence. IEEE Access 8, 203655–203666. https://doi.org/10.1109/ACCESS.2020.3036769 (2020).

Article Google Scholar

Choi, Y. J., Park, M. S. & Chu, C. N. Prediction of drill failure using features extraction in time and frequency domains of feed motor current. Int. J. Mach. Tools Manuf. 48, 29–39 (2008).

Article Google Scholar

Skalle, P., Aamodt, A. & Gundersen, O. E. Detection of symptoms for revealing causes leading to drilling failures. SPE Drill. Complet. 28, 182–193 (2013).

Article Google Scholar

Kumar, A., Ramkumar, J., Verma, N. K. & Dixit, S. Detection and classification for faults in drilling process using vibration analysis. In 2014 International Conference on Prognostics and Health Management, 1–6 (IEEE, 2014).

Henriquez, P., Alonso, J. B., Ferrer, M. A. & Travieso, C. M. Review of automatic fault diagnosis systems using audio and vibration signals. IEEE Trans. Syst. Man. Cybern. Syst. 44, 642–652. https://doi.org/10.1109/TSMCC.2013.2257752 (2014).

Article Google Scholar

Glowacz, A. Fault detection of electric impact drills and coffee grinders using acoustic signals. Sensors 19, 269 (2019).

Article ADS Google Scholar

Jing, Y., Su, H., Wang, S., Gui, W. & Guo, Q. Fault diagnosis of electric impact drills based on time-varying loudness and logistic regression. Shock Vib. 2021 (2021).

Hou, J., Sun, H., Xu, A., Gong, Y. & Ning, D. Fault diagnosis of synchronous hydraulic motor based on acoustic signals. Adv. Mech. Eng. 12, 1687814020916107 (2020).

Google Scholar

Wang, X., Mao, D. & Li, X. Bearing fault diagnosis based on vibro-acoustic data fusion and 1D-CNN network. Measurement 173, 108518 (2021).

Article Google Scholar

Polat, K. The fault diagnosis based on deep long short-term memory model from the vibration signals in the computer numerical control machines. J. Inst. Electron. Comput. 2, 72–92 (2020).

Article Google Scholar

González-Muñiz, A., Díaz, I. & Cuadrado, A. A. DCNN for condition monitoring and fault detection in rotating machines and its contribution to the understanding of machine nature. Heliyon 6, e03395 (2020).

Article Google Scholar

Verstraete, D., Ferrada, A., Droguett, E. L., Meruane, V. & Modarres, M. Deep learning enabled fault diagnosis using time-frequency image analysis of rolling element bearings. Shock Vib. 2017 (2017).

Zhang, S., Zhang, S., Wang, B. & Habetler, T. G. Deep learning algorithms for bearing fault diagnostics—A comprehensive review. IEEE Access 8, 29857–29881 (2020).

Article Google Scholar

Chen, Z., Chen, X., Li, C., Sanchez, R.-V. & Qin, H. Vibration-based gearbox fault diagnosis using deep neural networks. J. Vibroeng. 19, 2475–2496 (2017).

Article Google Scholar

Islam, M. M. & Kim, J.-M. Motor bearing fault diagnosis using deep convolutional neural networks with 2d analysis of vibration signal. In Canadian Conference on Artificial Intelligence, 144–155 (Springer, 2018).

Xueyi, L., Jialin, L., Yongzhi, Q. & David, H. Semi-supervised gear fault diagnosis using raw vibration signal based on deep learning. Chin. J. Aeronaut. 33, 418–426 (2020).

Article Google Scholar

Zhang, X., Zou, Y. & Shi, W. Dilated convolution neural network with LeakyReLU for environmental sound classification. In International Conference on Digital Signal Processing, DSP 2017-Augus, https://doi.org/10.1109/ICDSP.2017.8096153 (2017).

Davis, N. & Suresh, K. Environmental sound classification using deep convolutional neural networks and data augmentation. In 2018 IEEE Recent Advances in Intelligent Computational Systems (RAICS), 41–45, https://doi.org/10.1109/RAICS.2018.8635051 (2018).

Boddapati, V., Petef, A., Rasmusson, J. & Lundberg, L. Classifying environmental sounds using image recognition networks. Procedia Comput. Sci. 112, 2048–2056. https://doi.org/10.1016/j.procs.2017.08.250 (2017).

Article Google Scholar

Mushtaq, Z., Su, S. F. & Tran, Q. V. Spectral images based environmental sound classification using CNN with meaningful data augmentation. Appl. Acoust. 172, 107581. https://doi.org/10.1016/j.apacoust.2020.107581 (2021).

Article Google Scholar

Medhat, F., Chesmore, D. & Robinson, J. Masked conditional neural networks for environmental sound classification. In Artificial Intelligence XXXIV - 37th SGAI International Conference on Artificial Intelligence, AI 2017, Cambridge, UK, December 12–14, 2017, Proceedings, Vol. 10630 of Lecture Notes in Computer Science, (eds Bramer, M. & Petridis, M.) 21–33, https://doi.org/10.1007/978-3-319-71078-5_2 (Springer, 2017).

Chen, Y., Guo, Q., Liang, X., Wang, J. & Qian, Y. Environmental sound classification with dilated convolutions. Appl. Acoust. 148, 123–132. https://doi.org/10.1016/j.apacoust.2018.12.019 (2019).

Article Google Scholar

Wang, H., Zou, Y., Chong, D. & Wang, W. Environmental sound classification with parallel temporal-spectral attention. In Interspeech 2020, 21st Annual Conference of the International Speech Communication Association, Virtual Event, Shanghai, China, 25-29 October 2020, (eds Meng, H. et al.) 821–825, https://doi.org/10.21437/Interspeech.2020-1219 (ISCA, 2020).

Zhang, Z., Xu, S., Zhang, S., Qiao, T. & Cao, S. Attention based convolutional recurrent neural network for environmental sound classification. Neurocomputinghttps://doi.org/10.1016/j.neucom.2020.08.069 (2020).

Article PubMed PubMed Central Google Scholar

Torun, Y. & Pazarci, Ö. Parametric power spectral density estimation-based breakthrough detection for orthopedic bone drilling with acoustic emission signal analysis. Acoust. Aust. 48, 221–231 (2020).

Article Google Scholar

Seibold, M. et al. Real-time acoustic sensing and artificial intelligence for error prevention in orthopedic surgery. Sci. Rep. 11, 1–11 (2021).

Article ADS Google Scholar

Choi, K., Fazekas, G., Sandler, M. & Cho, K. A comparison of audio signal preprocessing methods for deep neural networks on music tagging. In 2018 26th European Signal Processing Conference (EUSIPCO), 1870–1874 (IEEE, 2018).

Maas, A. L., Hannun, A. Y. & Ng, A. Y. Rectifier nonlinearities improve neural network acoustic models. in ICML Workshop on Deep Learning for Audio, Speech and Language Processing, Vol. 28, (2013).

Hochreiter, S. & Schmidhuber, J. Long short-term memory. Neural Comput. 9, 1735–1780 (1997).

Article CAS Google Scholar

Vaswani, A. et al. Attention is all you need. In Advances in Neural Information Processing Systems, 5998–6008 (2017).

Raffel, C. & Ellis, D. P. W. Feed-forward networks with attention can solve some long-term memory problems. CoRR (2015). arXiv:1512.08756.

Salamon, J., Jacoby, C. & Bello, J. P. A dataset and taxonomy for urban sound research. In 22nd ACM International Conference on Multimedia (ACM-MM’14), 1041–1044 (2014).

Chollet, F. et al. Keras. https://github.com/fchollet/keras (2015).

Abadi, M. et al. TensorFlow: Large-scale machine learning on heterogeneous systems (2015). Software available from tensorflow.org.

McFee, B. et al. librosa/librosa: 0.8.1rc2, https://doi.org/10.5281/zenodo.4792298 (2021).

Kingma, D. P. & Ba, J. Adam: A method for stochastic optimization. In 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings (eds Bengio, Y. & LeCun, Y.) (2015).

Nordby, J. O. Environmental sound classification on microcontrollers using Convolutional Neural Networks. Master's thesis, Norwegian University of Life Sciences, Ås (2019).

Abdoli, S., Cardinal, P. & Koerich, A. L. End-to-end environmental sound classification using a 1D convolutional neural network. Expert Syst. Appl. 136, 252–263 (2019).

Article Google Scholar

Guzhov, A., Raue, F., Hees, J. & Dengel, A. Audioclip: Extending clip to image, text and audio. arXiv preprint arXiv:2106.13043 (2021).

He, K., Zhang, X., Ren, S. & Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 770–778 (2016).

Download references

This research was supported by the EU Regional Fund, the MiLo Project (No. 20201888), and the Acoustic sensor set for AI monitoring systems (AISound ) project. The authors would like to thank Valmet AB for providing the drill sound dataset. The authors are grateful to Matthias Seibold at the Technical University of Munich for sharing their dataset.

Open access funding provided by Mid Sweden University.

Department of Electronics Design, Mid Sweden University, Sundsvall, Sweden

Thanh Tran & Jan Lundgren

Division of Computational Mechatronics, Institute for Computational Science, Ton Duc Thang University, Ho Chi Minh City, Vietnam

Nhat Truong Pham

Faculty of Electrical and Electronics Engineering, Ton Duc Thang University, Ho Chi Minh City, Vietnam

Nhat Truong Pham

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

Conceptualization: T.T.; Methodology: T.T.; Software: T.T. and N.T.P.; Data curation: T.T.; Writing—original draft: T.T. and N.T.P.; Writing—review and editing: J.L.; Supervision: J.L.; All authors reviewed the manuscript.

Correspondence to Thanh Tran.

The authors declare no competing interests.

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and Permissions

Tran, T., Pham, N.T. & Lundgren, J. A deep learning approach for detecting drill bit failures from a small sound dataset. Sci Rep 12, 9623 (2022). https://doi.org/10.1038/s41598-022-13237-7

Download citation

Received: 01 January 2022

Accepted: 23 May 2022

Published: 10 June 2022

DOI: https://doi.org/10.1038/s41598-022-13237-7

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.

SHARE