Skip to main content
Log in

Multichannel convolution neural network for gas mixture classification

  • Original Research
  • Published:
Annals of Operations Research Aims and scope Submit manuscript

Abstract

Concomitant with people beginning to understand their legal rights or entitlement to complain, complaints of offensive odors and smell pollution have increased significantly. Consequently, monitoring gases and identifying their types and causes in real time has become a critical issue in the modern world. In particular, toxic gases that may be generated at industrial sites or odors in daily life consist of hybrid gases made up of various chemicals. Understanding the types and characteristics of these mixed gases is an important issue in many areas. However, mixed gas classification is challenging because the gas sensor arrays for mixed gases must process complex nonlinear high-dimensional data. In addition, obtaining sufficient training data is expensive. To overcome these challenges, this paper proposes a novel method for mixed gas classification based on analogous image representations with multiple sensor-specific channels and a convolutional neural network (CNN) classifier. The proposed method maps a gas sensor array into a multichannel image with data augmentation, and then utilizes a CNN for feature extraction from such images. The proposed method was validated using public mixture gas data from the UCI machine learning repository and real laboratory experiments. The experimental results indicate that it outperforms the existing classification approaches in terms of the balanced accuracy and weighted F1 scores. Additionally, we evaluated the performance of the proposed method in various experimental settings in terms of data representation, data augmentation, and parameter initialization, so that practitioners can easily apply it to artificial olfactory systems.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12

Similar content being viewed by others

References

  • Bagnall, A., Lines, J., Bostrom, A., Large, J., & Keogh, E. (2017). The great time series classification bake off: a review and experimental evaluation of recent algorithmic advances. Data mining and knowledge discovery, 31, 606–660.

    Article  Google Scholar 

  • Baydogan, M. G., Runger, G., & Tuv, E. (2013). A bag-of-features framework to classify time series. IEEE transactions on pattern analysis and machine intelligence, 35, 2796–2802.

    Article  Google Scholar 

  • Behera, B., & Kumaravelan, G., et al. (2019). Performance evaluation of deep learning algorithms in biomedical document classification, in 2019 11th International Conference on Advanced Computing (ICoAC), IEEE, pp. 220–224.

  • Carmel, L., Levy, S., Lancet, D., & Harel, D. (2003). A feature extraction method for chemical sensors in electronic noses. Sensors and Actuators B: Chemical, 93, 67–76.

    Article  Google Scholar 

  • Chen, Z., Chen, Z., Song, Z., Ye, W., & Fan, Z. (2019). Smart gas sensor arrays powered by artificial intelligence. Journal of Semiconductors, 40, 111601.

    Article  Google Scholar 

  • Deng, C., Lv, K., Shi, D., Yang, B., Yu, S., He, Z., & Yan, J. (2018). Enhancing the discrimination ability of a gas sensor array based on a novel feature selection and fusion framework. Sensors, 18, 1909.

    Article  Google Scholar 

  • Distante, C., Leo, M., Siciliano, P., & Persaud, K. C. (2002). On the study of feature extraction methods for an electronic nose. Sensors and Actuators B: Chemical, 87, 274–288.

    Article  Google Scholar 

  • Eklöv, T., Mårtensson, P., & Lundström, I. (1997). Enhanced selectivity of MOSFET gas sensors by systematical analysis of transient parameters. Analytica Chimica Acta, 353, 291–300.

    Article  Google Scholar 

  • Faleh, R., Othman, M., Gomri, S., Aguir, K., & Kachouri, A. (2016). A transient signal extraction method of WO 3 gas sensors array to identify polluant gases. IEEE Sensors Journal, 16, 3123–3130.

    Article  Google Scholar 

  • Fawaz, H. I., Forestier, G., Weber, J., Idoumghar, L., & Muller, P.-A. (2019). Deep learning for time series classification: a review. Data Mining and Knowledge Discovery, 33, 917–963.

    Article  Google Scholar 

  • Feng, S., Farha, F., Li, Q., Wan, Y., Xu, Y., Zhang, T., & Ning, H. (2019). Review on smart gas sensing technology. Sensors, 19, 3760.

    Article  Google Scholar 

  • Ferri, C., Hernández-Orallo, J., & Modroiu, R. (2009). An experimental comparison of performance measures for classification. Pattern Recognition Letters, 30, 27–38.

    Article  Google Scholar 

  • Fonollosa, J., Rodríguez-Luján, I., Trincavelli, M., Vergara, A., & Huerta, R. (2014). Chemical discrimination in turbulent gas mixtures with mox sensors validated by gas chromatography-mass spectrometry. Sensors, 14, 19336–19353.

    Article  Google Scholar 

  • Fonollosa, J., Sheik, S., Huerta, R., & Marco, S. (2015). Reservoir computing compensates slow response of chemosensor arrays exposed to fast varying gas concentrations in continuous monitoring. Sensors and Actuators B: Chemical, 215, 618–629.

    Article  Google Scholar 

  • García, V., Mollineda, R. A., & Sánchez, J. S. (2009). Index of balanced accuracy: A performance measure for skewed class distributions, in Iberian conference on pattern recognition and image analysis, Springer, pp. 441–448.

  • Han, L., Yu, C., Xiao, K., & Zhao, X. (2019). A new method of mixed gas identification based on a convolutional neural network for time series classification. Sensors, 19, 1960.

    Article  Google Scholar 

  • Hasibi, R., Shokri, M., & Dehghan, M. (2019). Augmentation scheme for dealing with imbalanced network traffic classification using deep learning, arXiv preprintarXiv:1901.00204.

  • He, K., Zhang, X., Ren, S., & Sun, J. (2016a). Deep residual learning for image recognition, in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778.

  • He, K., Zhang, X., Ren, S., & Sun, J. (2016a). (2016b), Identity mappings in deep residual networks, in European conference on computer vision, Springer, pp. 630–645.

  • Jamal, M., Khan, M., Imam, S. A., & Jamal, A. (2010). Artificial neural network based e-nose and their analytical applications in various field, in 2010 11th International Conference on Control Automation Robotics & Vision, IEEE, pp. 691–698.

  • Jeni, L. A., Cohn, J. F., & De La Torre, F. (2013). Facing imbalanced data–recommendations for the use of performance metrics, In 2013 Humaine association conference on affective computing and intelligent interaction, IEEE, pp. 245–251.

  • Jia, Y., Yu, B., Du, M., & Wang, X. (2018). Gas Composition Recognition Based on Analyzing Acoustic Relaxation Absorption Spectra: Wavelet Decomposition and Support Vector Machine Classifier, in 2018 2nd International Conference on Electrical Engineering and Automation (ICEEA 2018), Atlantis Press, pp. 126–130.

  • Kalantar-Zadeh, K. (2013). Sensors: an introductory course, Springer Science & Business Media.

  • Karakaya, D., Ulucan, O., & Turkan, M. (2020). Electronic nose and its applications: a survey. International Journal of Automation and Computing, 17, 179–209.

    Article  Google Scholar 

  • Kim, E., Lee, S., Kim, J. H., Kim, C., Byun, Y. T., Kim, H. S., & Lee, T. (2012). Pattern recognition for selective odor detection with gas sensor arrays. Sensors, 12, 16262–16273.

    Article  Google Scholar 

  • Kingma, D. P. & Ba, J. (2014). Adam: A method for stochastic optimization, arXiv preprintarXiv:1412.6980.

  • Laref, R., Losson, E., Sava, A., & Siadat, M. (2018). Support vector machine regression for calibration transfer between electronic noses dedicated to air pollution monitoring. Sensors, 18, 3716.

    Article  Google Scholar 

  • Moore, S., Gardner, J., Hines, E., Göpel, W., & Weimar, U. (1993). A modified multilayer perceptron model for gas mixture analysis. Sensors and Actuators B: Chemical, 16, 344–348.

    Article  Google Scholar 

  • Nakata, S., Akakabe, S., Nakasuji, M., & Yoshikawa, K. (1996). Gas sensing based on a nonlinear response: discrimination between hydrocarbons and quantification of individual components in a gas mixture. Analytical chemistry, 68, 2067–2072.

    Article  Google Scholar 

  • Oh, Y. & Kim, S. (2021), Multi-channel Convolution Neural Network for Gas Mixture Classification, In 2021 International Conference on Data Mining Workshops (ICDMW), IEEE, pp. 1094–1095.

  • O’Hara, S. & Draper, B. A. (2011), Introduction to the bag of features paradigm for image classification and retrieval, arXiv preprintarXiv:1101.3354.

  • Ohnishi, M. (1992). A molecular recognition system for odorants incorporating biomimetic gas-sensitive devices using Langmuir-Blodgett films. Sens. Mater, 4, 53–60.

    Google Scholar 

  • Olson, M., Wyner, A. J., & Berk, R. (2018), Modern neural networks generalize on small data sets, In Proceedings of the 32nd International Conference on Neural Information Processing Systems, pp. 3623–3632.

  • Pancoast, S. & Akbacak, M. (2012). Bag-of-audio-words approach for multimedia event classification, in Thirteenth Annual Conference of the International Speech Communication Association.

  • Pashami, S., Lilienthal, A. J., & Trincavelli, M. (2012). Detecting changes of a distant gas source with an array of MOX gas sensors. Sensors, 12, 16404–16419.

    Article  Google Scholar 

  • Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., & Dubourg, V., et al. (2011). Scikit-learn: Machine learning in Python, the Journal of machine Learning research, 12, 2825–2830.

  • Peng, P., Zhao, X., Pan, X., & Ye, W. (2018). Gas classification using deep convolutional neural networks. Sensors, 18, 157.

    Article  Google Scholar 

  • Persaud, K., & Dodd, G. (1982). Analysis of discrimination mechanisms in the mammalian olfactory system using a model nose. Nature, 299, 352–355.

    Article  Google Scholar 

  • Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., et al. (2015). Imagenet large scale visual recognition challenge. International journal of computer vision, 115, 211–252.

    Article  Google Scholar 

  • Schütze, H., Manning, C. D., & Raghavan, P. (2008). Introduction to information retrieval, vol. 39, Cambridge University Press Cambridge.

  • Shorten, C., & Khoshgoftaar, T. M. (2019). A survey on image data augmentation for deep learning. Journal of Big Data, 6, 1–48.

    Article  Google Scholar 

  • Sokolova, M., & Lapalme, G. (2009). A systematic analysis of performance measures for classification tasks. Information processing & management, 45, 427–437.

    Article  Google Scholar 

  • Sujono, H. A., Rivai, M., & Amin, M. (2018). Asthma identification using gas sensors and support vector machine. Telkomnika, 16, 1468–1480.

    Article  Google Scholar 

  • Tan, C. W., Petitjean, F., Keogh, E., & Webb, G. I. (2019). Time series classification for varying length series, arXiv preprintarXiv:1910.04341.

  • Thammarat, P., Kulsing, C., Wongravee, K., Leepipatpiboon, N., & Nhujak, T. (2018). Identification of volatile compounds and selection of discriminant markers for elephant dung coffee using static headspace gas chromatography-Mass spectrometry and chemometrics. Molecules, 23, 1910.

    Article  Google Scholar 

  • Um, T. T., Pfister, F. M., Pichler, D., Endo, S., Lang, M., Hirche, S., Fietzek, U., & Kulić, D. (2017). Data augmentation of wearable sensor data for parkinson’s disease monitoring using convolutional neural networks, in Proceedings of the 19th ACM International Conference on Multimodal Interaction, pp. 216–220.

  • Wang, Q., Li, L., Ding, W., Zhang, D., Wang, J., Reed, K., & Zhang, B. (2019). Adulterant identification in mutton by electronic nose and gas chromatography-mass spectrometer. Food Control, 98, 431–438.

    Article  Google Scholar 

  • Wei, G., Li, G., Zhao, J., & He, A. (2019). Development of a LeNet-5 gas identification CNN structure for electronic noses. Sensors, 19, 217.

    Article  Google Scholar 

  • Xiaobo, Z., Jiewen, Z., Shouyi, W., & Xingyi, H. (2003). Vinegar classification based on feature extraction and selection from tin oxide gas sensor array data. Sensors, 3, 101–109.

    Article  Google Scholar 

  • Yan, J., Guo, X., Duan, S., Jia, P., Wang, L., Peng, C., & Zhang, S. (2015). Electronic nose feature extraction methods: A review. Sensors, 15, 27804–27831.

    Article  Google Scholar 

  • Yan, J., Tian, F., He, Q., Shen, Y., Xu, S., Feng, J., & Chaibou, K. (2012). Feature extraction from sensor data for detection of wound pathogen based on electronic nose. Sensors and Materials, 24, 57–73.

    Google Scholar 

  • Yan, M., Tylczak, J., Yu, Y., Panagakos, G., & Ohodnicki, P. (2018). Multi-component optical sensing of high temperature gas streams using functional oxide integrated silica based optical fiber sensors. Sensors and Actuators B: Chemical, 255, 357–365.

    Article  Google Scholar 

  • Yang, J., Sun, Z., & Chen, Y. (2016). Fault detection using the clustering-kNN rule for gas sensor arrays. Sensors, 16, 2069.

    Article  Google Scholar 

  • Zhan, C., He, J., Pan, M., & Luo, D. (2021). Component Analysis of Gas Mixture Based on One-Dimensional Convolutional Neural Network. Sensors, 21, 347.

    Article  Google Scholar 

  • Zhang, Y., Jin, R., & Zhou, Z.-H. (2010). Understanding bag-of-words model: a statistical framework. International Journal of Machine Learning and Cybernetics, 1, 43–52.

    Article  Google Scholar 

  • Zhao, X., Wen, Z., Pan, X., Ye, W., & Bermak, A. (2019). Mixture gases classification based on multi-label one-dimensional deep convolutional neural network. IEEE Access, 7, 12630–12637.

    Article  Google Scholar 

Download references

Acknowledgements

This work was partly supported by Institute of Information & Communications Technology Planning & Evaluation (IITP) Grant funded by the Korea Government(MSIT) (No. 2021-0-02068, Artificial Intelligence Innovation Hub) and the National Research Foundation of Korea (NRF) Grant funded by the Korea government (MSIT) (NRF-2021R1F1A1061038).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Sungil Kim.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendices

Appendix A. The UCI dataset: data characteristics

  1. I.

    Sensor information to measure gas mixture

    The UCI data were collected using eight sensors, as shown in Table 10. In particular, the MOX gas sensor value used for data measurement varies according to the change in the resistance value of the semiconductor oxide film, at which time the oxide film of the semiconductor reacts to the gas molecule. Therefore, the response value of mixed gases cannot simply be expressed as a linear combination of the response values of individual gases and, to accurately predict the type and concentration of the desired gases, it must be able to reflect the characteristics inherent in mixed gases described earlier.

  2. II.

    Details of data collection and experiments

    The UCI dataset utilizes three types of gases: carbon monoxide (CO), methane, and ethylene. The experimental environment had two gas outlets: one outlet emitted 2500 ppm of ethylene, while the other outlet emitted 1000 ppm of methylene and 4000 ppm of CO. Each emission speed was varied, and the released gases were sprayed through turbines, making them a mixture of varying concentrations when the wind-blown gases touched the sensors. The concentrations of N, L, M, and H for each gas are listed in Table 11. The actual generated gas mixture samples were classified into five classes, as shown in Table 12.

Table 10 Sensor details and representation of UCI data
Table 11 Concentration levels in the UCI dataset
Table 12 Class information in the UCI dataset

Appendix B. Real data: data characteristics

Table 13 Sensor details and representation of real data
Table 14 Experimental setting of real data
Table 15 Concentration levels in the real data
  1. I.

    Sensor information to measure gas mixture

    The data used in this study were collected using a total of 11 gas sensor arrays, as shown in Table 13 as data from the Taesung Environmental Research Institute. There were two MOX type, one NDIR type, and eight ELECTRONICHEMICAL type sensors.

    The MOX sensor is based on the conductivity change of the gas-sensitive MOX semiconductor layer. The NDIR sensor recognizes gas molecules using infrared light. When infrared rays are absorbed by gas molecules, they vibrate the gas molecules at specific wavelengths. The electronic chemical sensor makes measurements by oxidizing or reducing the gas it wants to detect when it enters the internal sensor.

  2. II.

    Details of data collection and experiments

    The data from the Taesung Environmental Research Institute were created by varying the injection volumes of three types of standard gases: dimethyl sulfide, butyl acetate, and toluene. Standard gas was prepared at each concentration of 10,000 ppm using the following formula:

    $$\begin{aligned} V_i[\mu l] = \frac{C_t[ppm] \times V_t[L] \times w_m[g/mol]}{22.4 \times 1000 \times d[g/ml]} \times \frac{\textit{p}[\%]}{100}. \end{aligned}$$

    \(V_i\) represents the amount of injection gas (injection volume), \(C_t\) represents the target concentration (target concentration), \(V_t\) represents the target volume (Target Volume), \(w_m\) indicates the atomic volume of the injection gas (density), and p indicates the purity of the injection gas (purity). The gas was injected to match the concentration and volume of the target mixture, and the experiment was conducted, as shown in Table 14.

    A mixture of 1ml=0.001l of standard gas generated by the above process and 10l of nitrogen can produce 1ppm of mixed gas, and several standard gases can be used to produce mixed gases with different proportions and concentrations.

    A total of 100 samples were generated and classified into four classes according to the ratio of the injection amount, with the characteristics and numbers of each class equal to those in Table 15. The numbers in parentheses indicate the amount of standard gas injected during the sample generation process. For example, class 1 is a specimen injected with 1000 ppm of dimethyl sulfide and 1000 ppm of toluene. In this study, we focused on the existence of a specific chemical in a gas mixture rather than its ratio or amount. Therefore, in the case of class 4, diverse combinations of concentrations were treated as the same class.

Appendix C. Details of model architecture and learning strategy

  1. I.

    Model architectures

    There are several major differences between ResNet models for ImageNet classification and Multichannel ResNet models for gas mixture classification. First, the input size and normalization parameters differ. The ImageNet input images were cropped to a size of \(224\times 224\) with normalization with three channels: mean=[0.485, 0.456, 0.406] and std=[0.229, 0.224, 0.225]. The proposed model uses a larger number of channels. We use a larger resizing to \(256\times 256\) and the same normalization for all channels, such as mean=\([0.5, \cdots , 0.5]\) and std=\([0.5, \cdots , 0.5]\).

    The first input layer is changed because of the larger input size, which aggregates the multichannel input to the designated output dimension. The first convolutional layer uses a smaller kernel size and stride. Therefore, the proposed architecture does not reduce the input after the first convolutional layer. ResNet works as downsampling to reduce the image and computation. In our case, we used more expensive operations because of the small number of data and features. In addition, the last fully connected layer changed to the number of classes, from 1000 labels in ImageNet to four or five labels in our cases. This is a general modification of transfer learning. Therefore, we changed the major components in the model for our datasets and tasks.

  2. II.

    Training Parameters Setting

    Five-fold cross validation was used to avoid over-fitting. The models were trained with a batch size of 4, Adam optimizer, learning rate of \(10^{-6}\) with weight decay of \(10^{-8}\), and 50 epochs. The weighted cross-entropy loss function was used, and the best model was selected based on the minimum loss in the validation set.

Table 16 Comparison model: ResNet18 versus multichannel ResNet18
Table 17 Comparison model: ResNet50 versus multichannel ResNet50

Appendix D. The UCI dataset: Model Performances

  1. (I)

    Machine models with the Bag-of-Features representation

  2. (II)

    Comparison of Noise-added data augmentation

  3. (III)

    CNN models with the analogous image matrix representation

Table 18 The UCI dataset: performance metrics for BoF based models
Table 19 The UCI dataset: performance metrics for noisy augmented data
Table 20 The UCI dataset: performance metrics for analogous image matrix-based models

Appendix E. Real data: Model Performances

  1. (I)

    Machine models with the Bag-of-Features representation

  2. (II)

    Comparison of Noise-added data augmentation

  3. (III)

    CNN models with analogous image matrix representation

Table 21 Real data: Performance metrics for BoF-based models
Table 22 Real data: Performance metrics for noisy augmented data
Table 23 Real data: performance metrics for analogous image matrix-based models

Appendix F. The bag-of-features approach

The BoF approach represents complex objects with feature vectors of sub-objects. Originally, the main idea of the BoF approach is well-known as bag-of-words (BoW) and developed in document classification. The main idea is to define a code book and display it as a histogram, with each item being the number of code books that have occurred in the document. While ignoring word order information, BoW models are still very effective in capturing document information because they are good at navigating document code word frequency information (Zhang et al., 2010; Schütze et al., 2008) Also, BoF is a well-known dictionary-based feature extraction method for time-series classification (O’Hara & Draper 2011; Pancoast & Akbacak 2012; Baydogan et al., 2013). Furthermore, BoF can be easily applied when the observations have different sequence lengths (Tan et al., 2019).

Feature selection or feature extraction is a common approach to analyze high-dimensional data. We implemented BoF representation for multivariate time series. The BoF representation can represent the original data into lower-dimensional features. We had N samples of the gas sensor array in each dataset. In this study, the bag-of-features representation was the baseline for evaluating the proposed approach. compared to the image transformation, we used the individual observations in the gas sensor array as a feature:. \({\mathbf {X}} =\left[ {\mathbf {z}}_1\; {\mathbf {z}}_2\;\cdots \; {\mathbf {z}}_T\right] \), where \({\mathbf {z}}_j= (x_{1j}, \ldots , x_{Mj})^T\), \(j=1,\ldots ,T\). Here, we use \({\mathbf {z}}_j\) as a sample for the BoF.

To build a codebook, \({\mathbf {z}}_j\) can be quantized using a clustering algorithm, hence the K-means clustering model g with \(K=128\) was trained. K-means clustering generates K centroids, and each vector \({\mathbf {z}}_j\) is classified with the centroid index closest to it. In other words, \(g ({\mathbf {z}}_j) = k\), where \(j \in \left\{ 1, \cdots , T\right\} \) and \(k \in \left\{ 1, \cdots , K\right\} \). Subsequently, we used the centroid index as a codebook for each sample. Each centroid can be seen as a “word” in the dictionary and each sample as a “document, ” which has a different set of words. We used the frequency ratio of each centroid index as a feature of the samples. In this approach, K is the size of the codebook, which can affect the classifier model, but the effect of the codebook size was not investigated in this research.

Fig. 13
figure 13

Sample data plot

We used a sufficiently large K to represent each sample’s characteristics. Following the proposed approach, each sample was transformed into a K-dimensional vector. The original data \(\mathbf{X} \) became a feature vector \(\mathbf{f} \) with a size of K.

$$\begin{aligned} \mathbf{f} = \left[ f_k \mid f_k = \frac{\# (z \textit{ classified as cluster } k)}{T} \quad \textit{where } k \in \left\{ 1, \cdots , K\right\} \right] \end{aligned}$$

Through this process, we could extract K-dimensional features from the \(M \times T\) matrix. Once features were selected by using the bag-of-features, we used logistic regression (LR), support vector machine (SVM), Naive Bayes (NB), K-nearest neighbor (KNN), decision tree (DT), random forest (RF), extreme gradient boosting (XGBoost), light gradient boosting machine (LGBM), and multilayer perceptron (MLP) to classify the classes of samples.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Oh, Y., Lim, C., Lee, J. et al. Multichannel convolution neural network for gas mixture classification. Ann Oper Res (2022). https://doi.org/10.1007/s10479-022-04715-2

Download citation

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10479-022-04715-2

Keywords

Navigation