| Literature DB >> 32094366 |
Xintao Chai1, Hanming Gu2, Feng Li3, Hongyou Duan4, Xiaobo Hu4, Kai Lin1.
Abstract
Deep learning (DL) is a powerful tool for mining features from data, which can theoretically avoid assumptions (e.g., linear events) constraining conventional interpolation methods. Motivated by this and inspired by image-to-image translation, we applied DL to irregularly and regularly missing data reconstruction with the aim of transforming incomplete data into corresponding complete data. To accomplish this, we established a model architecture with randomly sampled data as input and corresponding complete data as output, which was based on an encoder-decoder-style U-Net convolutional neural network. We carefully prepared the training data using synthetic and field seismic data. We used a mean-squared-error loss function and an Adam optimizer to train the network. We displayed the feature maps for a randomly sampled data set going through the trained model with the aim of explaining how the missing data are reconstructed. We benchmarked the method on several typical datasets for irregularly missing data reconstruction, which achieved better performances compared with a peer-reviewed Fourier transform interpolation method, verifying the effectiveness, superiority, and generalization capability of our approach. Because regularly missing is a special case of irregularly missing, we successfully applied the model to regularly missing data reconstruction, although it was trained with irregularly sampled data only.Entities:
Year: 2020 PMID: 32094366 PMCID: PMC7040000 DOI: 10.1038/s41598-020-59801-x
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1Network architecture at an example of pixels in the lowest resolution. N, N, and N are the height, width, and the number of channels of the input data, respectively. Blocks show the calculated feature hierarchy. Each green box denotes multiple feature maps, and the number of feature maps (i.e., F) is marked on the right of the box. The height-width-size of a feature map is given around the box. The boxes with the same height have the same number of feature maps. The boxes with the same width indicate the same height-width-size of feature maps. The arrows and the right curly brace denote different operations. Numbers in [[⋅]] are labelled according to Table 1, which are in line with those shown in Fig. 2.
Model summary for the model architecture shown in Fig. 1, where N = 1, F1 = 64, F2 = 128, F3 = 256, F4 = 512, F5 = 1024, and KK = K × K. K = K = 5 denote the height and width of the convolution kernel, respectively. The first layer is an input layer. The layers numbered 3, 5, 8, 10, 13, 15, 18, 20, 23, 25, 29, 31, 35, 37, 41, 43, 47, and 49 are (18) activation layers. The layers numbered 6, 11, 16, and 21 are (4) max-pooling layers. The layers numbered 26, 32, 38, and 44 are (4) up-sampling layers. The layers numbered 27, 33, 39, and 45 are (4) concatenate layers. The total trainable parameters are 87,149,953.
| Layer number | Layer name | Number of trainable parameters |
|---|---|---|
| 2 | Conv01 | ( |
| 4 | Conv02 | ( |
| 7 | Conv03 | ( |
| 9 | Conv04 | ( |
| 12 | Conv05 | ( |
| 14 | Conv06 | ( |
| 17 | Conv07 | ( |
| 19 | Conv08 | ( |
| 22 | Conv09 | ( |
| 24 | Conv10 | ( |
| 28 | Conv11 | ( |
| 30 | Conv12 | ( |
| 34 | Conv13 | ( |
| 36 | Conv14 | ( |
| 40 | Conv15 | ( |
| 42 | Conv16 | ( |
| 46 | Conv17 | ( |
| 48 | Conv18 | ( |
| 50 | Conv19 | (1 × 1 × |
Figure 2Feature maps for a randomly sampled data set going through the trained model. Numbers in [[⋅]] are labelled according to Table 1, which are consistent with those shown in Fig. 1. The symbol “...” implies the omitted feature maps.
Figure 3Velocity models used to generate the synthetic data. Each velocity model has its own density model. (a) Adapted Pluto 1.5 model. (b) Down-sampled Marmousi2 model. The lateral spacing dx and the depth spacing dz are 5 m.
Figure 4Results on a synthetic training data set. (a) True data (a common-shot-point, CSP, gather). (b) Irregularly sampled data with 90% missing. (c) DL reconstruction result. (d) Difference between (a,c).
Figure 5Results on the Mobil Viking graben line 12 data set. (a) Reference data. (b) Irregularly sampled data with 60% missing. (c) Interpolated data using the FGFT interpolation method[23]. (d) DL reconstruction result.
Figure 6Model input-output pairs, which are randomly selected from the training data set. Each pair is composed of irregularly missing incomplete data (in the odd column, i.e., the model input X) and the corresponding complete data (in the even column, i.e., the model output Yref). The size of each panel is 112 × 112.
Figure 7During training, variation of with epoch.
Figure 8Results on a test data set generated with the Marmousi2 model. (a) True data. (b) Irregularly sampled data with 70% missing. (c) DL reconstruction result. (d) Difference between (a,c).
Figure 13Regularly missing data reconstruction. (a) Regularly sampled data from the synthetic training data set with a decimation factor of 10 in the space direction (90% missing). (b) DL result for (a). (c) Regularly sampled data from the Mobil Viking graben line 12 data set with a decimation factor of 3 in the space direction (66% missing). (d) DL result for (c).
Figure 9Results on a physical modelling data set[30]. (a) Reference data. (b) Irregularly sampled data with 70% missing. (c) Interpolated data using a fast-generalized Fourier transform (FGFT) method[23]. (d) DL reconstruction result.
Figure 12Results on a data set from the North Sea. (a) Reference data. (b) Irregularly sampled data with 50% missing. (c) Interpolated data using the FGFT interpolation method[23]. (d) DL reconstruction result.
Figure 10Results on a data set from the GeoFrame software. (a) Reference data. (b) Irregularly sampled data with 40% missing. (c) Interpolated data using the FGFT interpolation method[23]. (d) DL reconstruction result.