Professional Documents
Culture Documents
Seismic Facies Classification Using Supervised Convolutional Neural Networks and Semisupervised Generative Adversarial Networks
Seismic Facies Classification Using Supervised Convolutional Neural Networks and Semisupervised Generative Adversarial Networks
10.1190/GEO2019-0627.1
Downloaded 06/23/20 to 137.111.162.20. Redistribution subject to SEG license or copyright; see Terms of Use at http://library.seg.org/
Manuscript received by the Editor 22 September 2019; revised manuscript received 23 March 2020; published ahead of production 2 June 2020; published
online 10 June 2020.
1
Aramco Research Center — Aramco Services Company, Houston, Texas, USA and University of Wyoming, Department of Geology and Geophysics,
Laramie, Wyoming, USA. E-mail: mliu4@uwyo.edu (corresponding author).
2
EXPEC ARC, Saudi Aramco, Dhahran, Saudi Arabia. E-mail: albajervis@yahoo.com; philippe.nivlet@aramco.com.
3
Aramco Research Center — Aramco Services Company, Houston, Texas, USA. E-mail: weichang.li@aramcoservices.com.
© 2020 Society of Exploration Geophysicists. All rights reserved.
O47
O48 Liu et al.
Recently, deep learning has become an emerging subfield of challenges, we use a realistic 3D facies model and use only the in-
machine learning, benefitting from the advancement in computing formation at well locations (lithofacies upscaled from well logs and
resources, especially graphics processing units (GPUs), and access derived from core analysis) as labels to perform facies classification
to large amounts of data. This has led to a series of major break- from seismic reflection data with deep neural networks. Because the
throughs, especially in computer vision and speech recognition. The amount of well data can vary from relatively abundant in older fields
to extremely rare in new prospects, we propose two types of deep-
Downloaded 06/23/20 to 137.111.162.20. Redistribution subject to SEG license or copyright; see Terms of Use at http://library.seg.org/
performance of CNNs in seismic facies interpretation. From the overfitting. Overfitting is an undesirable behavior in which the
available drilled wells, seismic samples at well locations can be as- trained model achieves a perfect fit on the training data but does
signed with reliable facies labels derived from well logs and core not generalize well on new test data. However, it is often expensive
analysis. With sufficient samples with labels, seismic facies inter- or impractical to obtain a large number of labeled samples for train-
pretation can be formulated as a supervised classification problem. ing in seismic interpretation. In new prospects, drilled wells are
Downloaded 06/23/20 to 137.111.162.20. Redistribution subject to SEG license or copyright; see Terms of Use at http://library.seg.org/
Generally, CNNs use a set of convolutional layers to automatically rather limited. Additionally, due to class imbalance, some of the
extract the hierarchical feature maps of the input images. Each con- underrepresented facies could involve very few numbers of sam-
volutional layer consists of a stack of learnable filters that detect ples. Fortunately, 3D seismic volumes, however, provide millions
specific features present in the input data. Each convolution output to billions of unlabeled samples. It is known that the use of unla-
is passed through a typically nonlinear activation function to intro- beled data together with a small number of labeled samples can pro-
duce nonlinearity into the model. To reduce the dimension of ex- duce considerable improvement in learning accuracy by using
tracted feature maps and speed up the computation, a pooling layer, semisupervised learning. In this work, we propose a semisupervised
such as max pooling, is periodically inserted between successive method based on the generative adversarial network (GAN) (Sali-
convolutional layers. Pooling is a subsampling procedure that repla- mans et al., 2016) to deal with the challenging case of few drilled
ces the values over a certain area with their summary statistics, wells and insufficient labeled samples.
which is also helpful in making the feature representation approx- Semisupervised learning is a class of techniques that makes use
imately invariant to small translations and scale changes (Goodfel- of labeled and unlabeled data for training, typically a small amount
low et al., 2016). Then, the feature maps extracted from the of labeled data with a large amount of unlabeled data. Figure 1 il-
intermediate convolutional layer are fed into the subsequent fully lustrates how unlabeled data can help in classification. The seismic
connected layers to predict the probability of each class. input data or attributes are represented in a multivariate space, but
Training a neural network is an optimization problem, which is for the sake of simplicity, only two dimensions are shown. If we
equivalent to minimizing a predefined loss function. The loss func- only consider the labeled data shown in Figure 1a, the optimal de-
tion measures how good a model is and guides how to update the cision boundary could be a straight line (or a hyperplane) and then
model parameters during training, so it is critical to select proper the sample shown in gray would be classified as channel facies.
loss functions for different tasks. Specifically, seismic facies clas- However, if in addition to the labeled samples we also consider
sification is a typical multiclass classification problem in which the all of the unlabeled samples (the gray circles shown in Figure 1b),
multiclass cross-entropy loss (equation 1) commonly is used: it is obvious that there exists some structure to the underlying dis-
tribution of those unlabeled data, which would help us update the
1X N X K
decision boundary as shown in Figure 1c. Then, the gray sample
L¼− Iðk; yi Þ log Pðyi ¼ kjxi Þ; (1) shown in Figure 1a would be included in the nonchannel class.
N i¼1 k¼1
The semisupervised GAN is a variant of the original GAN, which
where K is the number of classes; N is the number of training sam- is specialized to deal with learning problems with very few labeled
ples; xi and yi represent the input (seismic patch or subcube) and data. The GANs are a class of generative models devised by Good-
label (facies type) of training sample i, respectively; Iðk; yi Þ is the fellow et al. (2016), and they have since become widely known for
indicator function that is equal to 1 if and only if sample i belongs to their application versatility and their often remarkable ability in gen-
class k; and Pðyi ¼ kjxi Þ is the output probability of the neural net- erating realistic data. Unlike traditional neural networks, GANs
work that sample i belongs to class k calculated from the output comprise two competing players each represented by a neural net-
layer through the softmax activation function. work: the generator and the discriminator. The generator tries to
create samples that come from a targeted, yet unknown, probability
distribution (for instance, data points from Gaussian distribution
Semisupervised generative adversarial networks
and channelized reservoir realizations) using random noise in a low-
Supervised CNNs typically have millions of trainable parameters dimensional space as the input, whereas the discriminator acts
to fit high-dimensional and nonlinear models. As a result, a tremen- like a judge to distinguish those samples produced by the generator
dous amount of labeled training samples is required to avoid from the real samples. In other words, the training objective of the
Figure 1. Illustration of semisupervised learning: (a) a possible decision boundary (dashed red line) using only one channel sample (white
circle) and one nonchannel sample (black circle), (b) the data distribution of the labeled and unlabeled samples (gray circles), and (c) the
updated decision boundary (dashed red curve) using the labeled and unlabeled samples.
O50 Liu et al.
generative network is to increase the error rate of the discriminative discarded after training. To make it applicable for semisupervised
network by producing samples that are as similar as possible to the learning, Salimans et al. (2016) transform the discriminator into a
real data, whereas it is exactly the opposite for the discriminative multiclass classifier that has dual roles in the label prediction as well
network that aims to minimize the same error rate. When trained in as in the discrimination of samples generated by the generator from
an alternative manner, the two networks would guide each other to the real data. To be more specific, the discriminator takes an input
sample and classifies it into K þ 1 classes, where K is the number of
Downloaded 06/23/20 to 137.111.162.20. Redistribution subject to SEG license or copyright; see Terms of Use at http://library.seg.org/
study, we can tune the hyperparameters to control the trade-off be- Synthetic data with CNN
tween labeled and unlabeled data.
In addition, to avoid the discriminator overconfidence in which The first example is a synthetic case of 3D lithofacies prediction
the network learns very limited features for classification and leads using supervised CNNs. The data set is generated from a realistic
to bad performance in prediction, it is desirable to add regularization synthetic model that is based on an existing on-shore field. The
3D facies model has been built by geostatistics using lithofacies
Downloaded 06/23/20 to 137.111.162.20. Redistribution subject to SEG license or copyright; see Terms of Use at http://library.seg.org/
In total, there are seven target lithofacies to be classified. Then, the To mimic the scenario of a mature field, we randomly select 256
corresponding 3D poststack seismic data are generated by 1D con- (16 × 16) traces from the synthetic geologic model as
volution with a 25 Hz Ricker wavelet of the zero-offset reflectivity pseudowells and split them into two subsets: 200 wells as a train-
derived from the P-wave velocity and density information. ing set and 56 wells as a validation set. Next, we generate labeled
samples along such pseudowells by taking
Downloaded 06/23/20 to 137.111.162.20. Redistribution subject to SEG license or copyright; see Terms of Use at http://library.seg.org/
the inference phase, the generator is discarded, and the discrimina- For a fair comparison, we also train a supervised 2D CNN on the
tor is used for the final prediction. Figure 12 shows the accuracy same training set. Given that both networks have difficulty in dis-
curves of the discriminator during training and validation. We criminating channels from levees, potentially due to the limited res-
choose the ideal point to stop training at approximately 20 epochs olution of the seismic data, the similarity of facies in terms of elastic
to minimize the possibility of overfitting. properties, and the insufficient number of total samples (labeled or
unlabeled) of both facies to adequately learn their
Downloaded 06/23/20 to 137.111.162.20. Redistribution subject to SEG license or copyright; see Terms of Use at http://library.seg.org/
Field data
Next, we apply the semisupervised GAN model to the F3 block
seismic data made publicly available by the Dutch government
through TNO and dGB Earth Sciences. Inline 339 has been inter-
preted manually with label data freely accessed from https://github.-
com/bolgebrygg/MalenoV. Unlike the definition of facies types in
the synthetic example based on lithofacies information from well
log and core analysis, in this case, nine different seismic facies are
defined according to their seismic reflection characteristics: eight
facies with distinct reflection patterns and one facies used to re-
present samples not belonging to the eight target facies (see Table 2).
Because the number and patterns of the seismic facies (i.e., width
and torsion of channels) vary with the seismic data, it is challenging
Figure 12. Training and validation accuracy. to directly transfer the trained network on the synthetic data to the
real field data. For this reason, in this case, we
reuse the network architecture with minor
changes but retrain it from scratch.
The neural network used in this study is sim-
ilar to the model in example 2 except (1) we
adopt 3D convolutional and deconvolutional
layers to build the generator and the discrimina-
tor and (2) we use a larger sample size of
64 × 64 × 64. We also alter the dimension of out-
put of the discriminator to 10 accordingly. To
generate the training set, we extract 20,000 sam-
ples from inline 339 that have been interpreted
previously (Figure 16) and from which we ran-
domly select 50 samples per facies class as la-
beled data with the remaining as unlabeled
Figure 13. Confusion matrix of the prediction on the validation set: (a) semisupervised data. For the latter, the label information is
GAN and (b) supervised CNN. masked only during training, and it is put back
Seismic facies analysis with CNN and GAN O55
and used during validation for model evaluation. After training of DISCUSSION
the networks, the generator is discarded and the discriminator is
used to perform facies prediction on the entire seismic volume. Fig- Although the proposed supervised CNN and semisupervised
ure 17a and 17b shows the classification results on inline 339 and GAN for facies classification work well on the synthetic and real
xline 775. It is evident that the predicted target facies distribution is
Downloaded 06/23/20 to 137.111.162.20. Redistribution subject to SEG license or copyright; see Terms of Use at http://library.seg.org/
very consistent with the seismic reflection characteristics. Table 2. Seismic facies defined in the F3 seismic data.
To evaluate the performance of the semisupervised GAN, we com-
pare its results to the supervised 3D CNN by using 10 and 50 labeled
samples per facies class, respectively. Figure 18 shows the t-SNE Facies number Seismic reflection characteristic
visualization of seismic samples in the validation set before being
transformed by deep neural networks, and Figure 19 shows the t- 1 Low coherency
SNE visualization of feature maps extracted from the last convolution 2 High-amplitude continuous
layers in 3D CNN and the discriminator of the GAN. We can see that 3 Low-amplitude dips
the validation samples become separable in the transformed feature 4 High-amplitude dips
space of CNN and GAN. Compared to the supervised CNN, the 5 Chaotic
semisupervised GAN can further separate the facies groups with
6 Low amplitude
the help of unlabeled data. The prediction accuracies of the semisu-
pervised GAN on the validation set with 10 and 50 labeled samples 7 High amplitude
per facies class are 85.8% and 97.5%, respectively, compared with 8 Salt
accuracies of 70.2% and 80.4% for the supervised CNN. 9 Everything else
seismic data sets, running inference on the entire 3D volume is very is to adopt lightweight networks such as MobileNet (Howard et al.,
time consuming, which would severely restrict the applications 2017), MobileNetV2 (Sandler et al., 2018), ShuffleNet (Zhang et al.,
when GPUs are not available. To solve this limitation, some at- 2018), and ShuffleNetV2 (Ma et al., 2018). In the future, we will com-
tempts have been made to convert facies classification from seismic prehensively investigate the performance of the lightweight networks
to an image segmentation problem using deep neural networks such and compare them with the standard CNNs in terms of accuracy and
Downloaded 06/23/20 to 137.111.162.20. Redistribution subject to SEG license or copyright; see Terms of Use at http://library.seg.org/
as UNet (Ronneberger et al., 2015; Di et al., 2019). Instead of point- computing time.
wise classification, those segmentation networks take a 2D seismic As discussed in the synthetic case, generally it is challenging to
section or 3D subvolume as input and are able to predict the facies discriminate the lithofacies that have similar P impedances, such as
at each pixel simultaneously. However, that usually requires com- the channel and levee, from poststack seismic data. To further im-
pletely interpreted seismic sections or subvolumes to train the segmen- prove the prediction accuracy, it would be helpful to train the net-
tation networks, which makes them not applicable for identification of works on prestack seismic data that contain more information about
rock and fluid types in which the label information is only available at rock and fluid types. Additionally, in real applications, the seismic
well locations. An alternative solution to speed up the computing time data are contaminated by noise, the lithofacies labels are derived
from well logs, and core analysis usually includes uncertainty. More
effort is necessary to address these challenges.
CONCLUSION
We have proposed two types of deep neural network frameworks
Downloaded 06/23/20 to 137.111.162.20. Redistribution subject to SEG license or copyright; see Terms of Use at http://library.seg.org/
REFERENCES
Alaudah, Y., S. Gao, and G. AlRegib, 2018, Learning
to label seismic structures with deconvolution net-
works and weak labels: 88th Annual International
Meeting, SEG, Expanded Abstracts, 2121–2125,
doi: 10.1190/segam2018-2997865.1.
Alfarraj, M., and G. AlRegib, 2019, Semisupervised
sequence modeling for elastic impedance inversion:
Interpretation, 7, SE237–SE249, doi: 10.1190/INT-
2018-0250.1.
Ao, Y., H. Li, L. Zhu, S. Ali, and Z. Yang, 2019, Iden-
tifying channel sand-body from multiple seismic
attributes with an improved random forest algorithm:
Journal of Petroleum Science and Engineering, 173,
781–792, doi: 10.1016/j.petrol.2018.10.048.
Araya-Polo, M., J. Jennings, A. Adler, and T. Dahlke,
2018, Deep-learning tomography: The Leading
Edge, 37, 58–66, doi: 10.1190/tle37010058.1.
Bond, C. E., A. D. Gibbs, Z. K. Shipton, and S. Jones,
2007, What do you think this is? “Conceptual uncer-
tainty” in geoscience interpretation: GSA Today, 17,
4–10, doi: 10.1130/GSAT01711A.1.
de Matos, M. C., P. L. Osorio, and P. R. Johann, 2006,
Unsupervised seismic facies analysis using wavelet
transform and self-organizing maps: Geophysics, 72,
no. 1, P9–P21, doi: 10.1190/1.2392789.
Di, H., D. Gao, and G. AlRegib, 2019, Developing a seis-
mic texture analysis neural network for machine-aided
seismic pattern recognition and classification: Geo-
physical Journal International, 218, 1262–1275, doi:
10.1093/gji/ggz226.
Di, H., Z. Wang, and G. AlRegib, 2018a, Seismic fault
Figure 19. A t-SNE visualization of the extracted features from (a) 3D CNN with 10 detection from post-stack amplitude by convolu-
labeled samples per class, (b) semisupervised 3D GAN with 10 labeled samples per tional neural networks: 80th Annual International
class, (c) 3D CNN with 50 labeled samples per class, and (d) semisupervised 3D Conference and Exhibition, EAGE, Extended Ab-
GAN with 50 labeled samples per class. stracts, 1–5, doi: 10.3997/2214-4609.201800733.
O58 Liu et al.
Di, H., Z. Wang, and G. AlRegib, 2018b, Deep convolutional neural net- Salimans, T., I. Goodfellow, W. Zaremba, V. Cheung, A. Radford, and X.
works for seismic salt-body delineation: AAPG Annual Convention and Chen, 2016, Improved techniques for training GANs: 30th Annual
Exhibition. Conference on Neural Information Processing Systems, 2234–2242.
Dramsch, J. S., and M. Lüthje, 2018, Deep-learning seismic facies on state- Sandler, M., A. Howard, M. Zhu, A. Zhmoginov, and L. C. Chen, 2018,
of-the-art CNN architectures: 89th Annual International Meeting, SEG, Mobilenetv2: Inverted residuals and linear bottlenecks: Proceeding of
Expanded Abstracts, 2036–2040, doi: 10.1190/segam2018-2996783.1. the 31st IEEE Conference on Computer Vision and Pattern Recognition,
Dumay, J., and F. Fournier, 1988, Multivariate statistical analyses applied to 4510–4520.
seismic facies recognition: Geophysics, 53, 1151–1159, doi: 10.1190/1 Saraswat, P., and M. K. Sen, 2012, Artificial immune-based self-organizing
Downloaded 06/23/20 to 137.111.162.20. Redistribution subject to SEG license or copyright; see Terms of Use at http://library.seg.org/
.1442554. maps for seismic facies analysis: Geophysics, 77, no. 4, O45–O53, doi: 10
Goodfellow, I., Y. Bengio, and A. Courville, 2016, Deep learning: MIT Press. .1190/geo2011-0203.1.
Guitton, A., 2018, 3D convolutional neural networks for fault interpretation: Shi, Y., X. Wu, and S. Fomel, 2019, SaltSeg: Automatic 3D salt segmenta-
80th Annual International Conference and Exhibition, EAGE, Extended tion using a deep convolutional neural network: Interpretation, 7, SE113–
Abstracts, 1–5. SE122, doi: 10.1190/INT-2018-0235.1.
Howard, A. G., M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, Simonyan, K., and A. Zisserman, 2014, Very deep convolutional networks
M. Andreetto, and H. Adam, 2017, Mobilenets: Efficient convolutional for large-scale image recognition: arXiv preprint, arXiv:1409.1556.
neural networks for mobile vision applications: arXiv preprint, ar- Strecker, U., and R. Uden, 2002, Data mining of 3D poststack seismic attrib-
Xiv:1704.04861. ute volumes using Kohonen self-organizing maps: The Leading Edge, 21,
Huang, L., X. Dong, and T. E. Clee, 2017, A scalable deep learning platform 1032–1037, doi: 10.1190/1.1518442.
for identifying geologic features from seismic attributes: The Leading van der Maaten, L. V. D., and G. Hinton, 2008, Visualizing data using t-
Edge, 36, 249–256, doi: 10.1190/tle36030249.1. SNE: Journal of Machine Learning Research, 9, 2579–2605.
Ioffe, S., and C. Szegedy, 2015, Batch normalization: Accelerating deep Waldeland, A. U., A. C. Jensen, L. J. Gelius, and A. H. S. Solberg, 2018,
network training by reducing internal covariate shift: arXiv preprint, ar- Convolutional neural networks for automated seismic interpretation: The
Xiv:1502.03167. Leading Edge, 37, 529–537, doi: 10.1190/tle37070529.1.
Kingma, D. P., and J. Ba, 2014, Adam: A method for stochastic optimiza- West, B. P., S. R. May, J. E. Eastwood, and C. Rossen, 2002, Interactive
tion: arXiv preprint, arXiv:1412.6980. seismic facies classification using textural attributes and neural networks:
Krizhevsky, A., I. Sutskever, and G. E. Hinton, 2012, Imagenet classification The Leading Edge, 21, 1042–1049, doi: 10.1190/1.1518444.
with deep convolutional neural networks: 26th Annual Conference on Wrona, T., I. Pan, R. L. Gawthorpe, and H. Fossen, 2018, Seismic facies
Neural Information Processing Systems, 1097–1105. analysis using machine learning: Geophysics, 83, no. 5, O83–O95,
LeCun, Y., B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard, doi: 10.1190/geo2017-0595.1.
and L. D. Jackel, 1989, Backpropagation applied to handwritten zip code re- Wu, X., L. Liang, Y. Shi, and S. Fomel, 2019a, FaultSeg3D: Using synthetic
cognition: Neural Computation, 1, 541–551, doi: 10.1162/neco.1989.1.4.541. data sets to train an end-to-end convolutional neural network for 3D seis-
Li, W., 2018, Classifying geological structure elements from seismic images mic fault segmentation: Geophysics, 84, no. 3, IM35–IM45, doi: 10.1190/
using deep learning: 88th Annual International Meeting, SEG, Expanded geo2018-0646.1.
Abstracts, 4643–4648, doi: 10.1190/segam2018-2998036.1. Wu, X., Y. Shi, S. Fomel, L. Liang, Q. Zhang, and A. Yusifov, 2019b, Fault-
Lubo-Robles, D., and K. J. Marfurt, 2019, Independent component analysis Net3D: Predicting fault probabilities, strikes, and dips with a single con-
for reservoir geomorphology and unsupervised seismic facies classifica- volutional neural network: IEEE Transactions on Geoscience and Remote
tion in the Taranaki Basin, New Zealand: Interpretation, 7, SE19–SE42, Sensing, 57, 9138–9155, doi: 10.1109/TGRS.2019.2925003.
doi: 10.1190/INT-2018-0109.1. Xiong, W., X. Ji, Y. Ma, Y. Wang, N. M. AlBinHassan, M. N. Ali, and Y.
Ma, N., X. Zhang, H. T. Zheng, and J. Sun, 2018, ShuffleNet V2: Practical Luo, 2018, Seismic fault detection with convolutional neural network:
guidelines for efficient CNN architecture design: 17th European Conference Geophysics, 83, no. 5, O97–O103, doi: 10.1190/geo2017-0666.1.
on Computer Vision, 116–131. Zeiler, M. D., and R. Fergus, 2014, Visualizing and understanding convolu-
Pyrcz, M. J., and C. V. Deutsch, 2014, Geostatistical reservoir modeling: tional networks: 13rd European Conference on Computer Vision, 818–
Oxford University Press. 833.
Qian, F., M. Yin, X. Y. Liu, Y. J. Wang, C. Lu, and G. M. Hu, 2018, Un- Zhang, X., X. Zhou, M. Lin, and J. Sun, 2018, Shufflenet: An extremely
supervised seismic facies analysis via deep convolutional autoencoders: efficient convolutional neural network for mobile devices: Proceeding
Geophysics, 83, no. 3, A39–A43, doi: 10.1190/geo2017-0524.1. of the 31st IEEE Conference on Computer Vision and Pattern Recogni-
Radford, A., L. Metz, and S. Chintala, 2015, Unsupervised representation tion, 6848–6856.
learning with deep convolutional generative adversarial networks: arXiv Zhao, T., 2018, Seismic facies classification using different deep convolu-
preprint, arXiv:1511.06434. tional neural networks: 88th Annual International Meeting, SEG, Ex-
Ronneberger, O., P. Fischer, and T. Brox, 2015, U-net: Convolutional net- panded Abstracts, 2046–2050, doi: 10.1190/segam2018-2997085.1.
works for biomedical image segmentation: International Conference on Zhao, T., V. Jayaram, K. J. Marfurt, and H. Zhou, 2014, Lithofacies clas-
Medical Image Computing and Computer Assisted Intervention, 234–241. sification in Barnett Shale using proximal support vector machines: 84th
Roy, A., A. S. Romero-Peláez, T. J. Kwiatkowski, and K. J. Marfurt, 2014, Annual International Meeting, SEG, Expanded Abstracts, 1491–1495,
Generative topographic mapping for seismic facies estimation of a carbon- doi: 10.1190/segam2014-1210.1.
ate wash, Veracruz Basin, southern Mexico: Interpretation, 2, SA31–SA47, Zhao, T., F. Li, and K. J. Marfurt, 2017, Constraining self-organizing map
doi: 10.1190/INT-2013-0077.1. facies analysis with stratigraphy: An approach to increase the credibility
Saggaf, M. M., M. N. Toksöz, and M. I. Marhoon, 2003, Seismic in automatic seismic facies classification: Interpretation, 5, T163–T171,
facies classification and identification by competitive neural networks: doi: 10.1190/INT-2016-0132.1.
Geophysics, 68, 1984–1999, doi: 10.1190/1.1635052.