Professional Documents
Culture Documents
CLL882 Endterm 2018CH70296
CLL882 Endterm 2018CH70296
CLL882 Endterm 2018CH70296
Generation of
Knowledge Graph
for PET
Introduction
• PET is a thermoplastic polymer made from
polyester
• Excellent mechanical, chemical, and thermal
stability
• Packaging, electronics etc. commercial usage of PET
• Extensive data on the research of PET production
and recycling Need for automated techniques
• Data isolation, diffusion, and heterogeneity
• Knowledge Graph (KG) for PET poses a solution for
that
• Huge data, therefore, abstractive text
summarization and other automated techniques
required
Shi, T., Keneshloo, Y., Ramakrishnan, N., & Reddy, C. K. (2018). Neural Abstractive Text Summarization with Sequence-to-Sequence Models. ArXiv.
/abs/1812.02303
Seq2Seq Model
• Feed-forward networks, CNNs, or RNNs can be used as encoders and decoders
• The most popular RNN designs for seq2seq models are GRU and LSTM.
• In the figure, a straightforward RNN seq2seq model that makes use of a bidirectional LSTM encoder and
decoder is shown
• Since a bi-directional LSTM typically provides better document representations than a forward LSTM, it
is taken into consideration.
• The model configuration used in training the model is:
CONFIGURATION VALUE
Optimizer RMS Prop
LSTM layers 4
Epochs 50
Loss function Sparse Categorical Cross entropy
Learning rate 5x10-4
BART Model
• BART stands for bidirectional autoregressive transformer.
• A Seq2Seq model with a left-to-right autoregressive decoder and a
bidirectional encoder over tainted text.
• BART is trained by first corrupting documents and then the loss (cross
entropy) is optimized between the original document and the
decoder’s output.
• BART is pretrained on large-scale corpora using self-supervised
learning techniques.
• Well-suited for tasks like machine translation, text summarization, and
dialogue generation.
• Our architecture employs an encoder and decoder with 12 layers
each.
Lewis, Mike, et al. "BART: Denoising Sequence-to-Sequence Pre-Training for Natural Language Generation, Translation, and Comprehension." ArXiv,
2019, /abs/1910.13461. Accessed 21 Jun. 2023.
BART Model
The model configuration used in training the model:
CONFIGURATION VALUE
max_length 300
min_length 30
temperature 1
epochs 20
α 10-4
GPT Model
1. CO2 is employed as a pyrolysis CO2 is employed as a pyrolysis medium for thermal treatment of polyethylene terephthalate
medium for thermal treatment of (PET) waste to recovery of energy from the PET waste and reduce acidic byproducts such as
polyethylene terephthalate (PET) benzoic acid. The generation of carbon monoxide (CO) for pyrolysis of the PET helps to
waste to recovery of energy. energy recovery.
2. the production of gaseous the production of gaseous products containing acetylene, ethylene and carbon monoxide
products containing acetylene, from PET particles using a rotating direct current arc plasma reactor with a product gas
ethylene and carbon monoxide containing 42% acetylene, 53% carbon monox and 4% ethylene.
from PET particles using a rotating
direct current are plasma reactor.
GPT Summaries
S No. Actual Summary Predicted Summary
1. depolymerization of poly(ethylene terephthalate) Depolymerization of poly(ethylene terephthalate) in hot compressed water to form
using Sulfonic acid ionic liquids as catalyst in hot terephthalic acidic acidene terephthalate compressed waterSulfonic acid ionic liquids
compressed water to yield terephthalic acid. as catalysts in hot compressed water liquids witholy acidic ionic acid acid assts in hot
compressed waterDepolymerization of poly(ethylene terephthalate) in hot
compressed wateric acid
2. In the presenece of xylene and an emulsifier, Simultaneous glycolysis and neutral hydrolysis of waste PET wastes PET flakes
glycolysis and neutral hydrolysis of waste PET was obtained from grinding postconsumer bottles was carried out in the presence of
carried out. After separating the product from EG, xylene and an emulsifier at 190A°C Simultaneous glycolysis and neutral hydrolysis of
water and xylene by filtration, water insoluble waste PET wastes PET wastes PET from PET PET grinding postconsumer bottles was
fraction was obtained. carried out in the presence of xylene and an emulsifier at 180A°C.
Knowledge Graph
Conclusion
• BART fine tuning hasn’t performed well enough on the dataset
• The predicted summaries by BART contain some unnecessary information
• For Seq2Seq model, the test loss is converging, suggesting complete training with the
average loss coming out to be 1.6
• Seq2Seq predicted summaries contain a lot of repetitive words making to ineligible to
use for practical purposes
• The performance of GPT model is not much different from that obtained by BART
• Summaries predicted by GPT are better than BART because it does contain valuable
information, although the output is repetitive
• Finally ,the summaries predicted by GPT are used for KG creation, by simply taking
only the first non-repetitive sentence
Future Work
• The size of the dataset can be increased to a few 1000
samples.
• A more complex transformers architecture can be tested on
the current dataset to increase the model performance.
• To extend this project for other polymers too besides PET
THANKYOU!
-Praveen Soni