Knowledge Discovery in Big Data From Astronomy and Earth Observation: Astrogeoinformatics 1St Edition Petr Skoda (Editor) - Ebook PDF

You might also like

Download as pdf or txt
Download as pdf or txt
You are on page 1of 69

Knowledge Discovery in Big Data from

Astronomy and Earth Observation:


Astrogeoinformatics 1st Edition Petr
Skoda (Editor) - eBook PDF
Go to download the full and correct content document:
https://ebooksecure.com/download/knowledge-discovery-in-big-data-from-astronomy-
and-earth-observation-astrogeoinformatics-ebook-pdf/
More products digital (pdf, epub, mobi) instant
download maybe you interests ...

Big Data in Astronomy: Scientific Data Processing for


Advanced Radio Telescopes 1st Edition Linghe Kong
(Editor) - eBook PDF

https://ebooksecure.com/download/big-data-in-astronomy-
scientific-data-processing-for-advanced-radio-telescopes-ebook-
pdf/

Multi-Scale Approaches in Drug Discovery. From


Empirical Knowledge to In Silico Experiments and Back
1st Edition Edition Alejandro Speck-Planche (Eds.) -
eBook PDF
https://ebooksecure.com/download/multi-scale-approaches-in-drug-
discovery-from-empirical-knowledge-to-in-silico-experiments-and-
back-ebook-pdf/

Big Data Application in Power Systems 1st Edition -


eBook PDF

https://ebooksecure.com/download/big-data-application-in-power-
systems-ebook-pdf/

Process Safety and Big Data 1st Edition - eBook PDF

https://ebooksecure.com/download/process-safety-and-big-data-
ebook-pdf/
Cognitive Big Data Intelligence with a Metaheuristic
Approach (Cognitive Data Science in Sustainable
Computing) 1st Edition - eBook PDF

https://ebooksecure.com/download/cognitive-big-data-intelligence-
with-a-metaheuristic-approach-cognitive-data-science-in-
sustainable-computing-ebook-pdf/

Big Data Mining for Climate Change 1st edition - eBook


PDF

https://ebooksecure.com/download/big-data-mining-for-climate-
change-ebook-pdf/

Progress in Heterocyclic Chemistry Volume 29 1st


Edition - eBook PDF

https://ebooksecure.com/download/progress-in-heterocyclic-
chemistry-ebook-pdf/

Big Data and Mobility as a Service 1st Edition Haoran


Zhang - eBook PDF

https://ebooksecure.com/download/big-data-and-mobility-as-a-
service-ebook-pdf/

Discovery and Development of Antidiabetic Agents from


Natural Products: Natural Product Drug Discovery -
eBook PDF

https://ebooksecure.com/download/discovery-and-development-of-
antidiabetic-agents-from-natural-products-natural-product-drug-
discovery-ebook-pdf/
Knowledge
Discovery in Big
Data from
Astronomy and
Earth Observation
Knowledge
Discovery in Big
Data from
Astronomy and
Earth Observation
AstroGeoInformatics
Edited by
Petr Škoda
Stellar Department
Astronomical Institute of the Czech Academy of Sciences
Ondřejov, Czech Republic

Fathalrahman Adam
Earth Observation Center
German Remote Sensing Data Center
DLR German Aerospace Center
Wessling, Germany
Elsevier
3251 Riverport Lane
St. Louis, Missouri 63043

Knowledge Discovery in Big Data from Astronomy and Earth Observation ISBN: 978-0-12-819154-5
Copyright © 2020 Elsevier Inc. All rights reserved.

MATLAB® is a trademark of The MathWorks, Inc. and is used with permission.


The MathWorks does not warrant the accuracy of the text or exercises in this book.
This book’s use or discussion of MATLAB® software or related products does not constitute endorsement or sponsorship by The Math-
Works of a particular pedagogical approach or particular use of the MATLAB® software.
No part of this publication may be reproduced or transmitted in any form or by any means, electronic or mechanical, including photo-
copying, recording, or any information storage and retrieval system, without permission in writing from the publisher. Details on how
to seek permission, further information about the Publisher’s permissions policies and our arrangements with organizations such as the
Copyright Clearance Center and the Copyright Licensing Agency, can be found at our website: www.elsevier.com/permissions.
This book and the individual contributions contained in it are protected under copyright by the Publisher (other than as may be noted
herein).

Notices

Practitioners and researchers must always rely on their own experience and knowledge in evaluating and using any information,
methods, compounds or experiments described herein. Because of rapid advances in the medical sciences, in particular, independent
verification of diagnoses and drug dosages should be made. To the fullest extent of the law, no responsibility is assumed by Elsevier,
authors, editors or contributors for any injury and/or damage to persons or property as a matter of products liability, negligence or
otherwise, or from any use or operation of any methods, products, instructions, or ideas contained in the material herein.

Publisher: Candice Janco


Acquisitions Editor: Marisa LaFleur
Editorial Project Manager: Andrea Dulberger
Production Project Manager: Sreejith Viswanathan
Designer: Alan Studholme
Contents

L I S T O F C O N T R I B U T O R S, vii 6 Surveys, Catalogues,


A WO R D FR O M TH E B I G - S K Y- E A RTH Databases/Archives, and
C H A I R, xi State-of-the-Art Methods for
P R E F A C E, xiii Geoscience Data Processing, 103
A C K N O W L E D G M E N T S, xvii Lachezar Filchev, Lyubka Pashova,
Vasil Kolev, Stuart Frye
PART I
D ATA 7 High-Performance Techniques for
Big Data Processing, 137
1 Methodologies for Knowledge Philipp Neumann, Julian Kunkel
Discovery Processes in Context of
AstroGeoInformatics, 1
8 Query Processing and Access
Peter Butka, Peter Bednár,
Methods for Big Astro and Geo
Juliana Ivančáková
Databases, 159
Karine Zeitouni, Mariem Brahem,
2 Historical Background of Big Data Laurent Yeh, Atanas Hristov
in Astro and Geo Context, 21
Christian Muller 9 Real-Time Stream Processing
in Astronomy, 173
PART I I Veljko Vujčić, Darko Jevremović
I N F O R M ATI O N
PART I I I
3 AstroGeoInformatics: From KNOWLEDGE
Data Acquisition to Further
Application, 31 10 Time Series, 183
Bianca Schoen-Phelan Ashish Mahabal

4 Synergy in Astronomy and 11 Advanced Time Series Analysis


Geosciences, 39 of Generally Irregularly Spaced
Mikhail Minin, Angelo Pio Rossi Signals: Beyond the Oversimplified
Methods, 191
Ivan L. Andronov
5 Surveys, Catalogues, Databases,
and Archives of Astronomical
Data, 57 12 Learning in Big Data: Introduction
Irina Vavilova, Ludmila Pakuliak, Iurii Babyk, to Machine Learning, 225
Andrii Elyiv, Daria Dobrycheva, Olga Melnyk Khadija El Bouchefry, Rafael S. de Souza

v
vi CONTENTS

13 Deep Learning – an Opportunity 20 International Database of


and a Challenge for Geo- and Neutron Monitor Measurements:
Astrophysics, 251 Development and Applications, 371
Christian Reimers, Christian Requena-Mesa D. Sapundjiev, T. Verhulst, S. Stankov

14 Astro- and Geoinformatics – 21 Monitoring the Earth Ionosphere by


Visually Guided Classification of Listening to GPS Satellites, 385
Time Series Data, 267 Liubov Yankiv-Vitkovska, Stepan Savchuk
Roman Kern, Tarek Al-Ubaidi, Vedran Sabol,
Sarah Krebs, Maxim Khodachenko, 22 Exploitation of Big Real-Time
Manuel Scherf GNSS Databases for Weather
Prediction, 405
15 When Evolutionary Nataliya Kablak, Stepan Savchuk
Computing Meets Astro- and
Geoinformatics, 283 23 Application of Databases Collected
Zaineb Chelly Dagdia, Miroslav Mirchev in Ionospheric Observations by
VLF/LF Radio Signals, 419
PART I V Aleksandra Nina
WISDOM
24 Influence on Life Applications
16 Multiwavelength Extragalactic of a Federated Astro-Geo
Surveys: Examples of Data Database, 435
Mining, 307 Christian Muller
Irina Vavilova, Daria Dobrycheva,
Maksym Vasylenko, Andrii Elyiv, Olga Melnyk
I N D E X, 445

17 Applications of Big Data in


Astronomy and Geosciences:
Algorithms for Photographic
Images Processing and Error
Elimination, 325
Ludmila Pakuliak, Vitaly Andruk

18 Big Astronomical Datasets and


Discovery of New Celestial Bodies
in the Solar System in Automated
Mode by the CoLiTec Software, 331
Sergii Khlamov, Vadym Savanevych

19 Big Data for the Magnetic Field


Variations in Solar-Terrestrial
Physics and Their Wavelet
Analysis, 347
Bozhidar Srebrov, Ognyan Kounchev,
Georgi Simeonov
List of Contributors

Tarek Al-Ubaidi, MSc Andrii Elyiv, Dr


DCCS – IT Business Solutions, Graz, Austria Main Astronomical Observatory of the National
Academy of Sciences of Ukraine, Kyiv, Ukraine
Ivan L. Andronov, DSc, Prof
Department of Mathematics, Physics and Astronomy, Lachezar Filchev, Assoc Prof, PhD
Odessa National Maritime University, Odessa, Ukraine Space Research and Technology Institute, Bulgarian
Academy of Sciences, Sofia, Bulgaria
Vitaly Andruk
Main Astronomical Observatory of the National Stuart Frye, MSc
Academy of Sciences of Ukraine, Kyiv, Ukraine National Aeronautics and Space Administration,
Washington, DC, United States
Iurii Babyk, Dr
Atanas Hristov, PhD
Main Astronomical Observatory of the National
Academy of Sciences of Ukraine, Kyiv, Ukraine University of Information Science and Technology
“St. Paul the Apostle”, Ohrid, North Macedonia
Peter Bednár, PhD
Juliana Ivančáková, MSc
Department of Cybernetics and Artificial Intelligence,
Department of Cybernetics and Artificial Intelligence,
Technical University of Košice, Košice, Slovakia
Technical University of Košice, Košice, Slovakia
Mariem Brahem, PhD
Darko Jevremović, Dr
DAVID Lab., University of Versailles
Astronomical Observatory Belgrade, Belgrade, Serbia
Saint-Quentin-en-Yvelines, Université Paris-Saclay,
Versailles, France Nataliya Kablak, DSc, Prof
Uzhhorod National University, Uzhhorod, Ukraine
Peter Butka, PhD
Department of Cybernetics and Artificial Intelligence, Roman Kern, PhD
Technical University of Košice, Košice, Slovakia Institute of Interactive Systems and Data Science,
Technical University of Graz, Graz, Austria
Zaineb Chelly Dagdia, Dr
Université de Lorraine, CNRS, Inria, LORIA, Sergii Khlamov, PhD
F-54000 Nancy, France Institute of Astronomy, V. N. Karazin Kharkiv National
LARODEC, Institut Supérieur de Gestion de Tunis, University, Kharkiv, Ukraine
Tunis, Tunisia Main Astronomical Observatory of the NAS of Ukraine,
Kyiv, Ukraine
Rafael S. de Souza, PhD
Maxim Khodachenko, PhD
Department of Physics & Astronomy, University of North
Carolina at Chapel Hill, Chapel Hill, NC, United States Space Research Institute, Austrian Academy of
Sciences, Graz, Austria
Daria Dobrycheva, Dr Skobeltsyn Institute of Nuclear Physics, Moscow State
Main Astronomical Observatory of the National University, Moscow, Russia
Academy of Sciences of Ukraine, Kyiv, Ukraine
Vasil Kolev, MSc
Khadija El Bouchefry, PhD Institute of Information and Communication
South African Radio Astronomy Observatory, Technologies, Bulgarian Academy of Sciences, Sofia,
Rosebank, JHB, South Africa Bulgaria

vii
viii LIST OF CONTRIBUTORS

Ognyan Kounchev, Prof, Dr Christian Requena-Mesa, MSc


Institute of Mathematics and Informatics, Bulgarian Computer Vision Group, Faculty of Mathematics and
Academy of Sciences, Sofia, Bulgaria Computer Science, Friedrich Schiller University Jena,
Jena, Germany
Sarah Krebs, MSc
Climate Informatics Group, Institute of Data Science,
Know-Center, Graz, Austria
German Aerospace Center, Jena, Germany
Julian Kunkel, Dr Department Biogeochemical Integration, Max Planck
University of Reading, Reading, United Kingdom Institute for Biogeochemistry, Jena, Germany
Ashish Mahabal, PhD Angelo Pio Rossi, Dr, PhD
California Institute of Technology, Pasadena, CA, Jacobs University Bremen, Bremen, Germany
United States
Vedran Sabol, PhD
Olga Melnyk, Dr Know-Center, Graz, Austria
Main Astronomical Observatory of the National
D. Sapundjiev, PhD
Academy of Sciences of Ukraine, Kyiv, Ukraine
Royal Meteorological Institute (RMI), Brussels, Belgium
Mikhail Minin, MSc
Vadym Savanevych, DSc
Jacobs University Bremen, Bremen, Germany
Main Astronomical Observatory of the NAS of Ukraine,
Miroslav Mirchev, Dr Kyiv, Ukraine
Faculty of Computer Science and Engineering,
Stepan Savchuk, DSc, Prof
Ss. Cyril and Methodius University in Skopje, Skopje,
North Macedonia Lviv Polytechnic National University, Lviv, Ukraine

Christian Muller, Dr Manuel Scherf, MSc


Royal Belgian Institute for Space Aeronomy, Space Research Institute, Austrian Academy of
Belgian Users Support and Operation Centre, Sciences, Graz, Austria
Brussels, Belgium Bianca Schoen-Phelan, PhD
Philipp Neumann, Prof, Dr Technical University Dublin, School of Computer
Helmut-Schmidt-Universität Hamburg, Hamburg, Science, Dublin, Ireland
Germany Georgi Simeonov, Assistant
Aleksandra Nina, PhD Institute of Mathematics and Informatics, Bulgarian
Institute of Physics Belgrade, University of Belgrade, Academy of Sciences, Sofia, Bulgaria
Belgrade, Serbia Bozhidar Srebrov, Assoc Prof, Dr
Ludmila Pakuliak, Dr Institute of Mathematics and Informatics, Bulgarian
Main Astronomical Observatory of the National Academy of Sciences, Sofia, Bulgaria
Academy of Sciences of Ukraine, Kyiv, Ukraine S. Stankov, PhD
Lyubka Pashova, Assoc Prof, PhD Royal Meteorological Institute (RMI), Brussels, Belgium
National Institute of Geophysics, Geodesy and Maksym Vasylenko, MSc
Geography, Bulgarian Academy of Sciences, Sofia,
Main Astronomical Observatory of the National
Bulgaria
Academy of Sciences of Ukraine, Kyiv, Ukraine
Christian Reimers, MSc
Irina Vavilova, Dr
Computer Vision Group, Faculty of Mathematics and
Main Astronomical Observatory of the National
Computer Science, Friedrich Schiller University Jena,
Academy of Sciences of Ukraine, Kyiv, Ukraine
Jena, Germany
Climate Informatics Group, Institute of Data Science, T. Verhulst, PhD
German Aerospace Center, Jena, Germany Royal Meteorological Institute (RMI), Brussels, Belgium
LIST OF CONTRIBUTORS ix

Veljko Vujčić Karine Zeitouni, Prof, PhD


Astronomical Observatory Belgrade, Belgrade, Serbia DAVID Lab., University of Versailles
Saint-Quentin-en-Yvelines, Université Paris-Saclay,
Liubov Yankiv-Vitkovska, Assoc Prof, Dr
Versailles, France
Lviv Polytechnic National University, Lviv, Ukraine
Laurent Yeh, Assoc Prof, PhD
DAVID Lab., University of Versailles
Saint-Quentin-en-Yvelines, Université Paris-Saclay,
Versailles, France
A Word from the BIG-SKY-EARTH Chair

In the summer of 2013, a small group of astronomers ers. The book was a challenge for authors to write, since
and Earth observation experts gathered around an in- the readers will surely come with a large range of back-
triguing idea – why not assemble a network of experts ground knowledge, and texts had to be balanced be-
from those two disciplines that have so much in com- tween the disciplines. But everyone should find at least
mon, but essentially do not communicate much as pro- some parts of the book intriguing and interesting, hope-
fessionals? An opportunity opened for this idea to be fully as much as we found our BIG-SKY-EARTH network
realized as the COST funding scheme opened a call inspiring and motivating.
for transdisciplinary networks. A proposal was put to-
gether and the Big Data Era in Sky and Earth Observa- Dejan Vinković
tion (BIG-SKY-EARTH) project was born.1 Funded by Chair of the BIG-SKY-EARTH network
COST, the project officially started in January 2015 and
ended in January 2019. In the end, the BIG-SKY-EARTH
network included 28 COST countries and hundreds of
researchers from all around the world. The project or-
ganized four training schools, two workshops and two
This book is based upon work from COST Action TD1403
conferences, exchange of numerous scientists between
- BIG-SKY-EARTH, supported by the European Cooperation
research groups, many working group meetings, and
in Science and Technology (COST).
formal and informal gatherings. The lively exchange of
COST2 is a funding agency for research and innovation
ideas and a mixture of people of various expertise fueled
networks. Our actions help connect research initiatives across
a plethora of collaborations between participants. It has
Europe and enable scientists to grow their ideas by sharing
been a mixture of experts from academia and the private
them with their peers. This boosts their research, career, and
sector, with special attention given to supporting young
innovation.
researchers still in the early stages of their careers.
Toward the end of the COST project, a suggestion
emerged to put together a book that would be an in-
teresting reading for astronomers curious about remote
sensing and vice versa. Essentially, the book would be
about AstroGeoInformatics – an amalgam of computer
science, astronomy, and Earth observation. The sugges-
tion was enthusiastically supported by lots of partici-
pants and now, finally, you have it in front of you –
the first book written for such transdisciplinary read-

1 https://bigskyearth.eu. 2 www.cost.eu.

xi
Preface

WHAT’S IN THIS BOOK? MOTIVATION AND SCOPE


This book has several parts reflecting various stages of Fundamental science at the beginning of our civiliza-
Big Data processing and machine learning following the tion was naturally treated as a multidisciplinary task.
Data–Information–Knowledge–Wisdom (DIKW) pyra- Most important ancient Greek scientists and philoso-
mid explained below. There are several main parts re- phers focused on mathematics, physics, astronomy, ge-
flecting the particular stages of the pyramid. ometry, biology, mineralogy, meteorology, and medi-
cine, as well as on social and political matters. The
Part I is an introductory section about the origin of
discovery of nature per se was primarily driven by cu-
Big Data and its history. It shows that Big Data is not an
riosity and not by the intention of exploiting it. Par-
entirely new concept, as humans have tended for a long ticularly famous are many philosophical discussions in
time to collect ever-growing amounts of data, and sum- the Pythagorean school or the Platonic Academy. The
marizes the general principles and recent developments principles of “Pansophism” as an educational goal were
of knowledge discovery. proclaimed by Comenius and the didactic principles he
Part II discusses the different stages of data acquisi- had introduced are probably the top achievement in
tion, preprocessing, and interpretation and data repos- teaching that seem to continue almost unchanged until
itories. We want to point out what basic steps have to these days. Our goal is to maintain some of his princi-
be made to reach end-to-end systems allowing an au- ples in this book, as well. That is why we put emphasis
tomated extraction of results based on instrument data on the understandability and on the joy of gaining new
and auxiliary databases. knowledge together with using (some) practical exam-
Part III addresses the primary goal of this book – ples. We also accentuate (in agreement with Comenius)
the role of color pictures that try to associate the read
the extraction of new wisdom about the Universe (in-
text with something familiar to the reader.
cluding our Earth) from Big Data. It presents various
The problem with current science is mainly in break-
approaches of data analysis, such as genetic program- ing with all the principles given above. The contem-
ming and machine learning techniques, including the porary scientific communities are very narrow-focused,
commercially overhyped deep learning. having their own terminology which presents a high en-
Part IV addresses the specific readership that we want try barrier for the intruders into their “sacred land.” An
to reach. We think of a wide range of experts ranging enthusiastic researcher searching similarities between
from students of different disciplines to practitioners his research and a completely different field of science
as well as theory-oriented researchers. Here, our aim is is quickly redirected into the “proper” corridors by a
to acquaint these readers with the wide variety of tasks number of ways (e.g., by funding agencies, referees of
that can be solved by modern approaches of Big Data scientific journals, or leaders of their home institution).
processing and give them some examples of an interdis- The result is that interdisciplinarity, although strong-
ciplinary approach using smart untraditional methods ly proclaimed by science policymakers, is very difficult
to practice in reality. Most researchers hardly get suffi-
and data sources. The advanced combination of appar-
cient funds to visit essential conferences and workshops
ently nonrelevant data may even help reveal unexpected
in their own narrow field, and here they usually meet
correlations and relationships having an impact on our the same colleagues they already know quite well, and
everyday life. So, for instance, the real-time monitoring everybody knows what topics (and even which objects)
of GPS signals helps in predicting the weather, and lis- will be presented. The big symposia are more promis-
tening to very long-wave transmitters can be crucial in ing to promote interdisciplinarity at least within the
predictions of natural disasters such as hurricanes or boundaries of one branch of science (e.g., astronomy,
earthquakes. space physics, and geophysics). However, the multitude

xiii
xiv PREFACE

of parallel sessions and the shortage of available time A typical astronomical example are old astronomical
enforces the attendees again to visit only sessions about plates with spectra of bright hot stars (secured at the
their own and closely related fields. So it seems that beginning of the 20th century) to measure ozone con-
many scientists remain locked in their small, highly spe- centrations in the Earth’s stratosphere at that time.
cialized communities. Fortunately, on the other hand, In the future, we expect that more complex scien-
there are also quite large meetings not based on lim- tific problems will lead to common “astro-/geo-” ap-
ited research subjects but particular methodologies and proaches to be applied in synergy. There are already
technology. several cases where astronomical instrumentation was
So, for instance, in astronomy, there is a tradi- used for analyzing terrestrial phenomena. For example,
tional Astronomical Data Analysis Software and Sys- the Earth’s aurora X-ray radiation was observed in 2015
tems (ADASS), conference where the common uniting by ESA Integral mission, ordinarily busy with observing
subject is modern computer technology in fields such as black holes, during calibration of the diffuse cosmic X-
satellite and telescope control systems, advanced math- ray background. Another example is the world’s largest
ematical algorithms, high-performance computing and radio telescope array LOFAR, which, when monitoring
databases, software development methodology, or the the three-dimensional propagation of a lightning 18 km
analysis of Big Data. The same holds for Earth obser- above the Netherlands with microsecond time resolu-
vation, where the International Geoscience and Remote tion, discovered interesting needle-shaped plasma struc-
Sensing Symposium (IGARSS) conferences are targeted tures, probably responsible for multiple strikes of the
to participants from numerous geoscientific and remote same lightning within seconds.
sensing disciplines. The astronomical observations also contribute to
Another example of a broad community from all meteorology and vice versa. In astronomical spec-
fields of astronomy are interoperability meetings of
troscopy, almost every spectrum is contaminated by
the International Virtual Observatory Alliance, where
atmospheric water vapor absorption lines, which pro-
the main goal is to define data formats, models, and
vides a direct way to measure the line-of-sight water
protocols allowing the global interoperability of all as-
content in the atmosphere. This water contamination
tronomical databases and archives. Their “terrestrial”
must be removed from the recorded spectra by compli-
counterpart are lots of standards and conventions pub-
cated methods, although in many cases, it is used to
lished by the Open Geospatial Consortium (OGC),
correct the wavelength calibration of the spectrograph
national and international space agencies (e.g., NASA,
as a reference source with negligible radial velocity in
ESA, and DLR), and even industrial consortia that are
comparison to stellar ones.
highly specialized in image processing, spectroscopy,
precision farming, database technologies, computer Current astronomy is making many new discoveries
networks, etc. thanks to the possibility of aggregating the observations
Nowadays, we can see that more and more quantita- in the whole electromagnetic spectrum (with recent ex-
tive physical and chemical parameters can be retrieved tensions to astroparticle physics – cosmic rays, neutri-
from advanced instruments and their processing chains; nos, and also gravitational wave astronomy) and also
a typical criterion is the attainable signal-to-noise level by cross-matching of gigantic multi-petabyte scaled sky
of an instrument. Here we can see considerable im- surveys. The new astronomical instruments LSST, SKA,
provements when we compare traditional instruments and EUCLID foresee to produce tens of petabytes of raw
with innovative devices. However, we must also con- data every year, and their future archives are expected to
sider the specific requirements of astro- and geoscience be operating at the edge of contemporary IT technol-
applications such as the long-term data availability and ogy, with the embedded state-of-the-art machine learn-
traceability of the results (data curation, reproducibil- ing functionality.
ity), proven scientific data quality with quantitative er- As for geosciences, we have to consider additional
ror levels, and appropriate tools to validate them. The application tasks with specific constraints, such as real-
importance of archives of historical records, namely, time traffic monitoring and rapid disaster analysis (e.g.,
photographic plates, is rising in accordance with the im- of flooding, lava flows, storm warnings, and volcanic
plementation of FAIR principles of data management plume dynamics) or the handling of large data vol-
(striving to make data findable, accessible, interopera- umes that we encounter in weather predictions or crop
ble, and reusable). Digitization of legacy sources brings yield calculations. In particular, some applications call
about a considerable amount of Big Data volumes but for highly reliable and carefully calibrated input data,
also new, unexpected opportunities of their analysis. for instance, for climate research and risk assessments,
PREFACE xv

while other forms put more emphasis on keeping up statistics and astroinformatics of both the International
with the data volume or access to cloud computing. Astronomical Union and the American Astronomical
The geosciences benefit a lot from satellite observa- Society, and also the recently established International
tions and remote sensing, where the continuous flow of AstroInformatics Association. In contrast, Geoinformat-
new hyperspectral data, motion and displacement mea- ics for Earth observation suffers from the misunder-
surements (based on speed and phase measurements standing of being confounded with geomatics and ge-
derived from synthetic-aperture radars or optical instru- ographical information systems; however, the impor-
ments) have to be combined with accurate “ground tance of machine learning and Big Data processing is
truth” reference data obtained by geophysics, hydro- already well understood and it is applied routinely.
logy, and terrain investigations. This presents challeng- Despite the lack of a proper taxonomy for these
ing opportunities for agriculture, forestry, water resource newly established fields, there are attempts to transfer
planning, and ore mining as well as new approaches some methods that were successfully applied in one
to infrastructure development and urban planning. The scientific field into a completely different one. Within
importance of aggregated Earth observation databases this context, a key methodology for transferring digi-
and so-called Virtual Earths is continuously increasing tized knowledge is transfer learning, allowing us to train
in the case of natural disasters for rescue and humani- a machine learning system in one domain and then
tarian aid purposes. apply the fully trained system within a completely dif-
The current Fourth Paradigm of science, character- ferent application area of data analytics. However, this
ized by Big Data, is data-driven research and exploratory approach calls for an appropriate design of the required
analysis exploiting innovative machine learning princi- databases. Currently, the first results could already be
ples to obtain new knowledge, e.g., about a vast tsunami demonstrated for image content classification in remote
threatening a coastal area, thus offering a qualitatively sensing applications.
new way of scientific methodologies, e.g., by simulation There are already world-renowned institutions where
runs and comparisons with existing databases. The Big such research is being conducted. For instance, the
Data phenomenon is common for every scientific dis- Center for Data-Driven Discovery at Caltech and some
cipline and presents a serious obstacle to conducting groups in NASA successfully applied methods devel-
efficient research by established methods. Therefore, a oped for astronomical analysis in medicine (e.g., pre-
new multidisciplinary approach is needed, which in- diction of autism in EEG records or the identification of
tegrates the particular discipline knowledge with ad- cancer metastases in histologic samples.).
vanced statistics, high-performance computing and data Another field where the methodological similarities
analytics, new computer software design principles, ma- with astronomy are even more striking is the complete
chine learning and other fields belonging rather to com- range of geosciences and remote sensing disciplines. Im-
puter science, artificial intelligence, and data science. age analysis using data recorded by CCD detectors, mul-
This change also requires a new kind of a career path tispectral analysis, hyperdata cubes, time series mea-
which is not yet fully established despite first exper- surements, data streams, various coordinate systems,
imental courses at renowned universities, where the deep learning, classification techniques, and federaliza-
graduates are “data scientists.” tion of resources – all these fields are applied in similar
The same term is used in business for the “sexiest ways both in astronomy and geosciences. So it seems
job of the 21st century,” and is understood rather as to be very useful for one community to learn from
a commercial data analysis task – using statistics and the other one, and all of them should acquire a lot of
simple machine learning for the analysis of data ware- practical skills from computer science and data science.
houses. The real scientific data scientist must, however, We are convinced that both communities (“astro” and
have a considerably broader knowledge of his scientific “geo”) will benefit from understanding the more com-
branch, in addition to excellent knowledge of statistics prehensive interdisciplinary view. An interesting topic
and machine learning. could be advanced algorithms to identify (and remove)
The first successful integration happened in bioin- the varying atmospheric background of satellite images.
formatics, which recently became an officially accepted The idea of this book originated during very produc-
branch of science. In astronomy, the Astroinformatics tive meetings of the transdisciplinary European COST
originated around the year 2010 and is still being treated Action TD1403 called BIG-SKY-EARTH. It would not
by the majority of astronomers like a kind of sorcery, be possible without the great enthusiasm of many peo-
despite the public outreach effort of the International ple who devoted a considerable amount of time to the
Astrostatistics Association, the working groups in astro- preparation of the state-of-the-art reviews of typical top-
xvi PREFACE

ics that, according to their feelings, will be important for As we are aware that there are limits in understand-
future data scientists. ing the terminologies of other fields, we asked all au-
The main goal of the book, which is a kind of ex- thors to avoid complex mathematics as well as deep
periment trying to show the potential synergy between details and the common jargon of each discipline, and
astronomy and geosciences, is to give some first-time try to treat their contribution more on an educational or
overview of the technologies and methodologies to- outreach level (also by using more illustrations) while
gether with references to the practical usage of many im- still maintaining a rigorous structure with proper refer-
portant fields related to knowledge discovery in astro- ences for every important statement. Thus, we can avoid
and geo-Big Data. Its purpose is to give a very general detailed discussions about very specific problems such
background and some ideas with numerous references as the avoidance of overfitting during image classifica-
that could be helpful in the design of fully operational tion; however, we have to be aware of typical decision
data analysis systems – and of an experimental data sci- making problems and imminent limitations: Shall we
ence book. believe in earthquake prediction based on deep learn-
The book tries to follow the pyramid called Data ing and evacuate a full region?
(from data acquisition to distributed databases), We hope that a wide scope of readers will find this
Information (data processing), Knowledge (machine book interesting, and that it will serve them as a starter
learning), and Wisdom (applied knowledge for appli- for an interdisciplinary way of individual thinking. This
cations) (DIKW). We intend to present a global picture should be an important characteristic of this book. The
seen from history, data gathering, data processing and future of humankind is dependent on a close collabora-
knowledge extraction to inferences of new wisdom. This tion between many scientific disciplines in synergy. Big
has to be understood in conjunction with the pros and Data is one example of global problems that must be
cons of every processing step and its concatenations (in- overcome by changes of paradigms of how the research
cluding user interfaces). was done in each discipline so far.
In the last part of the book, we address a funda- Data scientists will be indispensable leaders of these
mental question, namely, what kind of new knowledge changes. We hope that our book will help educate new
we could get if the data from both astronomical and graduates in this emerging field of science.
geo-research will be properly processed in close synergy,
and what impact this will have on human health, envi- Petr Škoda
ronmental problems, economic prosperity, etc. In the Fathalrahman Adam
following chapters, we present somewhat preliminary Editors
topics of a newly emerging scientific field which we try
to define as AstroGeoInformatics. With great help from Gottfried Schwarz, DLR.
Acknowledgments

We want to thank all the people who participated in interdisciplinary collaboration that tries to present the
the preparation of this book. The book is not just an- benefits of synergetic empowerment in natural sciences.
other collection of papers like typical conference pro- We are also indebted to Elsevier’s representatives,
ceedings, but the result of long-term planning, inspiring who were in direct contact with other authors and us,
discussions with experts in many fields of natural sci- editors, during the whole process of book preparation.
ences, asking personally more than fifty people we knew It was Marisa LaFleur who followed the book project
from various conferences, making open calls in different from its start and managed to arrange things in a com-
discussion groups and communities, and exchanging a fortable but determined way. Lots of people, including a
large number of e-mails with the world’s leaders in re- part of the referees during the initial review phase of the
lated fields. So in addition to the authors who finally book’s table of contents, did not believe in our vision of
mixing astronomy, geosciences, and computer science
wrote some chapters, we also thank those who estab-
in all chapters. They would have preferred the more clas-
lished contacts with potential authors or contributed by
sic approach – to split the book into two parts – one for
providing links to interesting articles, as well as those
astronomers and another second for geo-experts. De-
who have enthusiastically promised writing the chap- spite their skepticism, the Elsevier people trusted the
ter, but later were not able to do this due to more urgent strength of our vision and helped us finish the work suc-
matters. cessfully.
We are also grateful to the European Union for its So we thank Ashwathi Aravindakshan, our copy-
funding of the COST Action TD1403 BIG-SKY-EARTH, rights coordinator, who was keeping an eye on the
which succeeded in attracting such a nice group of ex- proper copyright status of each figure, as well as the
perts that benefited from the interdisciplinary nature tables and examples which were already published else-
of this action. This COST action has also shown the where, and Subramaniam Jaganathan, the contract co-
need of personal contacts in preparing exciting research ordinator, who helped us arrange the initial paperwork
ideas. The fruitful COST meetings helped amalgamate for signing the contract.
the initially heterogeneous group of researchers from Finally, we would like to express our deep gratitude
different fields and countries into a real task force ca- to Andrea Dulberger, the editorial project manager, as
pable of accepting each other’s visions, methodologies, well as Sreejith Viswanathan, the project manager, for
and technologies. This book is the result of such a new leading the project to a successful end.

xvii
PART I DATA

CHAPTER 1

Methodologies for Knowledge


Discovery Processes in Context of
AstroGeoInformatics
PETER BUTKA, PHD • PETER BEDNÁR, PHD • JULIANA IVANČÁKOVÁ, MSC

1.1 INTRODUCTION For a better understanding of KDPs, we can shortly


Whenever someone wants to apply data mining tech- describe how basic terms about data, information, or
niques to specific problem or data, it is useful to see knowledge are defined. We have to say that there are
anything done in a broader and more organized way. many attempts to explain them more precisely. One ex-
Therefore, successful data science projects usually fol- ample is the DIKW pyramid (Rowley, 2007). This model
low some methodology which can provide data scientist represents and characterizes information-based levels
with basic guidelines on how to challenge the problem (according to the area of information engineering) in
and how to work with data, algorithms, or models. This the chain of grading informativeness known as Data–
methodology is then a structured way to describe the Information–Knowledge–Wisdom (see Fig. 1.1). Simi-
knowledge discovery process. Without a flexible struc- lar models often apply such chains, even if some parts
ture of steps data science projects can be unsuccessful, are removed or combined. For example, very often such
or at least it will be hard to achieve a result that can be a model is simplified to Data–Information–Knowledge
easily applied and shared. A better understanding of at or even Data–Knowledge, but semantics is usually the
least an overview of the process is quite beneficial both same or similar as in the case of the DIKW pyramid.
to the data scientist and to anyone who needs to discuss Moreover, there are many models which describe not
results or steps of the process (such as data engineers, only its objects but also processes for their transitions,
customers, or managers). Moreover, in some domains, e.g., Bloom’s taxonomy (Anderson and Krathwohl,
including those working with data from astronomy and 2001), decision process models (Bouyssou et al., 2010),
geophysics, steps used in preprocessing and analysis of or knowledge management – SECI models (Nonaka et
data are crucial to understanding provided products. al., 2000). The description of methodology usually de-
From the 1990s, research in this area started to define fines what we understand under data, information, and
its terms more precisely, with the definition of knowl- knowledge level.
edge discovery (or knowledge discovery in databases While methodologies started from a more general
[KDD]) (Fayyad et al., 1996) as a synonym for knowl- view, logically more and more attempts were trans-
edge discovery process (KDP). It included data mining formed into a more structured way. Also, many of them
as one of the steps in the knowledge acquisition effort. became more tool-specific. When we try to look at the
KDD (or KDP) and data mining are even today often evolution of the KDP, the main further steps after the
seen as equal terms, but data mining is a subpart (step) creation of more general methodologies are two ba-
of the whole process dedicated to the application of al- sic concepts. First, in order to have more precise and
gorithms able to extract patterns from data. Moreover, formalized processes, many of them were transformed
KDD also becomes the first description of KDP as a into standardized process-based definitions with the au-
formalized methodology. During the next years, new ef- tomation of their steps. Such effort is logically achieved
forts bring more attempts which lead to other method- more easily by the application in specific domains (such
ologies and their applications. We will describe more as industry, medicine, science), with clear standards for
details on selected cases later. exchanging documents and often with the support of

Knowledge Discovery in Big Data from Astronomy and Earth Observation. https://doi.org/10.1016/B978-0-12-819154-5.00010-2 1
Copyright © 2020 Elsevier Inc. All rights reserved.
2 PART I Data

FIG. 1.1 DIKW pyramid – understanding the difference between Data, Information, Knowledge, and Wisdom.

specific tools used for the automation of processes. Sec- data into some useful artifacts. Hence, the real benefit
ond, when we have several standardized processes in is in our ability to extract such useful artifacts, which
different domains, it is often not easy to apply methods can be in the form of reports, policies, decisions, or rec-
from one area directly in another one. One of the solu- ommended actions. Before we provide more details on
tions is to support better cross-domain understanding processes that transform raw data into these artifacts, we
of steps using some shared terminology. This solution can start with the basic notion of data, information, or
leads to the creation of formalized semantic models like knowledge.
ontologies that are helpful in better understanding of As we already mentioned in the previous section,
terminology between domains. Moreover, another step there are different definitions with a different scope,
towards a new view of methodologies and sharing of i.e., from the DIKW pyramid with a more granular
information about them was proposed based on the on- view, to simpler definitions when there are only two
tologies of KDPs, like OntoDM (Panov et al., 2013). levels of data–knowledge relations. For our purposes
Therefore, if we summarize, generalized methodolo- we would stay with simpler versions of DIKW, where
gies are basic concepts related to KDPs. More specific we define Data–Information–Knowledge relations in
versions of them provide standards and automation in this way, adapted from broader Beckman definitions
specific domains, and on the other hand, cross-domain (Beckman, 1997):
models share domain-specific knowledge between dif- • Data – facts, numbers, pictures, recorded sound, or
another raw source usually describing real-world ob-
ferent domains. This basic overview also describes the
jects and their relations;
structure of sections for this chapter. In the next sec-
• Information – data with added interpretation and
tion, we provide some details on data–information–
meaning, i.e., formatted, filtered, and summarized
knowledge definitions and KDPs. In the following sec-
data;
tion, we describe existing more general methodologies.
• Knowledge – information with actions and applica-
In Section 1.4 we provide a look at methodologies in
tions, i.e., ideas, rules, and procedures, which lead to
a more precise way, through standardization and au- decisions and actions.
tomation efforts, as well as attempts to share knowledge While there are also extended versions of such rela-
in cross-domain view. In the following section, the as- tions, this basic view is quite sufficient with all method-
tro/geo context is discussed, mainly focusing on their ologies for KDPs. It is because raw data gathering (Data
specifics and shared aspects, and the possible transfer of part), their processing and manipulation (Informa-
knowledge. tion part), and creation of models that are suitable
for support of decisions and further actions (Knowl-
edge part) are all necessary aspects of standard data
1.2 KNOWLEDGE DISCOVERY PROCESSES analytical tasks. Hence, transformations in this Data–
Currently, we can store and access large amounts of Information–Knowledge chain represent a very gen-
data. One of the main problems is to transform raw eral understanding of the KDP or a simple version of
CHAPTER 1 Methodologies for Knowledge Discovery Processes in Context of AstroGeoInformatics 3

methodology. We have the input dataset (raw sources – of methods applied here is large, including methods
Data part), which is transferred using several steps (of- from data mining, machine learning (whenever out-
ten including data manipulation to get more interpreted put models are also applicable as actions), operation
and meaningful data – Information part) to knowledge research, optimization, computational modeling, or
(models containing rules or patterns – Knowledge part). expert (knowledge-based) systems.
For example, data from a customer survey are in the A nice feature of business analytics is that every
raw form of Yes/No answers, values on an ordinal scale, option can be applied separately, or we can combine
or numbers. If we put these data about customers in them in the chain as a step-by-step process. In this case,
the context of questions, combine them in infographics, we can see descriptive analytics mainly responsible for
and analyze their relations with each other, we trans- transformation between Data and Information. With
form raw data into information. In practice, we mine the addition of predictive analytics, we can enhance the
some rules on how these customers and their subgroups process of transformation to get Knowledge of our sys-
usually react in specific cases discussed in the survey. We tem. Our extracted knowledge is then applicable and
can try to understand their behavior (what they prefer, actionable simply as is, or we can extend it and make
buy), predict their future reactions (if they will be in- it part of the decision making process using methods
terested in a new product) in similar cases, and provide from the area of prescriptive analytics. Hence, we can
actionable knowledge in the form of a recommendation see Data–Information–Knowledge in a narrow view as
to the responsible actor (apply these rules to get higher part of predictive analytics in, let us say, traditional un-
income). derstanding (with KDPs as KDD), or we can see it in
The presented view of Data–Information–Knowl- broader scope with all analytics involved in transforma-
edge relations is also comparable to the view of business tion.
analytics. In this case, we have three options in analytics Now we can show differences in this example. Imag-
according to our expectations (Evans, 2015): ine that a company has several hotels with casinos,
• Descriptive analytics – uses data aggregation and de- and they want to analyze customers and optimize their
scriptive data mining techniques to see what hap- profit. Within descriptive analytics they use data ware-
pened in the system (business), so the question housing techniques to make reports about hotel pack-
“What has happened?” is answered. The main idea is ing in time, activities in the casino and its incomes,
to use descriptive analytics if we want to understand and infographics of profit according to different aspects.
at an aggregate level what is going on, summarize These methods will help them to understand what is
such information, and describe different aspects of happening in their casinos and hotels. Within predic-
the system in that way (to understand present and tive analytics, they can create a predictive model that
historical data). The methods here lead us to ex- forecasts hotel and casino packing in the future, or they
ploration analysis, visualizations, periodic or ad hoc can use data about customers and segment them into
reporting, trend analysis, data warehousing, and cre- groups according to their behavior in casinos. The re-
ation of dashboards. sult is a better understanding of what will happen in
• Predictive analytics – basically tasks from this part the future, what will be occupancy of the hotel in differ-
examine the future of the system. They answer the ent months, and what is the expected behavior of cus-
question “What could happen according to historical tomers when they come to the casino. Moreover, within
data?” We can see this as a predictor of states accord- prescriptive analytics, they can identify which decision-
ing to all historical information. It is an estimation based input setup (and how) to optimize their profit.
of the normal development of the characteristics of It means that according to the prediction of hotel oc-
our system. This part of analytical tasks is closest to cupancy they can change prices accordingly, set up the
the traditional view of KDPs. The methods here are allocation of rooms, or provide benefits to some seg-
the same as in the case of any KDP methodology, sta- ments of customers. For example, if someone is playing
tistical analysis, and data mining methods. a lot, we can provide him/her with some benefits to sup-
• Prescriptive analytics – here are all attempts when port his/her return like a better apartment for a lower
we select some model about the system and try to price or free food.
optimize its possible outcomes. It means that we As we already mentioned, people often exchange the
analyze what we have to do if we want to get the KDP with data mining, which is only one step. More-
best efficiency for some output model values. The over, for knowledge discovery, some other names were
name came from the word prescribe, so it is pre- also used in literature, like knowledge extraction, infor-
scription or advice for actions to be done. The set mation harvesting, information discovery, data pattern
4 PART I Data

processing, or even data archeology. The mostly used source) software tools for particular steps or one unified
synonym for KDPs is then obviously KDD, which is log- analytical platform of tools.
ical due to the beginnings of KDP with the processing of Before we move to a description of selected method-
structured data stored in standard databases. The basic ologies in the next section, we summarize the motiva-
properties are even nowadays the same as or similar to tion for the use of standardized KDP models (method-
KDD basics from the 1990s. Therefore we can summa- ologies) (Kurgan and Musilek, 2006):
rize them accordingly (Fayyad et al., 1996): • The output product (knowledge) must be useful for
• The main objective of KDP is to seek new knowledge the user, and ad hoc solutions more often failed in
in the selected application domain. yielding valid, novel, useful, and understandable re-
• Data are a set of facts. The pattern is the expression in sults.
some suitable language (part of the outcome model, • Understanding of the process itself is important. Hu-
e.g., rule written in some rule-based language) about mans often lack a perception of large amounts of
a subset of facts. untapped and potentially valuable data. A process
• KDP is a nontrivial process of identifying valid, model that is well structured and logical will help
novel, potentially useful, and ultimately understand- to avoid these issues.
able patterns in data. A process is simply a multistep • An often underestimated factor is providing support
approach of transformations from data to patterns. for management problems (this also includes cases
The pattern (knowledge) mentioned before is: of a larger project in the science area, which needs
• valid – pattern should be true on new data with efficient management). Whenever KDP projects in-
some certainty, volve large teams, requiring careful planning and
• novel – we did not know about this pattern be- scheduling, a management specialist in such projects
fore, is often unfamiliar with terms from the data min-
• useful – pattern should lead to actions (the pat- ing area – KDP methodology can then be helpful in
tern is actionable), managing the whole project.
• comprehensible – the process should produce • Standardization of KDP provides a unified view of
patterns that lead to a better understanding of the current process description and allows an appropri-
underlying data for human (or machine). ate selection and usage of technology to solve cur-
KDP is easily generalized also to sources of data rent problems in practice, mostly on an industrial
which are not in databases or not in structured form, level.
which induces methodology aspects of a similar type
also to the area of text mining, Big Data analysis, or
data streams processing. Knowledge discovery involves 1.3 METHODOLOGIES FOR KNOWLEDGE
the entire process, including storage and access of data, DISCOVERY PROCESSES
application of efficient and scalable data processing al- In this section, we provide more details on selected
gorithms to analyze large datasets, interpretation and methodologies. From the 1990s, several of them were
visualization of outcome results, and support of the developed, starting basically from academic research,
human–machine or human–computer interaction, as but they very quickly moved on to an industry level.
well as support for learning and analyzing the domain. As we already mentioned, the first more structured way
The KDP model, which is then called methodology, was proposed as KDD in Fayyad et al. (1996). Their ap-
consists of a set of processing steps followed by the proach was later modified and improved by both the
data analyst or scientist to run a knowledge discovery research and the industry community. The processes
project. The KDP methodology usually describes proce- always share a multistep sequential way in processing
dures for each step of such a project. The model helps input data, where each step starts after accessing the re-
organizations (represented by the data analyst) to un- sult of the successful completion of the previous step as
derstand the process and create a project roadmap. The its input. Also, it is common that activities within steps
main advantage is reduced costs for any ad hoc analy- cover understanding of the task, data, preprocessing or
sis, time savings, better understanding, and acceptation preparation of data, analysis, evaluation, understand-
of the advice coming from the results of the analysis. ing of results, and their application. All methodolo-
While there are still data analysts who apply ad hoc gies also emphasize their iterative nature by introducing
steps to their projects, most of them apply some com- feedback loops throughout the process. Moreover, they
mon framework with the help of (commercial or open are often processed with a strong influence of human
CHAPTER 1 Methodologies for Knowledge Discovery Processes in Context of AstroGeoInformatics 5

data scientists and therefore acknowledge its interactiv- to the result of polls from years 2007 and 2014, more
ity. The main differences between the methodologies than 42% of data analysts (most of all votes) are using
are in the number and scope of steps, the characteris- CRISP-DM methodology in their analytics, data mining,
tics of their inputs and outputs, and the usage of various or data science projects, and the usage of the methodol-
formats. ogy seems to be stable.
Several studies compared existing methodologies,
their advantages and disadvantages, the scope of their 1.3.1 First Attempt to Generalize Steps –
application, the relation to software tools and stan- Research-Based Methodology
dards, and any other aspects. Probably the most ex- Within the starting field of knowledge discovery in the
tensive comparisons of methodologies can be found in 1990s, researchers defined the multistep process, which
Kurgan and Musilek (2006) and Mariscal et al. (2010). guides users of data mining tools in their knowledge
Other papers also bring ideas and advice, including discovery effort. The main idea was to provide a se-
their applicability in different domains; see, for exam- quence of steps that would help to go through the KDP
ple, Cios et al. (2007), Ponce (2009), Rogalewicz and in an arbitrary domain. As mentioned before, in Fayyad
Sika (2016). et al. (1996) the authors developed a model known as
Before we describe details of some selected method- KDD process.
ologies, we provide some information on two aspects, In general, KDD provides a nine-step process, mainly
i.e., the evolution of methodologies and practical usage considered as a research-based methodology. It involves
of them by data analysts. both the evaluation and interpretation of the patterns
According to the history of methodologies, in (possibly knowledge) and the selection of preprocess-
Mariscal et al. (2010) one can find quite a thorough de- ing, sampling, and projections of the data before the
scription of such evolution. As we already mentioned, data mining step. While some of these nine steps fo-
the first attempts were fulfilled by Fayyad’s KDD pro- cus on decisions or analysis, other steps are data tran-
cess between the years 1993–1996, which we will also sitions within the data–information–knowledge chain.
describe in the next subsection. This approach inspired As mentioned before, KDD is a “nontrivial process of
several other methodologies, which came in the years identifying valid, novel, potentially useful, and ulti-
after the KDD process, like SEMMA (SAS Institute Inc., mately understandable patterns in data” (Fayyad et al.,
2017), Human-Centered (Brachman and Anand, 1996), 1996). The KDD process description also provides an
or approaches described in Cabena et al. (1998) and outline of its steps, which is available in Fig. 1.2.
Anand and Buchner (1998). On the other hand, also The model of the KDD process consists of the follow-
some other ideas evolved into methodologies including ing steps (input of each step is output from the previous
the 5As or Six Sigma. Of course, some issues were iden- one), in an iterative (analysts apply feedback loops if
tified during those years and an answer to them was in necessary) and interactive way:
the development of CRISP-DM standard methodology, 1. Developing and understanding the application do-
which we will also describe in one of the following sub- main, learning relevant prior knowledge, identifying
sections. CRISP-DM became the leading methodology of the goals of the end-user (input: problem to be
and quite a reasonable solution for a start in any data solved/our goal, output: understanding of the prob-
mining project, including new projects with Big Data lem/domain/goal).
and data streams processing. Any new methodology or 2. Creation of a target dataset – selection (querying) of
some standardized description of processes usually fol- the dataset, identification of subset variables (data
lows a similar approach to one defined by CRISP-DM attributes), and the creation of data samples for the
(some of them are available in the review papers men- KDP (output: target data/dataset).
tioned before). 3. Data cleaning and preprocessing – dealing with out-
The influential role of CRISP-DM is evident by liers and noise removal, handling the missing data,
the polls evaluated on KDnuggets,1 a well-known collecting data on time sequences, and identifying
and widely accepted community-based web site re- known changes to data (output: preprocessed data).
lated to knowledge discovery and data mining. Gregory 4. Data reduction and projection – finding useful fea-
Piatetsky-Shapiro, one of the authors of the KDD pro- tures that represent the data (according to goal), in-
cess methodology, showed in his article2 that according cluding dimension reductions and transformations
1 https://www.kdnuggets.com/. (output: transformed data).
2 https://www.kdnuggets.com/2014/10/crisp-dm-top-methodology- 5. Selection of data mining task – the decision on
analytics-data-mining-data-science-projects.html. which methods to apply for classification, cluster-
6 PART I Data

FIG. 1.2 The KDD process.

ing, regression, or another task (output: selected 1.3.2 Industry-Based Standard – the
method[s]). Success of CRISP-DM
6. Selection of data mining algorithm(s) – select meth- Shortly after the KDD process definition, the indus-
od for pattern search, deciding on appropriate mod- try produced methodologies more suitable for their
els and their parameters, and matching methods needs. One of them is CRISP-DM (CRoss-Industry Stan-
with the goal of the process (output: selected algo- dard Process for Data Mining) (Chapman et al., 2000),
rithms). which became the standard for many years and is still
7. Data mining – searching for patterns of interest in widely used in both the industry and the research area.
specific form like classification rules, decision trees, CRISP-DM was originally developed by a project con-
regression models, trends, clusters, and associations sortium under the ESPRIT EU funding initiative in
(output: patterns). 1997. The project involved several large companies,
8. Interpretation of mined patterns – understanding which cooperated in its design: SPSS, Teradata, Daimler
and visualizations of patterns based on the extracted AG, NCR Corporation, and OHRA. Thanks to the differ-
models (output: interpreted patterns). ent knowledge of companies, the consortium was able
9. Consolidation of discovered knowledge – use of dis- to cover all aspects, like IT technologies, case studies,
data sources, and business understanding.
covered patterns into a system analyzed by the KDD
CRISP-DM is an open standard and is available for
process, documenting and reporting knowledge to
anyone to follow. Some of the software tools (like SPSS
end-users, and checking and resolving conflicts if
Modeler/SPSS Clementine) have CRISP-DM directly in-
needed (output: knowledge, actions/decisions based
corporated. As we already mentioned, CRISP-DM is the
on the results).
most widely used KDP methodology. While it still has
The authors of this model declared its iterative fash- some drawbacks, it became a part of the most success-
ion, but they gave no specific details. The KDD process ful story in the data mining industry. The central fact
is a simple methodology and quite a natural model for behind this success is that CRISP-DM is industry-based
the discussion of KDPs. There are two significant draw- and neutral according to tools and application. One of
backs of this model. First, lower levels are too abstract the drawbacks of this model is that it does not perform
and not explicit and formalized. This lack of detail was project management activities. One major factor behind
changed in later methodologies using more formalized the success of CRISP-DM is that it is an industry tool,
step descriptions (in some cases using standards, au- and it is application-neutral (Mariscal et al., 2010).
tomation of processes, or specific tools or platforms). The CRISP-DM model (see Fig. 1.3) consists of the
The second drawback is its lack of business aspects de- following six steps, which are then described in more
scription, which is logical due to the research-based idea details and can be iteratively applied, including feed-
at the start of its development. back in some places (where necessary):
CHAPTER 1 Methodologies for Knowledge Discovery Processes in Context of AstroGeoInformatics 7

FIG. 1.3 Methodology CRISP-DM.

1. Business understanding – focuses on the under- tion of data, software tools, technical deploy-
standing of objectives and requirements from a busi- ment).
ness perspective, and also converts them into the 2. Data understanding – initial collection of data, un-
technical definition and prepares the first version of derstanding the data quality issues, exploration anal-
the project plan to achieve the objectives. Therefore, ysis, detection of interesting data subsets. If under-
substeps here are: standing shows a need to reconsider business under-
a. determination of business objectives – here it standing substeps, we can move back to the previous
is important to define what we expect as busi- step. Hence, the substeps of data understanding are:
ness goals (costs, profits, better support of cus- a. collection of initial data – the creation of the
tomers, and higher quality of the data prod- first versions of the dataset or its parts,
uct), b. description of data – understanding the mean-
b. assessment of the situation – understanding the ing of attributes in data, summary of the initial
actual situation within the objectives, defining dataset(s), extraction of basic characteristics,
the criteria of success for business goals, c. exploration of data – visualizations, descrip-
c. determination of technical (data mining) tions of relations between attributes, correla-
goals – business goals should be transformed tions, simple statistical analysis on attributes,
into technical goals, i.e., what data mining exploration of the dataset,
models we need to achieve business goals, what d. verification of data quality – analysis of missing
the technical details of these models are, how values, anomalies, or other issues in data.
we will measure it, 3. Data preparation – after finishing the first steps, the
d. generation of a project plan – the analyst cre- most important step is the preparation of data for
ates the first version of the plan, where details data mining (modeling), i.e., the preparation of the
on next steps are available. The analysts should final dataset for modeling using data manipulation
address different issues, from business aspects methods which can be applied. We can divide them
(how to discuss and transform data mining re- into:
sults, deployment issues from a management a. selection of data – a selection of tables, records,
point of view) to technical aspects (how to and attributes, according to goal needs and re-
achieve data, data formats, security, anonymiza- duction of dimensionality,
8 PART I Data

b. integration of data – identification of the same plication of KDP (lifecycle applications). This step
entities within more tables, aggregations from consists of:
more tables, redundancy checks, and processing a. plan deployment – the deployment strategy
of and detection of conflicts in data, is provided, including the necessary steps and
c. cleansing of data – processing of missing values how to perform them,
(remove records or imputation of values), pro- b. plan monitoring and maintenance – strategy
cessing of anomalies, removing inconsistencies, for the monitoring and maintenance of deploy-
d. construction (transformation) of data – the cre- ment,
ation of new attributes, aggregations of values, c. generation of the final report – preparation of
transformation of values, normalizations of val- the final report and final presentation (if ex-
ues, and discretization of attributes, pected),
e. formatting of data – preparation of data as in- d. review of the process substeps – summary of
put to the algorithm/software tool for the mod- experience from the project, unexpected prob-
eling step. lems, misleading approaches, interesting solu-
4. Modeling – various modeling techniques are ap- tions, and externalization of best practices.
plied, and usually more types of algorithms are used, CRISP-DM is relatively easy to understand and has
with different setup parameters (often with some good vocabulary and documentation. Thanks to its gen-
metaapproach for optimization of parameters). Be- eralized nature, this methodology is a very successful
cause methods have different formats of inputs and and extensively used model. In practice, many advanced
other needs, the previous step of data preparation analytic platforms are based on this methodology, even
could be repeated in a small feedback loop. In gen- if they do not call it the same way.
eral, this step consists of: In order to help in understanding the process, we
a. selection of modeling technique(s) – choose can provide a simple example. One of the possible ap-
the method(s) for modeling and examining plications of the CRISP-DM methodology is to provide
their assumptions, tools in support of clinical diagnosis in medicine. For
b. generation of test design – plan for training, example, our goal is to improve breast cancer diagnos-
testing, and evaluating the models, tics using data about patients. In terms of CRISP-DM
c. creation of models – running the selected meth- methodology we can describe the KDP in the following
ods, way:
d. assessment of generated models – analysis of 1. Business understanding – from a business perspec-
models and their qualities, revision of param- tive, our business objective goal is to improve the
eters, and rebuild. effectiveness of breast cancer diagnostics. Here we
5. Evaluation – with some high-quality models (ac- can provide some expectation in numbers related
cording to the data analysis goal), such models are to diagnostics effectiveness and costs of additional
evaluated from a business perspective. The analyst re- medical tests, in order to set up business goals –
views the process of model construction (to find in- for example, if our diagnosis using some basic setup
sufficiently covered business issues) and also decides will be more effective, it reduces the costs by 20%.
on the next usage of data mining results. Therefore, Then data mining goals are defined. In terms of data
we have: mining, it is a classification task with the binary tar-
a. evaluation of the results – assessment of results get attribute, which will be tested using a confusion
and identification of approved models, matrix, and according to business goals we want to
b. process review – summarize the process, iden- achieve at least 95% accuracy of the classifier to ful-
tify activities which need another iteration, fill the business goal. According to the project plan,
c. determination of the next step – a list of further we know that data are available in CSV format, and
actions is provided, including their advantages data and models are processed in R using RStudio,
and disadvantages, with the Rshiny web application (on available server
d. decision – describe the decision as to how to infrastructure) providing the interface for doctors in
proceed. their diagnostic process.
6. Deployment – discovered knowledge is organized 2. Data understanding – in this example, let us say we
and presented in the form of reports or some com- have data collected from the Wisconsin Diagnosis
plex deployment is done. Also, this can be a step that Breast Cancer (WDBC) database. We need to under-
finishes one of the cycles if we have an iterative ap- stand the data themselves, and what are their at-
CHAPTER 1 Methodologies for Knowledge Discovery Processes in Context of AstroGeoInformatics 9

tributes and what is their meaning. In this case, we 1.3.3 Proprietary Methodologies – Usage of
have 569 records with 32 attributes, which mostly Specific Tools
describe original images with/without breast cancer. While the research or open standard methodologies are
The first attribute is ID and the second attribute is tar- more general and tool-free, some of the leaders in the
get class (binary – the result of diagnosis). The other area of data analysis also provide to their customers pro-
30 real-valued attributes describe different aspects of prietary solutions, usually based on the usage of their
cells in the image (shape, texture, radius). We also software tools.
find no missing values, and we do not need any pro- One of such examples is the SEMMA methodol-
cedure to clean or transform data. We also explore ogy from the SAS Institute, which provided a process
data, visualize them, and describe relations between description on how to follow its data mining tools.
attributes and correlations, in order to have enough SEMMA is a list of steps that guide users in the im-
information for the next steps. plementation of a data mining project. While SEMMA
3. Data preparation – any integration, cleaning, and provides still quite a general overview of KDP, authors
transformation issues are solved here. In our ex- claim that it is a most logical organization of their tools
ample, there are no missing values other issues in to cover core data mining tasks (known as SAS Enter-
WDBC. There is only one data table, we will select prise Miner). The main difference of SEMMA with the
traditional KDD overview is that the first steps of appli-
all records, and we will not remove/add an attribute.
cation domain understanding (or business understand-
The data format is CSV, suitable for input in RStu-
ing in CRISP-DM) are skipped. SEMMA also does not
dio for the modeling step. We can also select subsets
include the knowledge application step, so the business
of data according to expected modeling and evalua-
aspect is out of scope for this methodology (Azevedo
tion, in this case, let us say a simple hold-out method
and Santos, 2008). Both these steps are in the knowl-
with different ratios for the size of training and test
edge discovery community considered as crucial for the
samples (80:20, 70:30, 60:40). success of projects. Moreover, applying this methodol-
4. Modeling – data mining models are created. In our ogy outside SAS software tools is not easy. The phases
case, we want classification models (algorithms), of SEMMA and related tasks are the following:
i.e., C4.5, Random Forests, neural networks, k-NN, 1. Sample – the first step is data sampling – a selection
SVM, and naive Bayes. We create models for differ- of the dataset and data partitioning for modeling;
ent hold-out selections and parameters of algorithms the dataset should be large enough to contain rep-
to achieve the best models. Then we evaluate mod- resentative information and content, but still small
els according to test subsets and select the best of enough to be processed efficiently.
them for further deployment, i.e., the SVM-based 2. Explore – understanding the data, performing ex-
model with more than 97% accuracy with 70:30 ploration analysis, examining relations between the
hold-out. variables, and checking anomalies, all using simple
5. Evaluation – the best models are analyzed from a statistics and mostly visualizations.
business point of view, i.e., whether we can achieve 3. Modify – methods to select, create, and transform
the business goal using such a model and its suffi- variables (attributes) in preparation for data model-
ciency for application in the deployment phase. We ing.
decide on how to proceed with the best model, and 4. Model – the application of data mining techniques
what the advantages and disadvantages are. For ex- on the prepared variables, the creation of models
ample, in this case, the application of the selected with (possibly) the desired outcome.
model can support doctors and remove one intru- 5. Assess – the evaluation of the modeling results, and
sive and expensive test out of diagnostics, in some of analysis of reliability and usefulness of the created
the new cases. models.
6. Deployment – a web-based application (based on IBM Analytics Services have designed a new method-
Rshiny) is created and deployed on the server, which ology for data mining/predictive analytics named An-
contains an extracted model (SVM classifier) and a alytics Solutions Unified Method for Data Mining/Pre-
user interface for the doctor in order to input results dictive Analytics (also known as ASUM-DM),3 which is
a refined and extended CRISP-DM. While strong points
of image characteristics from new patients (records)
and provide him/her with a diagnosis of such new 3 https://developer.ibm.com/predictiveanalytics/2015/10/16/have-you-
samples. seen-asum-dm/.
10 PART I Data

of CRISP-DM are on the analytical part, due to its open is Architecture-centric Agile Big data Analytics (AABA)
standard nature CRISP-DM does not cover the infras- (Chen et al., 2016), which addresses technical and orga-
tructure or operations side of implementing data min- nizational challenges of Big Data with the application
ing projects, i.e., it has only few project management of agile delivery. It integrates Big Data system Design
activities, and has no templates or guidelines for such (BDD) and Architecture-centric Agile Analytics (AAA)
tasks. with the architecture-supported DevOps model for ef-
The primary goal of ASUM-DM creation was to solve fective value discovery and continuous delivery of value.
the disadvantages mentioned above. It means that this The authors validated the method based on case studies
methodology retained CRISP-DM and augmented some from different domains and summarized several recom-
of the substeps with missing activities, tasks, guidelines, mendations for Big Data analytics:
and templates. Therefore, ASUM-DM is an extension or
• Data analysts should be involved already in the busi-
refinement of CRISP-DM, mainly in the more detailed
ness analysis phase.
formalization of steps and application of (IBM-based)
• There should be continuous architecture support.
analytics tools. ASUM-DM is available in two versions –
• Agile steps are important and helpful due to fast
an internal IBM version and an external version. The in-
technology and requirements changes in this area.
ternal version is a full-scale version with attached assets,
and the external version is a scaled-down version with- • Whenever possible, it is better to follow the reference
out attached assets. Some of these ASUM-DM assets or a architecture to make development and evolution of
modified version are available through a service engage- data processing much easier.
ment with IBM Analytics Services. Like SEMMA, it is a • Feedback loops need to be open and should include
proprietary-based methodology, but more detailed and both technical and business aspects.
with a broad scope of covered steps within the analyti- As we already mentioned, processing of data and
cal project. their lifecycle is quite an important aspect in this area.
At the end of this section, we also mention that KDPs Moreover, the setup of processing architecture and tech-
can be easily extended using agile methods, initially nology stack is probably of the same importance in
developed for software development. The main appli- the Big Data context. One approach for solving such
cation of agile-based aspects is logically in larger teams issues is related to the Big Data Integrator (BDI)Plat-
in the industrial area. Many approaches are adapted form (Ermilov et al., 2017), developed within the Big
explicitly for some company and are therefore propri- Data Europe H2020 flagship project, which provides
etary. Generally, KDP is iterative, and the inclusion of distribution of Big Data components as one platform
more agile aspects is quite natural (Nascimento and with easy installation and setup. While there are several
de Oliveira, 2012). The AgileKDD method fulfills the other similar distributions, authors of this platform also
OpenUP lifecycle, which implements Agile Manifesto. provided to potential users a methodology for devel-
The project consists of sprints with fixed deadlines (usu- oping Big Data stack applications and several use cases
ally a few weeks). Each sprint must deliver incremental from different domains. One of their inspirations was to
value. Another example of an agile process description use the CRISP-DM structure and terminology and apply
is also ASUM-DM from IBM, which combines project
them to a Big Data context, like in Grady (2016), where
management and agility principles.
the author extends CRISP-DM to process scientific Big
1.3.4 Methodologies in Big Data Context Data. In the scope of the BDI Platform, authors pro-
posed a BDI Stack Lifecycle methodology, which sup-
Traditional methodologies are usually applied also in
Big Data projects. The problem here is that none of the ports the creation, deployment, and maintenance of the
traditional standards support the description of the ex- complex Big Data applications. The BDI Stack Lifecycle
ecution environment or workflow lifecycle aspects. In consists of the following steps (they developed docu-
the case of Big Data projects, it is an important issue mentation and tools for each of the steps):
due to the complex cluster of distributed services im- 1. Development – templates for technological frame-
plemented using the various technologies (distributed works, most common programming languages, dif-
databases, frameworks for distributed processing, mes- ferent IDEs applied, distribution formalized for the
sage queues, data provenance tools, coordination, and needs of users (data processing task).
synchronization tools). An interesting paper discussing 2. Packaging – dockerization and publishing of the de-
these aspects is Ponsard et al. (2017). One of the men- veloped or existing components, including best prac-
tioned methodologies related to Big Data in this paper tices that can help the user to decide.
CHAPTER 1 Methodologies for Knowledge Discovery Processes in Context of AstroGeoInformatics 11

3. Composition – assembly of a BDI stack, integration The primary standards designed for the process mod-
of several components to address the defined data eling are flowcharting techniques which represent the
processing task. process using a graph diagram. Nodes of the diagram
4. Enhancement – an extension of BDI stack with en- correspond to the performed process activities, and
hancement tools (daemons, logging) that provides edges represent control flow. This flowchart represents
monitoring. the execution ordering of the activities or data flow,
5. Deployment – instantiation of a BDI stack on physi- i.e., how the data objects pass from one operation to
cal or virtual servers. another one. Examples of the standards based on the
6. Monitoring – observing the status of a running BDI graphical notation include the Business Process Model
stack, repetition of BDI components, and architec- and Notation (BPMN4 ) or the Unified Modeling Lan-
ture development when need. guage (UML5 ) notation. BPMN models consist of sim-
ple diagrams constructed from a limited set of graph-
ical elements with the flow objects (graph nodes) and
1.4 METHODOLOGIES IN ACTION connecting objects (graph edges). The flow objects rep-
resent activities and gateways which determine forking
In practice, when it is necessary to apply the method-
and merging of connection paths, depending on the
ology, specific views and needs are expected for users
conditions expressed. We can group flow objects using
(data analyst). The general data analysis methodologies
the swim lanes representing, for example, the organi-
are not very formalized, i.e., their direct application for
zation units or different roles in the process. A part of
machine-readable sharing or automation of data analy-
the process model for data processing can be additional
sis processes is not easy. We must look at ways how an-
annotations representing the data objects generated or
alysts brought methodologies in action within a more received by the activities. Activities can be atomic tasks,
precise context. This section will look at such aspects, or they can consist of further decomposed subprocesses.
especially on the automation of KDP and understand- BPMN is a graphical modeling notation, but version
ing of their steps through shared ontologies. 2.0 also specifies the basic execution semantics, and the
workflow engines can directly execute BPMN diagram
1.4.1 Standardization and Automation of modeling in order to automatize the processes. Addi-
Processes – Process Models tionally, BPMN models can be directly mapped to work-
The primary goal of process modeling is to represent flow execution languages, such as Web Services Business
the process in such a way that it can be analyzed, im- Process Execution Language (WS-BPEL6 ). The main dis-
proved, or automatized. In the scope of data analysis, advantage of the BPMN is the lack of direct support
the data analytical work is organized itself as the pro- for knowledge creation processes and support for deci-
cess consisting of the various steps, such as process and sion rules, and some ambiguity in the sharing of BPMN
data understanding, data preprocessing, and modeling. models.
Process modeling is also crucial for the data provenance, In comparison to BPMN, UML is a general purpose
where it is necessary to capture how the data were trans- modeling language which provides many types of dia-
formed using the sequence of operations represented as grams from two categories: types representing the struc-
the data flow process model. Additionally, the analyzed tural information and types representing the general
domain can be process-oriented, as is the case for ex- types of behavior, including types representing differ-
ample in the process industries, i.e., process models can ent aspects of interactions. The behavior types can be
be an essential part of the domain knowledge shared directly used for the process modeling using the activity
by the domain experts and data scientist. Depending diagrams or in some cases sequence diagrams. A UML
on the complexity of the model, the process model- activity diagram generally describes step-by-step opera-
ing is typically performed by the process analysts, who tional activities of the components in a modeled system
provide expertise in the modeling discipline in cooper- using a similar flowcharting technique like BPMN dia-
ation together with the domain experts. In the case of grams. The activity diagram consists of nodes represent-
data analysis, data scientists typically perform the pro- ing the activities or decision gateway with support for
cess analysis. Models based on some formalism, like in choice, iteration, and concurrency. For data flow mod-
the form of the sequence diagrams, can be designed di- eling, diagrams can be additionally annotated with the
rectly by the domain experts. For the alternative, the 4 http://www.bpmn.org.
process model can be derived directly from the observed 5 http://www.uml.org.
process events using the process mining tools. 6 http://docs.oasis-open.org/wsbpel/2.0/wsbpel-v2.0.html.
12 PART I Data

FIG. 1.4 Example of a PMML file (from DMG PMML examples).

references to the structural entities. The main diagram tive models such as Predictive Model Markup Language
for structural modeling is the class diagram, which al- (PMML7 ) or Portable Format for Analytics (PFA8 ). The
lows modeling classes and instances of entities together core of the PMML standard is a structural description of
with the data types of their properties and interdepen- the input and output data and parameters of the mod-
dencies. The main advantage of the UML is its general els, but the format also allows to specify a sequence
applicability to the different aspects which can be mod- of data processing transformation which allow for the
eled, including the structural aspects not included in the mapping of user data into a more desirable form to be
BPMN. used by the mining model. This sequence together with
The flowcharting techniques were also directly in- the data dictionary (structural specification of the in-
corporated into the tools for data analytics. Tools such put and output data) can be used to represent data flow
as IBM SPSS Modeler or RapidMiner provide a visual and data provenance. An example of a structure of one
interface which allows users to leverage statistical and PMML file representing a regression task is shown in
data mining algorithms without programming. The Fig. 1.4.
data processing and modeling process is represented In comparison to PMML, the PFA standard is a
as the graph chart with the nodes representing data more generic functional execution model which pro-
vides control structures, such as conditionals and loops
sources, transformation operations, machine learning
(like a typical programming language). In PFA, data
algorithms, or build models applied to the data. The
processing operations are represented as the function
data flowchart (or data stream) is stored using the
proprietary format, but most of the modeling tools 7 http://dmg.org/pmml/v4-3/GeneralStructure.html.
also support standards for exchanging of the predic- 8 http://dmg.org/pfa.
CHAPTER 1 Methodologies for Knowledge Discovery Processes in Context of AstroGeoInformatics 13

with inputs and outputs. The standard provides the vo- the upper-level ontology and specific entities defined
cabulary of common operations for the basic data types in the domain ontologies. Upper-level and mid-level
and the language to specify user-defined functions. The ontologies are designed to be able to provide a mech-
data analysis process is then the composition of the anism for mapping subjects across domains. Mid-
functions. Although this approach is very flexible, it level ontologies usually provide a more specific ex-
lacks comprehensiveness of the graphical models. pression of abstract entities found in upper-level on-
tologies.
1.4.2 Understanding Each Other – Semantic • Domain ontologies – Domain ontologies specify en-
Models tities relevant to the domain and represent the most
During the data analysis process, domain experts have specific knowledge from the perspective of one do-
to share the domain knowledge with the data scientists main.
in order to understand business or research problem, The upper-level ontologies are especially important
identify goals of the data analysis tasks, identify relevant for the integration and sharing of the knowledge across
data, and understand relations between them. Data ana- multiple domains and provide a framework through
lysts also exchange knowledge in the opposite direction, which different systems can use a common base. The
i.e., they communicate with the domain experts during entities in upper-level ontologies are basic and univer-
the interpretation and the validation of the data analy- sal and are usually limited to meta, generic, abstract,
sis results. In order to capture the exchanged knowledge, and philosophical concepts. The following list describes
data analysts use various knowledge representations for some commonly used generic upper ontologies:
the externalization process. • Suggested Upper Merged Ontology (SUMO) –
Currently, the most elaborated knowledge represen- SUMO (Niles and Pease, 2001) was created by merg-
tation techniques are ontologies known from the Se- ing several public ontologies into one coherent struc-
mantic web area. Semantic web technologies cover the ture. Ontology is used for search and applied re-
whole stack for the representation of both knowledge search, linguistics, and logic. The SUMO core con-
about structure representing classes and instances of the tains approximately 1000 classes and 4000 axioms. It
entities and their relationship or procedural knowledge consists of SUMO core ontology, mid-level ontology,
in the form of the inference or production rules. The and a set of domain ontologies such as communica-
structural formalisms are based on the application of tion, economics, finance, and physical elements.
logic and were standardized as the ontology languages • CYC ontology – CYC (Lenat, 1995) provides a com-
such as Ontology Web Language (OWL9 ), developed by mon ontology, an upper-level language for defining
the World Wide Web Consortium (W3C). Its predeces- and creating arguments using ontology. CYC ontol-
sor is an RDF scheme developed as a standard language ogy is used in the field of natural language process-
for representing ontology. The highest priority during ing, word comprehensibility, answers to questions,
the design phase of OWL was to achieve better exten- and others.
sibility, modifiability, and interoperability. OWL is now • Descriptive Ontology for Linguistic and Cognitive
striving to achieve a good compromise between scala- Engineering (DOLCE) – A very significant top ontol-
bility and expressive power. ogy is DOLCE (Gangemi et al., 2002), which focuses
Semantic models based on the ontologies can be on capturing the ontological categories needed to
generally divided depending on the scope and level of represent the natural language and human reason.
specificity of the knowledge into the following three lev- Established upper-level categories are considered as
els:
cognitive artifacts that depend on human perception,
• Upper-level ontologies – Upper ontologies describe
cultural impulses, and social conventions. Categories
the most common entities, contain only general
include abstract quality, abstract area, physical ob-
specifications, and are used as a basis for specializa-
ject, the quantity of matter, physical quality, physical
tions. Typical entries in top ontology are, e.g., “en-
area, and process. DOLCE ontology applications in-
tity,” “object,” and “situation,” which include more
clude searching for multilingual information, web
specific concepts. Boundaries expressed by the top
systems and services, and e-learning.
levels of ontologies consist of general world knowl-
• Basic Formal Ontology (BFO) – The BFO focuses
edge that is not acquired by language.
on the role of providing a true upper-level ontol-
• Middle-level ontologies – Mid-level ontologies serve
ogy that can be used to support domain ontologies
as a bridge between the general entities defined in
developed, for example, for scientific research such
9 https://www.w3.org/OWL. as biomedicine (Smith et al., 2007). The BFO rec-
14 PART I Data

ognizes the basic distinction between the following tion of the proposed study, and documentation of the
two types of entities: the essential entities that per- results achieved. The OBI ontology uses rigid logic and
sist over time while preserving their identity and the semantics because it uses higher levels of ontology rela-
procedural entities that represent the entities that tionships to define higher levels and a set of relation-
are becoming and developing over time. The char- ships. OBI defines processes and contexts (materials,
acteristic feature of process entities is that they are tasks, tools, functions, properties) relevant to biomed-
expanded both in space and in time (Grenon and ical areas.
Smith, 2004). In comparison to OBI, EXPO (Soldatova and King,
• General Formal Ontology (GFO) – GFO (Herre et al., 2006) is a more generic ontology and is not specific
2006) is a basic ontology integrating objects and pro- to the biological domain. The EXPO ontology includes
cesses. GFO has a three-layer ontological architecture general knowledge of scientific experimental design,
consisting of an abstract top level, an abstract core methodology, and representation of results. The in-
level, and a basic level. This ontology involves ob- vestigator, method, outcome, and conclusion are the
jects as well as processes integrated into one coherent main results with which EXPO defines the types of two
framework. GFO ontology is designed to support in- main investigations, i.e., computational investigations
teroperability based on ontological mapping and re- and physical investigations. Ontology uses a subset of
duction principles. GFO is designed for applications, SUMO as the highest classes and minimizes the set of
especially in the medical, biological, and biomedical relationships to ensure compliance with existing for-
fields, but also in the field of economics and sociol- malisms. The EXPO ontology is a valuable resource for
ogy. a description of the experiments from various research
• Yet Another More Advanced Top-level Ontology
areas. The authors used EXPO ontology to describe high
(YAMATO) – The YAMATO ontology (Mizoguchi,
energy and phylogenetic experiments.
2010) was developed mainly to address the deficien-
The EXPO ontology was further extended to the
cies of other upper-level ontologies, such as DOLCE,
LABORS ontology, which defines research units such as
BFO, GFO, SUMO, and CYC. It concerns the solu-
investigation, study, testing, and repetition. These are
tions of qualities and quantities dealing with the
needed to describe the complex multilayer examina-
representation and content of things and the differ-
tions performed by the robot in a fully automatic way.
entiation of the processes of the process. The current
LABORS is used to create experimental robot survey de-
version of YAMATO has been widely used in several
scriptions that result in the formalization of more than
projects, such as the development of medical ontol-
ogy. 10,000 research units in a tree structure that is 10 levels
Regarding the mid-level ontologies, in recent years, deep. Formalization describes how the robot has con-
there is an increased need for formalized representa- tributed to the discovery of new science-related knowl-
tions of the data analytics processes and formal repre- edge through the process (Soldatova and King, 2006).
sentation of outcomes of research in general. Several Semantic technologies were also applied directly to
formalisms for describing scientific investigations and formalize knowledge about the data analytics processes
outcomes of research are available, but most of them are and KDD. The initial goal of this effort was to build an
specific for the particular domain (e.g., biomedicine). intelligent data mining assistant that combines plan-
Examples of such formalisms include Ontology of ning and metalearning for automatic design of data
Biomedical Investigations (OBI), or ontology of exper- mining workflows. The assistant relies on the formal-
iments (EXPO). These ontologies specify useful con- ized ontologies of data mining operators which spec-
cepts, which describe general processes producing out- ify constraints, required inputs, and provided outputs
put data given some input data and formalize outputs for various operations in data preprocessing, modeling,
and results of the data analytics investigations. and validation phases. Examples of the ontologies for
The goal of OBI ontology (Schober et al., 2007) is data mining/data analytics are the Data Mining OPti-
to provide a standard for the representation of biologi- mization Ontology (DMOP) (Hilario et al., 2011) and
cal and biomedical examinations. OBI is entirely in line the Ontology of Data Mining (OntoDM) (Panov et al.,
with existing formalizations in biomedical areas. On- 2013). The main concepts describe the following enti-
tology promotes consistent annotation of biomedical ties:
research regardless of the specific field of study. OBI de- • Datasets, consisting of data records of the specified
fines the investigation as a multipart process, including type, which can be primitive (nominal, Boolean, nu-
the design of a general design study, the implementa- meric) or structured (set, sequence, tree, graph).
CHAPTER 1 Methodologies for Knowledge Discovery Processes in Context of AstroGeoInformatics 15

FIG. 1.5 Example of an EXPO ontology instance from the domain of high energy physics.

• Data mining tasks, which include predictive mod- discovery and traceability/reproducibility of the scien-
eling, pattern discovery, clustering, and probability tific experiments, EXPO also allows the reasoning (logi-
distribution estimation. cal inference) about the consistency and validity of the
• Generalization, the output of a data mining algo- conclusions stated in the articles or automatic genera-
rithm, which can be: predictive modeling, pattern tion of the new hypothesis for the further research.
discovery, clustering, and probability distribution es- The following example from the use of EXPO ontol-
timation. ogy, with the main properties of the structured record
• Data mining algorithms, which solve a data min- and their values, is illustrated in Fig. 1.5. The figure
ing task, produce generalizations from a dataset, and describes the fragment of the EXPO structured record
include components of algorithms such as distance (ontology instance) created by the annotation of the
functions, kernel functions, or refinement operators. scientific paper from the domain of high energy physics
describing the new estimate of the mass of the top quark
1.4.2.1 Example – EXPO (Mtop ) authored by the “D0 Collaboration” (approxi-
Scientific experiments and their results are usually de- mately 350 scientists). The experiment was unusual as
scribed in papers published in the scientific journals. In no new observational data were generated. Instead, it
these papers, the main aspects needed for the precise presented the results of applying a new statistical analy-
interpretation and reproducibility of the experiments sis method to existing data. No explicit hypothesis was
are presented in the natural language free text ambigu- put forward in the paper. However, the structured record
ously or implicitly. Therefore, it is difficult to search for includes the formalized description of the paper’s im-
the relevant experiments automatically, interpret their plicit experimental hypothesis, i.e., given the same ob-
results, or capture their reproducibility. The EXPO on- served data, the use of the new statistical method will
tology enables one to describe experiments in a more produce a more accurate estimate of Mtop than the orig-
explicit and unambiguous structured way. Besides the inal method.
16 PART I Data

The record consists of three main parts. The first specified algorithm parameter settings. All these aspects
part is the experiment classification according to the have to be covered in the description to achieve trace-
type (ComputationalExperiment: Simulation) and do- ability and reproducibility of the data mining process.
main. The classification terms are specified as the en- Fig. 1.6 presents the process aspect of a data mining al-
tries from the controlled vocabularies of existing clas- gorithm in more detail.
sification schemes or external ontologies (e.g., Dewey Each process has defined input and output entities
Decimal Classification [DDC]). This part of the EXPO which are linked to the process via has-specified-input
record allows efficient retrieval of the experiments rel- and has-specified-output relations. An input to an ap-
evant for the specified domain or problem. The next plication of a data mining algorithm is a dataset and
property describes the research hypothesis of the ex- parameter values, and as output, we get a generaliza-
periment in the natural language and (optionally) in tion, i.e., a data mining model such as a decision tree.
the artificial (logic) language which allows structural A dataset has as parts data items that are characterized
matching of the hypotheses during the retrieval/com- with a data type which can be primitive (i.e., nomi-
parison of experiments and automatic validation or nal values, numeric values, Boolean values) or com-
generation of new potential hypotheses for further re- bined into rows/ tuples. Besides the description of the
search. The third part describes the procedure or ap- building of data mining models, the OntoDM also sup-
plied method of the experiment, i.e., in this case, the ports the description of applying the model to the new
applied method was a statistical factored model. EXPO dataset (e.g., prediction using a decision tree). It al-
records explicitly describe its inputs, target variable, and lows us to describe very complex experimental processes
assumptions, which are necessary for the traceability built by the composition of multiple models and steps
and reproducibility of the experiment process. It also mixing the data mining method with other scientific
describes the conclusion of the experiment in the nat- methods, such as simulations.
ural language, but it is possible to use also an artificial
language for structural matching. 1.4.3 Knowledge Discovery Processes in
Astro/Geo Context
1.4.2.2 Example – OntoDM In research and practice of both domains related to sky
EXPO ontology provides concepts for the description and Earth observations, data analysis usually follows
of the scientific experiments on the upper level and similar steps as previously defined methodologies, even
describes hypotheses, assumptions, and results. It also if their application is in a more ad hoc way and termi-
defines elements for descriptions of the experimen- nology differs between particular cases. Usage of a stan-
tal methods for processing of the measured/simulated dard methodology (like CRISP-DM) is quite rare. Such
data, statistical testing of defined hypotheses, or pre- ad hoc implementation could bring problems when an-
diction of the expected results. The OntoDM ontology alysts do not recognize some of the issues usually ad-
extends the description of the experimental method for dressed by methodology. On the other hand, in recent
the application of data mining methods. years it looks like more experienced data scientists are
When applying the data mining method on the data, extending project teams (especially for large projects)
it is necessary to describe input datasets, data mining and their processes evolve closer to the standard appli-
tasks (i.e., if we are dealing with predictive or descriptive cation of KDP.
data mining), and the operation applied on data during While there are of course standard applications of
the preprocessing and methods used for the modeling data mining techniques, from data to extracted knowl-
(i.e., applied algorithms and their parameters settings). edge as a scientific result or engineering output for busi-
The description of data mining algorithms in OntoDM nesses, both domains share one specific type of projects.
covers three different aspects. The first aspect is the data In both astronomy and remote sensing, one of the spec-
mining algorithm specification (e.g., the C4.5 algorithm ified outputs of some data processing process can be a
for decision tree induction), which describes declarative data product. This concept is similar to the data ware-
elements of an algorithm, e.g., it specifies that analysts housing area (Ponniah, 2010), where the company pro-
can use the algorithm for solving a predictive modeling vides data from its infrastructure through data mart –
data mining task. The second aspect is the implementa- client-based structured access to a subpart of the avail-
tion of an algorithm in some tool or software library able data storage, often with rich query and search in-
(e.g., WekaJ48 algorithm implementation). The third terface.
aspect is the process aspect, which describes how to ap- Creation and maintenance of data product can be
ply a particular data mining algorithm to a dataset with a complex process and represents KDP itself, even if it
CHAPTER 1 Methodologies for Knowledge Discovery Processes in Context of AstroGeoInformatics 17

FIG. 1.6 Example of OntoDM instantiation.

needs additional analysis to provide new knowledge. AstroGeoInformatics, it will be even more important to
Moreover, classic KDP is usually extended with lifecycle apply methods for querying data products, set up partic-
management aspects (see Section 1.3.4 for an example). ular processes, and share their understanding well. This
In both astronomy and remote sensing, new projects means that both process modeling and ontology-related
(new large telescopes and satellites with high-resolution aspects might be helpful in the combined effort of data
data) bring data-intensive research to a new level, and scientists from both areas. We will address some of them
even data product projects include almost all steps of in the following subsections.
KDP. For example, raw data of images are often pre-
processed using calibrations, denoising, and transfor- 1.4.3.1 Process Modeling Aspects
mations, and often some basic patterns are recognized From the process point of view, astronomy and remote
(in the modeling step using some machine learning) to sensing are going in a very similar way. The most cru-
provide better and more reliable data products for ana- cial aspect in process modeling is to support workflow-
lysts, who will use the product in data mining for new based approaches to Big Data produced by the instru-
knowledge. ments and their lifecycle management. In the 2020s, it
Therefore we have three different KDP instantiations will be the most critical data analytical process for both
in both domains, i.e., first, full KDP project from raw areas and it will affect any AstroGeoInformatics data
data to extracted knowledge (scientific result or business mining project.
output); second, data product projects, which reduce Regarding the process modeling aspects, there are
traditional KDP at the end of the process, i.e., the final three main lines in both domains:
result is the reliable data source for other analysts; third, • Some types of projects are supported by more indi-
slightly reduced KDP at the start, where some level of vidual software packages, suitable for querying data
data product is used, i.e., some of the preprocessing products, visualization, and modeling, thus provid-
steps are not needed, but usually querying the data is ing simple (often ad hoc or user-centric) process-
still an important step. It seems that the first case, which based analysis of particular data. This approach
splits into the second and third case, will be even more brings two types of tools:
rare with more data coming in the next years in both • general purpose packages – usually some
areas. In the projects combining data and methods for toolkit(s) with a connection to different data
18 PART I Data

products from different projects or data product development and land topography, weather forecast,
sites; a good example is an idea of the virtual ob- and security monitoring. In astronomy, the main focus
servatory, extensively applied in astronomy (see is on scientific results and the domain is quite narrow.
the CDS set of tools10 ), but also in the remote This leads to a more diverse view of tools and stan-
sensing area (see TELEIOS11 ); many of the pack- dards within remote sensing than in astronomy; also
ages support not only virtual observatory query- business-oriented applications force more business pro-
ing functions, but also analysis and visualization; cess modeling tools in remote sensing. On the other
• project-specific packages – especially with large hand, thanks to advancements in Big Data processing
projects, much effort is made to provide users architectures on the industry level, both areas will soon
not only with the data product but also with nec- share similar software and distributed systems as an in-
essary tools which help data scientists with the dustry. This effort will help in the development of any
analysis of the project data, e.g., Sentinel Tool- AstroGeoInformatics projects.
boxes12 in remote sensing.
• Instead of the previous case, which provides a more 1.4.3.2 Ontology-Related Aspects
classical KDP approach (without some standards of The usage of the ontologies for KDP in astronomy or re-
process modeling), an effort in data processing with mote sensing is sporadic, especially if we think about
grid/cloud services and high-performance comput- the complete process and the description of its steps
ing has been made. The most prominent examples as provided in ontologies like OntoDM. It is different
here were related to the execution of workflows, from some other domains, like bioinformatics, where
known under the term “scientific workflow systems” ontologies are also used to understand KDP steps in the
(e.g., DiscoveryNet, Apache Airavata, and Apache analysis itself. In both areas of astronomy and remote
Taverna). Here analysts were able to create (often sensing, ontologies are usually used in the following
through a graphical designer), maintain, and share
cases:
their workflows for processing of data. Again we have
• Light-weight taxonomies for classification of ob-
here both:
jects – the most frequent application of ontolo-
• general purpose systems – tools that are not con-
gies for a better understanding of objects in par-
nected to some specific domain(s); here we can
ticular domains. In astronomy, it means taxonomy
also put many of the classical business workflow
about astronomical objects, data of their observa-
systems;
tion and their properties. It is quite straightforward
• domain- or project-specific systems – again, some
and includes efforts like ontology for virtual obser-
scientific workflow systems more specific to some
domain(s) or project. vatory objects, or space object ontology, or others.
• As an evolution of scientific workflow systems, the In remote sensing, it is more diverse, with domain-
third line can be seen in more advanced platforms, specific semantic models related to the different ap-
which instead of only workflows also support ana- plication domains, e.g., an ontology for geoscience,
lysts with more features related to data science and agriculture, climate monitoring, and many others.
KDP. It means that more traditional scientific work- A more specific domain usually provides some se-
flow systems evolved into e-Science platforms, which mantic standard on how to share data, models,
assist scientists to see the evolution of data, their pro- methods, and application data for its purposes. The
cesses, intermediate results, and also final results and light-weight taxonomies bring specific data formats
potential publication of results. We can see this evo- (e.g., GeoJSON13 in remote sensing for geographic
lution as the advancement of both previous cases, or features), as well as controlled vocabulary for an-
simply as their enhanced combination. notations in order to achieve better understanding,
While previous aspects are shared, there are also composition, and execution of queries on real data.
some differences. Remote sensing provides more do- • Semantic models for processing architectures – some
mains of applications and also has a large number ontology-based standards can be used in the process-
of business-oriented applications, e.g., from science- ing of workflows, especially for the composition of
related topics to applied research in agriculture, urban services. It is not specific for these two areas, but it
generally follows the application of dynamic work-
10 Centre de Données astronomiques de Strasbourg – http://cds.u-
strasbg.fr/.
flow models in Big Data and data stream processing.
11 http://www.earthobservatory.eu/.
12 https://sentinel.esa.int/web/sentinel/toolboxes. 13 http://geojson.org/.
CHAPTER 1 Methodologies for Knowledge Discovery Processes in Context of AstroGeoInformatics 19

Especially in the remote sensing area, ontologies Fayyad, U., Piatetsky-Shapiro, G., Smyth, P., Uthurusamy,
now became interesting due to the effort to better un- R. (Eds.), Advances in Knowledge Discovery and Data Min-
derstand hyperspectral imaging and the combination of ing. AAAI Press, Menlo Park, CA, USA, pp. 37–58.
different (and often dynamic) data products in applica- Cabena, P., Hadjinian, P., Stadler, R., Verhees, J., Zanasi, A.,
1998. Discovering Data Mining: From Concept to Im-
tions. It is also crucial due to large initiatives towards
plementation. Prentice-Hall, Inc., Upper Saddle River, NJ,
the open data; in geodata one of the most prominent USA.
initiatives is the Open Geospatial Consortium,14 which Chapman, P., Clinton, J., Kerber, R., Khabaza, T., Reinartz, T.,
is responsible for the creation, maintenance, and provi- Shearer, C., Wirth, R., 2000. Crisp-dm 1.0 step-by-step data
sion of standards, including some semantic models for mining guide. Technical report. The CRISP-DM Consor-
annotations. This effort for standardization and sharing tium.
of knowledge between application domains can be ben- Chen, H.-M., Kazman, R., Haziyev, S., 2016. Agile big data an-
eficial for any attempt to create an AstroGeoInformat- alytics development: an architecture-centric approach. In:
ics project. Moreover, any attempt to reuse approaches Proceedings of the 2016 49th Hawaii International Confer-
ence on System Sciences (HICSS). HICSS ’16. IEEE Com-
which tries to share more KDP-like information (e.g.,
puter Society, Washington, DC, USA, pp. 5378–5387.
ontologies like EXPO, LABORS, OntoDM) can be bene- Cios, K.J., Swiniarski, R.W., Pedrycz, W., Kurgan, L.A., 2007. The
ficial for analysts and researchers from both domains in Knowledge Discovery Process. Springer US, Boston, MA,
their work on such projects (because even terminology pp. 9–24.
in KDP steps may differ). Ermilov, I., Ngonga Ngomo, A.-C., Versteden, A., Jabeen, H.,
In conclusion, we can say that KDP methodologies Sejdiu, G., Argyriou, G., Selmi, L., Jakobitsch, J., Lehmann,
are a concept which can help especially in the prepa- J., 2017. Managing lifecycle of big data applications. In:
ration of cross-domain models and projects. Such am- Rozewski, P., Lange, C. (Eds.), Knowledge Engineering and
bition is also the case of AstroGeoInformatics projects Semantic Web. Springer International Publishing, Cham,
pp. 263–276.
proposed in this book. In the next chapters, most of the
Evans, J.R., 2015. Business Analytics. Pearson.
methodology steps will be addressed both from an as- Fayyad, U., Piatetsky-Shapiro, G., Smyth, P., 1996. From data
tronomy and a remote sensing point of view; many of mining to knowledge discovery in databases. AI Maga-
them will bring shared ideas, methods, and tools, but zine 17, 37–54.
also describe some differences and specifics. Moreover, Gangemi, A., Guarino, N., Masolo, C., Oltramari, A., Schnei-
some of the examples of AstroGeoInformatics ideas and der, L., 2002. Sweetening ontologies with DOLCE. In:
projects will be shown at the end of the book on se- Knowledge Engineering and Knowledge Management: On-
lected use cases. In the future, such synergy can be sup- tologies and the Semantic Web. Springer Berlin Heidelberg,
ported by the methodologies for KDPs with the ontolo- Berlin, Heidelberg, pp. 166–181.
Grady, N.W., 2016. Kdd meets big data. In: 2016 IEEE Interna-
gies used for better understanding of particular steps,
tional Conference on Big Data. Big Data, pp. 1603–1608.
data, methods, and models. Grenon, P., Smith, B., 2004. Snap and span: towards dynamic
spatial ontology. Spatial Cognition and Computation 4 (1),
69–103.
REFERENCES Herre, H., et al., 2006. General Formal Ontology (GFO):
Anand, S.S., Buchner, A.G., 1998. Decision Support Using Data A Foundational Ontology Integrating Objects and Pro-
Mining. Financial Times Management, London, UK. cesses. Part I: Basic Principles. OntoMed, Leipzig.
Anderson, L.W., Krathwohl, D.R. (Eds.), 2001. A Taxonomy for Hilario, M., Nguyen, P., Do, H., Woznica, A., Kalousis, A.,
Learning, Teaching, and Assessing. A Revision of Bloom’s 2011. Ontology-Based Meta-Mining of Knowledge Discov-
Taxonomy of Educational Objectives, 2 edn. Allyn & Bacon, ery Workflows. Springer Berlin Heidelberg, Berlin, Heidel-
New York. berg, pp. 273–315.
Azevedo, A., Santos, M.F., 2008. KDD, semma and CRISP-DM: Kurgan, L.A., Musilek, P., 2006. A survey of knowledge discov-
a parallel overview. In: Abraham, A. (Ed.), IADIS European ery and data mining process models. Knowledge Engineer-
Conf. Data Mining. IADIS, pp. 182–185. ing Review 21 (1), 1–24.
Beckman, T., 1997, International Association of Science and Lenat, D.B., 1995. CYC: a large-scale investment in knowl-
Technology for Development. A Methodology for Knowl- edge infrastructure. Communications of the ACM 38 (11),
edge Management. IASTED. 33–38.
Bouyssou, D., Dubois, D., Prade, H., Pirlot, M., 2010. Decision Mariscal, G., Marban, O., Fernandez, C., 2010. A survey of
Making Process: Concepts and Methods. ISTE, Wiley. data mining and knowledge discovery process models and
Brachman, R.J., Anand, T., 1996. The process of knowledge methodologies. Knowledge Engineering Review 25 (2),
discovery in databases: a human-centered approach. In: 137–166.
Mizoguchi, R., 2010. YAMATO: Yet Another More Advanced
14 http://www.opengeospatial.org/. Top-level Ontology.
20 PART I Data

Nascimento, G.S., de Oliveira, A.A., 2012. An agile knowl- Rogalewicz, M., Sika, R., 2016. Methodologies of knowledge
edge discovery in databases software process. In: Data discovery from data and data mining methods in mechan-
and Knowledge Engineering. Springer Berlin Heidelberg, ical engineering. Management and Production Engineering
pp. 56–64. Review 7 (4), 97–108.
Niles, I., Pease, A., 2001. Towards a standard upper ontology. Rowley, J., 2007. The wisdom hierarchy: representations of the
In: Proceedings of the International Conference on Formal DIKW hierarchy. Journal of Information Science 33 (2),
Ontology in Information Systems - Volume 2001. FOIS ’01. 163–180.
ACM, New York, NY, USA, pp. 2–9. SAS Institute Inc., 2017. SAS Enterprise Miner 14.3: Reference
Nonaka, I., Toyama, R., Konno, N., 2000. SECI, Ba and leader- Help. SAS Institute Inc., Cary, NC, USA.
ship: a unified model of dynamic knowledge creation. Long Schober, D., Kusnierczyk, W., Lewis, S., Lomax, J., 2007. To-
Range Planning 33, 5–34. wards naming conventions for use in controlled vocabulary
Panov, P., Soldatova, L.N., Dzeroski, S., 2013. OntoDM-KDD: and ontology engineering. In: Proceedings of BioOntolo-
ontology for representing the knowledge discovery pro- gies SIG ISMB. Oxford University Press, pp. 2–9.
cess. In: Discovery Science - 16th International Confer- Smith, B., Ashburner, M., Rosse, C., Bard, J., Bug, W., Ceusters,
ence. Proceedings. DS 2013, Singapore, October 6–9, 2013, W., Goldberg, L.J., Eilbeck, K., Ireland, A., Mungall, C.J.,
pp. 126–140. Leontis, N., Rocca-Serra, P., Ruttenberg, A., Sansone, S.-A.,
Ponce, Julio (Ed.), 2009. Data Mining and Knowledge Discov- Scheuermann, R.H., Shah, N., Whetzel, P.L., Lewis, S., 2007.
ery in Real Life Applications. IntechOpen. The OBO Foundry: coordinated evolution of ontologies to
Ponniah, P., 2010. Data Warehousing Fundamentals for IT Pro- support biomedical data integration. Nature Biotechnol-
fessionals. Wiley, New Jersey, USA. ogy 25 (11), 1251–1255.
Ponsard, C., Touzani, M., Majchrowski, A., 2017. Combining Soldatova, L.N., King, R.D., 2006. An ontology of scientific ex-
process guidance and industrial feedback for successfully periments. Journal of the Royal Society Interface 3 (11),
deploying big data projects. Open Journal of Big Data 3 (1), 795–803.
26–41.
CHAPTER 2

Historical Background of Big Data in


Astro and Geo Context
CHRISTIAN MULLER, DR

2.1 HISTORY OF BIG DATA AND with the flooding of the Nile (Nickiforov and Petrova,
ASTRONOMY 2012). In this respect, the corpus of Egyptian observa-
2.1.1 Big Data Before Printing and the tions led to the Egyptian calendar, which was adopted
Computer Age by Julius Caesar when he became the ruler of Egypt and
Astronomy began in prehistoric times by the observa- which, with a minor revision in the late 16th century,
tion of the apparent motions of the sun, moon, stars, became our current calendar.
and planets in the Earth’s sky. Big Data was not easy to Big Data cannot exist without data preservation.
define before the computer age. According to Zhang and The first steps were the compilation of star catalogues,
Zhao (2015), Big Data is defined by ten characteristics, which began in Hellenistic times when the infrastruc-
each beginning with a V: volume, variety, velocity, ve- ture of the Alexandria museum and library were avail-
racity, validity, value, variability, venue, vocabulary, and able (Thurmond, 2003). Star catalogues not only give
vagueness. This list comes from previous studies of Big the name of stars but also their position. Hipparcos
Data in other domains, such as marketing and quality combined previous Babylonian observations, Greek
control. They consider that the four terms volume, vari- geometrical knowledge, and his own observations in
ety, velocity, and value apply to astronomy. Rhodes, and he was the first to correct his data for pre-
Velocity did not exist before the computer age, as cession, but as his manuscripts are lost, he is essentially
the acquisition of observations and their recording and known by his distant successor, Claudius Ptolemy from
publishing were entirely manual until the printing and Alexandria, whose catalogue, the Almagest, has been
the electric telegraph. preserved (Grasshoff, 1990). The number of stars of the
However, volume was already present in the early de- original manuscript is not absolutely clear; Thurmond
scriptions of the night sky: “as the stars of heaven” is a indicates a number of 1028. See Fig. 2.1.
biblical equivalent of infinite, meaning “which cannot The precision of the observations was sometime un-
be counted.” equal as different observation sites had been used and
Variety corresponds to the different observed ob- refraction was not corrected for. The Almagest became
jects; the ancients had only optical observations, but the main reference until the end of the Middle Ages,
the conditions of these observations differed between when several versions made their way to the Arab world
observer sites. Their eye sight was also probably better and Arab astronomers both added new observations
trained. An ethnological study (Griaule and Dieterlen, and adapted the book to their own epochs. Finally, the
1950) revealed that the cosmogony of the Dogon peo- Almagest came back to the Western world by the Latin
ple in present-day Mali indicated two companions of translation from an Arabic version of Gerard of Cre-
Sirius, four satellites of Jupiter, a ring around Jupiter, mona in 1175. None of the Arabic versions increased
and knowledge of Venus phases. This kind of oral tradi- the number of stars; some, due to the observation lat-
tion will probably be difficult to verify in other ethnic itude, even mention less stars than Ptolemy. The first
groups, as more and more literacy is spreading over new catalogue to appear was endeavored by Ulugh Beg
the entire world and oral transmission is disappear- in Samarkand in the 15th century using only original
ing. observations from an especially designed large obser-
Value corresponds to the progress in general knowl- vatory, correcting the errors made by Ptolemy in con-
edge associated with the observations and to their prac- verting the Hipparcos observations. This time, only 996
tical application in navigation or the calendar. For ex- stars were observed. This catalogue was fully translated
ample, the heliacal rise of Sirius had an extreme impor- in Europe only in 1665, but it was known in the Arab,
tance for the Egyptian calendar related to its coincidence Persian, and Ottoman worlds.

Knowledge Discovery in Big Data from Astronomy and Earth Observation. https://doi.org/10.1016/B978-0-12-819154-5.00011-4 21
Copyright © 2020 Elsevier Inc. All rights reserved.
22 PART I Data

FIG. 2.1 Almagest zodiac in the first complete printing by Petri Liechtenstein (1515), United States Library
of Congress. Printing had two advantages: the multiplication of copies and thus a better dissemination, and
protection against copyist’s errors or “improvements.”

2.1.2 The Printing and Technological accurate mechanical clocks, and the development of
Renaissance Revolution astrology. All these necessitated better star catalogues
The 16th century was marked by three important evo- and planetary ephemerides. At the same time, the print-
lutions: the generalization of open sea navigation using ing technology allowed the diffusion of the astronom-
astronomical positioning techniques, the appearance of ical writings and was followed by a real explosion of
CHAPTER 2 Historical Background of Big Data in Astro and Geo Context 23

the number of publications (Houzeau and Lancaster, were first given by Adam Smith (1778) at the end of the
1887). Printing secured two important elements of Big 18th century. Astrologers would rely on a feeling based
Data: preservation of controlled copies and availability on their knowledge which they could not quantify for
to a larger number of users. everything outside their analysis of the sky. Astrology
Astronavigation was already used in the 15th century became suspected of being linked to superstition dur-
by the Portuguese, Arab, and Chinese navigators, but ing the English Reformation, but luckily, astronomy be-
proved to be very risky during the first intercontinen- came a respected science in Great Britain. For example,
tal explorations. It is in this context that the Ottoman the founder of the London stock exchange, Thomas Gre-
sultan Murad III ordered the construction of a large ob- sham, established Gresham College in the late 16th cen-
servatory in Constantinople superior to the Ulugh Beg tury for the education of the young bankers and traders
observatory and equipped with mechanical clocks. The with the following professorships: astronomy, divinity,
chief astronomer, Taqi ad-Din, wanted to correct the geometry, law, music, physics, and rhetoric. “The astron-
previous catalogues and ephemerides to promote im- omy reader is to read in his solemn lectures, first the
provement in cartography (Ayduz, 2004). He improved principles of the sphere, and the theory of the planets,
and designed new instruments much superior to pre- and the use of the astrolabe and the staff, and other
vious versions. Unfortunately, the observatory was de- common instruments for the capacity of mariners.” This
stroyed in 1570 due to a religious decree condemning program did not make any mention of astrology and its
astrology. use as a predictive tool in commodity trading.
Almost simultaneously, Tycho Brahe equipped a Robert Hooke, who was professor at Gresham col-
huge observatory in Denmark with instruments and not lege, insisted on the use of telescopic observations in
only used up to 100 assistants, but also spent for 30 order to increase the number of stars and their posi-
years about 1% of the total budget of Denmark (Couper tional accuracy, but this important progress was only
et al., 2007), Tycho Brahe was the first to take refrac- initiated by John Flamsteed, the first Astronomer Royal
tion into account and to analyze observational errors. who exceeded the precision of Tycho Brahe’s observa-
His huge accomplishments were transferred to Prague tions and published a catalogue of 2866 stars in 1712
where he became the astronomer of emperor Rodolph (Thurmond, 2003). At that time, a marine chronometer
II and was assisted by Johannes Kepler, who succeeded accurate by one minute in six hours existed and an able
him. Kepler demonstrated the existence of the helio- seaman was for the first time able to determine an ac-
centric system and determined the parameters of the curate position by using the sextant without any other
planetary elliptical orbits using Tycho’s data. The quan- information. Better marine chronometers were progres-
tity of data measured and reduced by Tycho Brahe and sively developed (Landes, 1983), but due to their high
their accuracy were an order of magnitude greater than price, their generalization had to wait until the 19th cen-
what existed before, representing maybe the first in- tury. Flamsteed got a commission to build the Green-
stance of Big Data improving science. wich observatory in close connection with the British
Astrology was the main application of this scientific Admiralty; the accurate chronometers designed by John
project and the tables produced by Kepler. The Rudol- Harrison for this observatory were essential to the ex-
phine Tables are still used by present-day astrologers, ploration of the Southern hemisphere oceans by Cap-
who usually do not have the means to adapt the epoch. tain Cook and his followers.
Astrology at the time was the equivalent of present- Later, in the 18th and 19th century saw the astro-
day business intelligence and was commonly used for nomical observations being extended to the Southern
any kind of forecasts. Galileo taught medical students hemisphere. At the end of the 19th century, the pho-
the art of establishing the horoscopes of their patients. tographic technique allowed to win again an order of
Galileo was in this respect accused in a first inquisition magnitude in the number of stars. At the beginning of
trial of fatalism, which is the catholic sin of believing the 20th century, about 500,000 stars had been iden-
that the future can be certainly known to human in- tified and several catalogues were under development.
telligence (Westman, 2011). At the same time, Lloyd’s The last catalogue before the space age was the Smithso-
of London were determining marine insurance fees by nian Astrophysical Observatory catalogue in 1965, with
the expected technique of inspecting the ships and crew 258,997 stars listed with 15 description elements for
records, but the last judgment was left to astrologers each. The SAO catalogue used electronic data treatment
(Thomas, 2003). Astrology cannot be considered a pre- since the middle of the 1950s and is the first to fully
cursor of Big Data and their role in business intelligence, meet the definition of Big Data given in the first para-
as large-scale statistical treatments of economic data graph.
24 PART I Data

the structuring of these early sources. Aristotle was also


the successor of the Greek natural philosophers who at-
tempted to relate the observations to their causes so that
they could explain them and even attempt forecasts.
Aristotle was the first to describe the hydrologic cycle.
His knowledge of prevailing winds as a function of sea-
son proved to be essential to the conquest of Greece by
the Macedonian army, the Greek islands being unable
to send troops to their allies in the continental cities in
time due to contrary winds. The meteorology of Aristo-
tle covered a wider context than now because it included
everything in the terrestrial sphere up to the orbit of the
moon and thus would have included geology and what
is now called space weather (Frisinger, 1972).
Unfortunately, Aristotle’s efforts were not continued
for long. His successor Theophrastus compiled signs
which in combination could lead to a weather forecast.
These progresses did not prevent most of the population
to attribute weather to divine intervention and when
Christianity and Islam took over, the pagan gods were
replaced by demons. No systematic records of weather
were kept, and present climate historians have to resort
to agricultural records or indications in chronicles. Dur-
ing the Renaissance, the revival of Hippocratic medicine
led physicians to consider the relation between the en-
vironment and health and record meteorological data
again; similarly the logbooks of the ships at sea be-
came more systematic, leading in the 18th century to
the first large set of meteorological data which began to
follow a standardization process, as exemplified by the
FIG. 2.2 Frontispiece of the Rudolphine Tables: Tabulae Societas Meteorologica Palatina (Meteorological Society
Rudolphinae, quibus Astronomicae scientiae, temporum of Mannheim) (Kington, 1974) which started in 1780,
longinquitate collapsae Restauratio continetur by Johannes and established a network of 39 weather observation
Kepler (1571–1630) (Jonas Saur, Ulm, 1627). stations; 14 in Germany, and the rest in other coun-
tries, including four in the United States, all equipped
It is now succeeded by the new efforts based on space with comparable and calibrated instruments: barome-
age techniques and the massive use of large databases ters, thermometers, hygrometers, and some with a wind
which constitute the basis of the BigSkyEarth COST ac- vane. During the 19th century, more meteorological ob-
tion. See Fig. 2.2. servatories were established in Europe, North America,
and in the British Empire. The progress of telegraphic
communications led to consider the establishment of a
2.2 BIG DATA AND METEOROLOGY: A LONG synoptic database of identical meteorological parame-
HISTORY ters measured at different observatories.
2.2.1 Early Meteorology
The study and comparison of large amounts of observa- 2.2.2 Birth of International Synoptic
tions constituted the early base of meteorology. The rep- Meteorology
etition of phenomena proved very early to be less regu- The breakthrough occurred with Leverrier in 1854. Lev-
lar than astronomical events, and even extreme events errier was a French astronomer who reached celebrity by
were the unpredictable action of the gods. The Baby- predicting the position of Neptune from perturbations
lonians and Egyptians compiled a lot of observations of the Uranus orbit. Gelle at the Berlin observatory was
without relating them. A big step forward was the clas- then able to observe the planet at the predicted position.
sifications and typologies assembled by Aristotle and Following a disastrous storm in the Black Sea during
CHAPTER 2 Historical Background of Big Data in Astro and Geo Context 25

the Anglo-French siege of Sebastopol, the French gov- who began by adopting common definitions of the me-
ernment commissioned Leverrier to determine if with teorological parameters. See Fig. 2.3.
an extensive network of stations, the storm could have
been predicted. After analysis, he determined that the 2.2.3 Next Step: Extension of Data
storm had originated in the Atlantic several days be- Collection to the Entire Globe
fore the disaster and that a synoptic network would The distribution of stations of this first network was
have allowed to follow it and to make a raw forecast heavily biased to Western Europe and the Eastern
of its arrival in the Black Sea (Lequeux, 2013). Unfortu- United States. It was clear at the beginning that a real
nately, Leverrier could never assemble the legions of la- network should extend to the entire world, includ-
borers necessary to study the long-term physical causes ing the Southern hemisphere. As a permanent exten-
of weather and climate. His efforts were however the sion was beyond the means of the early International
first steps to the creation of an international synoptic Meteorological Organization, periodic campaigns for
network in parallel to the geomagnetic network already the study of polar regions were proposed by several
developed by Gauss, Sabine, and Quetelet (Kamide and countries, combining exploration and maritime obser-
Chian, 2007). The extension of the geophysical network vations. The first one, in 1883, was concentrated on
to meteorology was rapid due to the establishment of the Arctic ocean and a few sub-Antarctic stations. The
meteorological services in most observatories and the observations took place between 1881 and 1884 and
development of the electric telegraph. These founding demonstrated the feasibility of a network extension.
fathers made an unpreceded effort to internationalize The success of the first campaign led to the second
the effort, and most notably, the Dutch meteorologist International Polar Year in 1932–1933. This campaign
Buys-Ballot, founder of the Royal Dutch Meteorological was initiated and led by the International Meteorolog-
Institute, published the empirically discovered relation ical Organization and extended to geomagnetism and
between cyclones, anticyclones, and wind direction, in- ionospheric studies; more countries participated, and
troducing fluid physics to meteorology and the basis of the program included simultaneous observations at low
future forecasting models (WMO, 1973). latitudes. This campaign should have included a net-
These early networks hardly fit the definition of Big work of Antarctic stations, but the financial crisis of
Data: the telegraphic systems of the different coun- the time limited the funding means of the participat-
tries were not standard, the archiving of the data was ing countries. The collection and use of Big Data was
not uniform, and a lot of parameters were station- or already envisaged by the establishment of World Data
operator-dependant. The exchange of processed data Centers centralizing data by themes.
as hourly averages was not evident. However, around The Second World War extended to the entire North-
1865, the generalization of the Morse telegraphic pro- ern hemisphere and parts of the Southern Pacific. Me-
tocols together with the application of the newly dis- teorological forecasts were essential, and the allies de-
covered Maxwell equations improved the reliability of cided on a very wide synoptic network. This effort was
the telegraph, and regular exchange of data between led by the UK Met Office, which exfiltrated qualified me-
stations became the norm. International meteorologi- teorologists from Norway and several other occupied
cal conferences regularly met, beginning in 1853 at the European countries. The Germans took a more theo-
initiative of the United States Naval Observatory, the retical approach, demanding less stations. The Anglo-
first one presided by the director of the Brussels Ob- American meteorological forecasts, with a better time
servatory, Adolphe Quetelet. Even though fewer than resolution, were essential in planning successful am-
15 countries were represented, no explicit resolutions phibious operations at the end of the war, as well as air
came from this first meeting because any recommen- force support. After the war, the extension of this net-
dation would have led to modifications of the practice work to the Southern hemisphere led to the 1947 US
of the signatories; the wording was very general, e.g., Navy Highjump operation, combining the exploration
“that some uniformity in the logs for recording ma- of Antarctica and the establishment of stations. This ex-
rine meteorological observations is desirable.” Anyway, pedition led to numerous accidents, which confirmed
a process was started, which led in 1873 to the foun- that military claim and occupation of Antarctica were
dation of the International Meteorological Organiza- beyond the means of any nation. Most of these acci-
tion at a Vienna international conference led by Buys- dents were related to errors in the positioning of ships
Ballot (WMO, 1973). This new organization proved to and aircrafts related to the proximity of the South Pole
be strong enough to standardize practices in the entire and weather conditions. The staff of this huge expedi-
world. It established a permanent scientific committee tion included the ionospheric scientist Lloyd Berkner,
26 PART I Data

FIG. 2.3 Early synoptic map of Swedish origin (https://en.wikipedia.org/wiki/Timeline_of_meteorology#/


media/File:Synoptic_chart_1874.png). Sea level pressure is indicated, as well as an indication of surface
winds, demonstrating the success of the International Meteorological Organization at its foundation in 1873.
Until the early 1970s, isobar lines were drafted by hand to fit the results of the stations; it was only in the
last quarter of the 20th century that they were automatically plotted integrating data from other origins as
airplanes and satellites.

who after designing the radio communications of the scientifically active countries (National Academy Press,
1929 Byrd Antarctica expedition multiplied the execu- 1992). See Fig. 2.5.
tive roles in scientific unions while continuing research. The Second World War had seen an increase in
He later played an important role in coordinating elec- the number of weather ships, as these supported also
tronic operations for the US Navy in the Second World transatlantic air traffic. This network was officialized,
and these stations are shown in Fig. 2.4. Unfortunately,
War. His positions as a rear admiral, a presidential ad-
their high cost led to their progressive retirement af-
viser, and the president of the International Council of
ter IGY when their function was taken over by instru-
Scientific Unions (ICSU) helped him to initiate in 1950 mented merchant ships and commercial airliners. Also,
the International Geophysical Year (IGY) project and to beginning in 1960, experimental satellites were devoted
take the first steps of the Antarctic treaty. The purpose of to meteorological observations until evolving into the
IGY was to extend the observations to the entire globe present network of civilian operational meteorological
with the cooperation of the Soviet Union and all other satellites operated by both EUMETSAT and NOAA.
CHAPTER 2 Historical Background of Big Data in Astro and Geo Context 27

FIG. 2.4 The 12 Arctic stations of the 1883 International Polar Year, NOAA, https://www.pmel.noaa.gov/
arctic-zone/ipy-1/index.htm.

FIG. 2.5 Photograph of one of the first preparatory meetings of the IGY at the US Naval Air Weapons
Station at China Lake (California) in 1950. The scientists present around Lloyd Berkner and Sydney
Chapman on this image represent three quarters of the world authorities on ionosphere and upper
atmosphere at the time. A similar group today would include much more than 10,000 participants
(Pr. Nicolet private archive).
28 PART I Data

FIG. 2.6 Extension of the network of WMO stations from a European network in the middle of the 19th
century to the current network. The stations are color-coded to indicate the first year in which they provided
12 months of data (Hashemi, 2009).

EUMETSAT is a consortium of meteorological or- models are run in parallel and in which the final analy-
ganizations regrouping most of Western and Central sis uses statistical techniques (WMO, 2012).
Europe, including Turkey. It operates both its own net-
work of geostationary METEOSAT satellites and METOP
in polar orbit. Since the 2010s, it collaborates with REFERENCES
COPERNICUS Sentinel satellites managed by ESA for Ayduz, S., 2004. Science and Related Institutions in the Ot-
the European Union. The data are used for forecasts by toman Empire During the Classical Period. Foundation for
the European Centre for Medium Range Weather Fore- Science, Technology and Civilisation, London.
casts (ECMWF) to produce forecast maps for the en- Couper, H., Henbest, N., Clarke, A.C., 2007. The History of As-
tronomy. Firefly Books, Richmond Hill, Ontario.
tire world. In 2019 these have a 20 km resolution, and
Frisinger, H.H., 1972. Aristotle and his meteorology. Bulletin
should reach the 5 km resolution during the 2020s. See of the American Meteorological Society 53, 634–638.
Fig. 2.6. https://doi.org/10.1175/1520-0477(1972)053<0634:
The total amount of data coming from all these AAH>2.0.CO;2.
sources is difficult to estimate as the definition of Grasshoff, G., 1990. The History of Ptolemy’s Star Catalogue.
data covers all aspects of the raw and processed data. Springer Verlag.
Currently, COPERNICUS, which is not yet in com- Griaule, M., Dieterlen, G., 1950. Un Systéme soudanais de Sir-
ius. Journal de la Société des Africanistes 20 (2), 273–294.
plete operation, generates about 10 petabytes per year;
Hashemi, K., 2009. http://homeclimateanalysis.blogspot.be/
ECMWF claimed in 2017 to have archived more than 2009/12/station-distribution.html. climate blog. Brandeis
130 petabytes of meteorological data, beginning essen- University.
tially in the 1980s, when EUMETSAT and NOAA data Houzeau, J.C., Lancaster, A., 1887. Bibliographie Générale de
flows started their exponential increase.1 L’Astronomie, Hayez, Bruxelles.
Big Data have clearly become a part of the obser- Kamide, Y., Chian, A., 2007. Handbook of the Solar-Terrestrial
vational database. More and more, Big Data enter the Environment. Springer Science & Business Media.
world of forecasts by techniques as assimilation, where Kington, J.A., 1974. The Societas Meteorologica Palatina:
an eighteenth-century meteorological society. Weather 29,
the model is tuned to minimize the gaps between obser-
416–426. https://doi.org/10.1002/j.1477-8696.1974.
vations and the forecast and the ensemble techniques tb04330.x.
in which a large number of instances of one or several Landes, David S., 1983. Revolution in Time. Belknap Press
of Harvard University Press, Cambridge, Massachusetts.
1 http://copernicus.eu/news/what-can-you-do-130-petabytes-data. ISBN 0-674-76800-0.
CHAPTER 2 Historical Background of Big Data in Astro and Geo Context 29

Lequeux, J., 2013. Le Verrier and meteorology. In: Le Verrier— Westman, R.S., 2011. The Copernican Question, Prognostica-
Magnificent and Detestable Astronomer. In: Astrophysics tion, Skepticism and Celestial Order. University of Califor-
and Space Science Library, vol. 397. Springer, New York, NY. nia Press, Berkeley.
National Academy Press, 1992. Biographical Memoirs, V.61. WMO, 1973. One hundred years of international co-operation
ISBN 978-0-309-04746-3. in meteorology (1873-1973): a historical review. https://
Nickiforov, M.G., Petrova, A.A., 2012. Heliacal rising of Sir- library.wmo.int/opac/doc_num.php?explnum_id=4121.
ius and flooding of the Nile. Bulgarian Astronomical Jour- World Meteorological Organization.
nal 18 (3), 53. WMO, 2012. Guidelines on Ensemble Prediction Systems
Smith, A., 1778. An Inquiry Into the Nature and Causes of the and Forecasting. http://www.wmo.int/pages/prog/www/
Wealth of Nations. W. Strahan and T. Cadell, London. Documents/1091_en.pdf. Publication 1091. World Meteo-
Thomas, K., 2003. Religion and the Decline of Magic: Stud- rological Organization.
ies in Popular Beliefs in Sixteenth and Seventeenth-Century Zhang, Y., Zhao, Y., 2015. Astronomy in the big data era. Data
England. Penguin=History. Science Journal 14 (11), 1–9. https://doi.org/10.5334/dsj-
Thurmond, R., 2003. A history of star catalogues. http:// 2015-011.
rickthurmond.com/HistoryOfStarCatalogs.pdf.
Another random document with
no related content on Scribd:
NAILS.

STEEL WIRE BARBED NAILS are superior to any cut nail made.
Are lighter, stronger and easier to drive. Holding firmer in the wood,
and having a larger flat head, they hold the tin to the roof boards
much better, are nearly double the quantity to the pound, and
therefore fully as cheap.
DEADENING AND FIRE-PROOFING
FLOORS.
This can be accomplished in a
simple and inexpensive manner
by nailing strips 1 × 1½ inches
along the side of each joist, the
top edge of each strip to be two
inches below the top surface of
the joist. On top of these strips
lay a floor of rough boards.
Second-hand material answers every purpose for the floor, as the
boards need not be matched or of even thickness.
On the top of this rough floor spread a covering of coarse mortar
gauged level with the top edge of the joist. When the mortar is dry
lay the floor in the usual manner. It takes a long time for fire to burn
through a floor laid in this manner, and sound is deadened
completely by the mass of non-conducting mortar and the double
floors. For fire-proofing and deadening partitions in an effective way,
build a four-inch brick wall between the studding. A nail driven every
ten or twelve inches in the side of the studding, on the line of the
mortar joint, will hold the wall firmly between the studding. Where
brick is scarce or expensive, sun dried brick made from stiff clay is a
good substitute. They can be made 12 × 12 inches, and thick
enough to come within ⅜ of an inch of the lathing. The mortar keys
impinging against the wall make the plastering firmer and
consequently stronger.
The expense of fire-proofing and deadening in this manner is so
small, it is surprising it is not more generally adopted. If the flooring
is laid before the mortar used between the joists is dry, it will cause
the boards to swell and may bulge them; and when they finally
become dry will leave ugly looking cracks where they join.
Remember, you cannot have a well built house without giving Nature
time to do her work.
DIRECTIONS FOR LAYING WALTER’S
PATENT METALLIC SHINGLES.

Cut I.

Shows manner of finishing with Climax Ridge Stop


Block and Gable End Finish.

Commence at the lower left-hand corner. In starting be particular to see that


you start straight with the eaves of the building. To do this it is best to draw a
chalk line about twelve inches from the eaves; this distance leaves two inches
to project from the eaves, which in many cases is more than sufficient. If you
use our Gable End Finish (Cut I) it saves the trouble of fitting the shingles to
the verge board, and adds to the appearance of the roof. After nailing the
Gable End Finish to its place, press the left-hand edge of the first shingle well
under the fold of the Gable End Finish, and before nailing it hook two or three
shingles with the top edge on a line with your chalk line; then remove the
loose shingles, and nail the first one which is held to its proper place by the
Gable End Finish. By doing this in starting each course you are sure of a
straight line if you follow correctly the gauge lines at the top of each shingle.
Cut A.

Shows commencement of first two


courses.

If your roof has a pitch of six inches to the foot, or steeper, let the
bottom edge of the shingle rest just above the lower gauge line on the
top of the underlying shingle, as shown in Cut A. But if the pitch of the roof
is less than six inches to the foot, let the shingle entirely cover the lower
gauge line, as shown in Cut B.
Cut B.

We advise against the use of our shingles on roofs of less pitch than five
inches to the foot, unless the lap is increased. As a rule, we think any roof that
can be walked over with safety is too flat for shingles of either wood or metal.

Cut F.
Should there be a gutter formed at the eave, let the shingle rest on it as you
would in using the ordinary wood shingle. (Remember every other course
commences with a half shingle, as shown in Cuts A and B.) The same
rules that govern the laying of slate or the common wood shingle along
valleys, or about chimneys and dormer windows, are applicable to ours,
except the tin shingles are bent up against the sides of chimneys, which
cannot be done with wood or slate. When our shingles turn up against the
sides of chimneys or brick walls, insert flashing by sawing out the mortar joint
above the line of tin work; where the upper edge of our shingles butts against
the brick wall, as they do on the lower side of chimneys, cut them off on the
line where the chimney comes to the roof, and use a strip of tin bent in this
manner. (See Cut F.) The upper end at A is to fit in mortar joint. The lower
edge, B, is nailed to the sheathing before the shingles are put on. The upper
ends of shingles are then to be pressed up under the fold, C. Great care
should be used in finishing about chimneys and dormers, the details of which
cannot well be explained to suit each case; but a workman of ordinary skill can
suggest the proper manner in which the work should be done to secure
thoroughly tight work.
It is much easier to secure this result with the use of our shingles than it is
with either wood or slate.
Where the upper end of shingle butts against the side of a frame house, use
the same means as on the lower side of chimneys, only let there be no bend
at the point A, as shown in Cut F; but let it extend an inch or so up under the
weather boarding. Where the weather boarding is vertical there is no way of
making tight work but to put the tin work, as before described, back of the
vertical weather board.
In laying the valley, cut the tin so it extends to about one-half inch over the
lock, and bend it under, as shown in Cut D. We furnish to each customer a
small pair of hand-tongs, which is handy to turn this edge over and pinch it
together after the shingle is laid.
Cut D.

The Cut D represents the shingles laid to and from the valley. In starting
from the valley it is best to hold several shingles together, or tack them at the
top, then with a straight-edge mark and cut where they overlap the valley; and
with the hand-tongs edge and lock them to the valley, as shown in cut D.
Use the Hip Coping by nailing the edges to the roof boards, and press the
shingle up under the folds on each side after they are cut to suit the angle of
the hip. (See Cut E.)
Cut E.

As plain as this appears, we have known men to nail the Hip Coping
through the folds and on the top of the shingles. We are, therefore, particular
to say wherever this Hip, or our Plain Ridge Coping (which is the same thing)
is used it should be nailed to the roof boards before the shingles are put on.
The fold is made expressly to receive the edge of the shingles. Cut E shows
this coping and the manner of applying it.
We desire to impress upon our customers, who live in the Northern States,
where blizzards and severe snow storms are frequent, the necessity of using
close sheathing, and if the sheathing is not close, the use of sheathing
paper, to be laid underneath the shingles; it adds greatly to the warmth of
the house in winter, and prevents small particles of snow from entering, it
costs but little, and should always be used under wood, slate or tin shingles
where the best protection is desired.
DO NOT Hammer Down the Joints or Lock.
PAINTING DEPARTMENT
PAINTING.
After the mason, bricklayer, carpenter, tinsmith and plasterer
comes the painter. At this stage of progress in the work, the owner
usually thinks of moving in; in fact, he thinks he should have been
living in his new house several weeks ago.
If weather conditions have been good, the representatives of other
branches of house building have got along smoothly excepting a little
shaking up now and then, caused mostly by the men occasionally
yielding to the seductive influences of Old John Barleycorn or his
rival King Growler. The owner, having previously engaged or
contracted for inside painting, now calls on the painter. The man of
colors and brushes is always ready to promise quick work (and we
are free to say the mysteries of his trade are equal to any reasonable
emergency), and is apt to humor his employer’s impatience, and
meekly submits to his wishes knowing that his employer’s hurry is
his gain. House-builder, did it ever occur to you that first-class inside
painting is a slow process? Think of this; after the first coat is laid, it
should stand long enough to become thoroughly dry, dry enough to
be sand-papered. The second coat should be a little heavier than the
first or priming coat. It does not dry as fast as the first. (We are
speaking of good materials properly proportioned.) Each coat must
be thoroughly dry before it is sand-papered. Any kind of good work
requires three coats. Extra good work—four, besides varnishing.
If you succeed in doing this work inside of four weeks you are
fortunate. Thus you see it requires time to do good work. Time for
the men to do theirs, and time for Nature to do hers. Sometimes
Nature frowns and lowers a curtain of humidity for days at a time,
which prevents the oils from drying. It is possible for your painter to
finish all four coats in as many days; but if he is honest and you have
not hurried him unreasonably, his conscience will suffer. Good
painting, like other good things, is durable, looks well, and is the
cheapest in the end.
Good painters pride themselves on doing good work. So if you
desire first-class work, you should give the painter a reasonable
price, and sufficient time in which to do the work properly.
There is doubtless more adulteration in paint than in any materials
used in house-finishing. So general is this the case that it is often
difficult to get pure materials. This is caused by the low price of the
products of petroleum, the earth materials used to adulterate mineral
paints, and the demand for cheap work.

ATTENTION, ROOFERS!
Our Broad Rib Roofing and Steel Plate Shingles are not
protected with a coat of richer metal like those made of tin or terne
plate, and consequently their durability depends upon the paint with
which they should be covered.
For this purpose nothing, so far as we know, is equal to a mixture
of pure unboiled linseed oil and the oxide of iron. The Shingles
should have two coats, one before they are put on, on both sides
(this we do), the other immediately after they are laid. This last coat
should be repeated every four or five years.
Remember, it is the rust-preventing qualities of linseed oil,
combined with the oxide of iron, that makes steel or iron sheets
resist the corrosive action of oxygen, which is ever present in the
atmosphere. (See page 101.)
A PERFECT ROOF.
The best can only be obtained by using good materials, worked
into the best form that theory and experience can suggest.
WALTER’S PATENT “STANDARD” and
COOPER’S PATENT “QUEEN ANNE”
METALLIC SHINGLES AND SIDING,
made from Tin Plate, Steel Plate, Galvanized Tin Plate, Bronze
Metal, and Copper, are unrivaled for roof covering.
Because they have a perfect lock, using no cleats, springs or
exposed seams.
Because they have no exposed fastenings; all nail heads are
covered as perfectly as those used for wood shingles.
Because they can be applied without the necessity of soldering, and
with no other tools than a hammer, a small pair of hand-tongs
and tin shears.
Because they break joints by starting each alternate course with a
half shingle, thus bringing the centre of the bottom of each
shingle astraddle of the locked shingle below, securely binding
the same and preventing rattling.
Because they are easily and rapidly put on, requiring no odd pieces
at the eaves, gables or comb.
Because they lessen the expense of insurance, which extends not
only to the building but to furniture and goods contained therein.
Because they do not crack, split, warp, fall off, rust or burn.
Because there are no cross seams or joints where rain or moisture
can settle and cause decay.
Because we furnish with our shingles, at moderate cost, Valleys,
Gable Strips, Ridge and Hip Coping, that not only improves the
finish of exterior, but assists the workman to an extent that
lessens the cost of applying.
Manufactured by
THE NATIONAL SHEET METAL ROOFING CO.,
510 to 520 East 20th Street, New York City,
and for sale by Dealers in all the principal cities of the United States
and Canada.

From BARTLETT HARDWARE CO.,


Freeport, Ill., January 10th, 1888.
The National Sheet Metal Roofing Co., New York City.
Dear Sirs:—I send you photo. of my house that you may
see how handsome a Roof the 7 × 10 Shingles make. You
thought the 10 × 14 would have been better, but this is the
finest roof in this country, and we expect to sell some the
present year. Everybody admires it. Please quote for the
coming season’s trade.
Very respectfully,
F. BARTLETT, President.
WHAT OUR GOODS ARE.
In offering our goods to the public, we desire to say, briefly, that
the Walter’s Patent Metallic Shingles and Siding Plates have now
been before this country for nearly eight years, upwards of twelve
thousand buildings have been covered with them in the United
States and Canada, and we feel warranted in referring to the owners
of every building so covered.
Our shingles have a concealed nailing flange the entire length of
the shingle on one side, and a perfect lock with a concealed gutter at
the side of the nailing flange that provides for expansion and
contraction, with sufficient ventilation to prevent sweating on the
under-side, (causing rust,) so common in the cross seams of flat lock
and standing seam, now in use.
Our Galvanized (Re-dipped) Shingle is our Standard Tin Shingle
galvanized; and, as every square has twenty pounds of zinc coating,
in addition to the first coat of tin, it will be seen why it is superior to
any galvanized iron made.
Our “Old Process” prices are designed to cover such grades of
double-coated plates as “Old Style,” “M.F.,” “Phelp’s Triple Plate,”
“Gilbertson’s Old Method,” and “Tregoning Old Process,” all superior
to the Standard Grades, and higher in price. We quote all but
Galvanized, painted both sides. If any size or kind is ordered
unpainted, the price will be 25 cents per square less.
Our Steel Plate Shingles are made from sheets of Bessemer steel
rolled as smooth as tin plate, they are painted the same as the
Standard Shingles, and when laid on a roof cannot be distinguished
from them. They should be painted every five years with pure linseed
oil and oxide of iron (brown mineral). They will not bear neglect; in
this respect they differ from shingles made from tin plate. The price
is somewhat less, but the greater durability of the Tin Shingles is
worth more than the difference in price.
We make our shingles in four sizes: The smallest, 7 × 10,
(showing an exposed surface, after laid, of 5½ × 9 inches,) is
designed for Mansards, Spires, Siding, and other upright work. The
10 × 14, (exposed surface, 8¼ × 11½ inches,) is the popular size;
and while it works to the best advantage on complicated roofs, it is
equally good for plain ones. The 14 × 20, (exposed surface, 12 ×
17½ inches,) is used largely on the plainer roofs, where a saving in
price and labor of putting on is an item. The 20 × 28, (exposed
surface, 17½ × 25½ inches,) is made from heavier metal, and is
designed for warehouses, and other large surfaces; and any of the
larger shingles can be used for siding equally as well.
We are prepared to make any of the above sizes from any special
brand of tin plate in the market, but for our Standard goods, which
we carry in stock, we use a first-class grade of roofing tin plate,
warranted perfect.

SAMPLES.
We will send free, and charges paid, samples of our Walter’s
Patent Shingles (three pieces) to any address in the United States
on receipt of five two-cent stamps; this does not pay us one-half the
expense, but we propose to make the cost as light as possible to
those interested in building.
THE NATIONAL SHEET METAL ROOFING CO.,
510 to 520 East 20th Street,
New York City.
BROAD RIB ROOFING,
Like Our
Shingles, has a
nailing flange the
entire length of each
sheet. It forms a
Continuous Rib
when applied from
eaves to comb with
cross seams twenty-
five inches apart,
which gives it
strength and rigidity
not found in roofing
where longer sheets
are used. It has no
protruding anchors,
cleats or rivets; all
nail heads are
covered as perfectly
as those used for wood shingles. There is no waste in using, either
at comb or eaves. The Side Nailing Flange admits of the roof being
fastened to roof boards with any required degree of strength. The
Cross Seams are formed with a folded lock which allows for
contraction and expansion. Material.—Each sheet of this roofing is
made from Best Roofing Tin, or Imported Siemen’s Sheet Steel,
which is hard rolled, tough, durable and thoroughly coated on both
sides with the best Oxide of Iron and Linseed Oil Paint. The size of
sheets are 20 × 28 inches. Tools.—We send with each order one
pair of Metal Shears, one Mallet, and a Draw Tool, for which we
make a charge of two dollars. These tools can be returned to us by
express at our expense, and the purchaser credited, or the money
paid for them returned. Important Improvements in the
manufacture of Steel Sheets enable us to put this on the market at
the same price as Sheet Iron Roofing, to which it is greatly superior.
The Rib is formed of the Walter’s Patent Lock, and is the same we
have used on our shingles for the last seven years. This Lock has
been thoroughly tested, and it is significant that over 25,000,000
square feet of shingles constructed upon this plan are already in use
in the United States and Canada. The application of this Lock does
away with a very large proportion of the work necessary in laying
standing seam roofs as ordinarily applied, such as tonging up the
seams, double seaming, capping or riveting the ribs, and making and
using cleats or other fastenings. This roofing costs no more than
wood shingles in sections where good timber is scarce. Each sheet
is made from plates 20 × 28 inches. WRITE FOR PRICES.

The National Sheet Metal Roofing Co., 510-520 E. 20th St., N. Y.


City.
THE “ATTIC” SKY-LIGHT AND
VENTILATOR.
This is a novelty in roof
ventilation, and possesses the
advantage of admitting light as
well as ventilating rooms in the
garret or immediately under the
roof.
It can be used on any kind of a slope roof having a pitch of three
inches or more to the foot. The front is of double strength glass
which is held to its place by a single catch. When used as a
ventilator the glass is drawn back to the slant of the roof, and sheds
what water would enter the opening during a rain storm. All the
carpenter work necessary in setting the “Attic” is to saw a square
hole through the roof boards at the time of putting on the roof-
covering. We make them to form one plate of our Broad Rib Roofing
with which it interlocks. As shown on page 55, they can be used
equally as well on roofs to be covered with slate or wood shingles as
with our Tin Shingles.
HOW TO SET A FINIAL.
Bore a hole in the upper end of the centre-staff and
insert in it the lower end of finial rod; when the
centre-staff is put to its place, plumb the finial rod by
moving the lower end of the centre-staff before it is
nailed to the floor or cross timbers. (See page 74.)

You might also like