Professional Documents
Culture Documents
EN116 Web
EN116 Web
EN116 Web
ERCIM
www.ercim.eu
NEWS
Special theme:
Transparency
in Algorithmic
Decision Making
Research and Society:
Ethics in Research
Contents
Ethics in Research
by Claude Kirchner (Inria) and James Larrus (EPFL) tribute scientifically or technologically, to develop social
projects, to elaborate rules (regulations, codes) that apply to
Science is in revolution. The formidable scientific and it, to be able to contest bad or unwise approaches, or to be
technological developments of the last century have dra- able to alert public opinion when necessary” are gaining visi-
matically transformed the way in which we conduct sci- bility and, we hope, the necessary impact.
entific research. The knowledge and applications that sci-
ence produces has profound consequences on our To contribute to raising awareness, this section, jointly coor-
society, both at the global level (for example, climate dinated by ERCIM and Informatics Europe, features four
change) and the individual level (for example, impact of contributions on “ethics in research”, focusing on infor-
mobile devices on our daily lives). These developments matics and mathematics and the possible interactions with
also have a profound impact on the way scientists are other scientific disciplines. The contributions develop the
working today and will work in the future. In particular, following topics:
informatics and mathematics have changed the way we • how informatics, logic and mathematics can help to under-
deal with data, simulations, models and digital twins, stand ethics in machine learning techniques;
publications, and importantly, also with ethics. • why and how reproducible research should be central to
research developments;
Ethics in research has been a field of inquiry since antiquity, • education in ethics and scientific integrity is now recog-
in particular in health science and more recently in physics. nized as a priority, but how should we organize the train-
Today all scientific disciplines, from medicine to biology, ing for doctoral candidates;
humanities, informatics, mathematics, physics and chemistry • how shall we organize research to avoid waste and bias in
are troubled by ethical issues. They have been particularly the exploration of scientific knowledge?
popularized by dilemmas posed by the rise of machine
learning, AI, and autonomous mechanized entities such as These contributions address a small but important part of the
drones, robots, and vehicles. overall ethical issues, and we believe it is important and
urgent to set up a joint working group between ERCIM and
In the fields of informatics and mathematics, there is a surge Informatics Europe, to contribute to the development of
of interest in understanding and studying emerging ethical ethics research in our fields. We invite scientists to join the
issues and methods for teaching them. Scientists are cur- initiative by contacting us.
rently engaged in dialogues about these ethical issues with
peers as well as the general population. Professional organi- Link:
sations, such as ERCIM, Informatics Europe, ACM, IEEE [L1] https://www.montrealdeclaration-responsibleai.com
and IFIP, are discussing these issues and many advisory doc-
uments are being produced at the national and international References
levels. Let us point to a few examples of such contributions [1] AIHLEG. Draft ethics guidelines for trustworthy AI
[3, 2, 1]. [online]. December 2018. https://kwz.me/hdq
[2] CERNA. Research Ethics in Machine Learning.
Of course, ethics in research is fundamental, but it is only Research report, February 2018.
one of the corner stones for the investigation of the ethical https://hal.archives-ouvertes.fr/hal-01724307.
consequences of the impact of sciences, technologies, [3] J. Larus, et al.: “When computers decide: European rec-
usages, and innovations on the digital evolutions. Ethical ommendations on machine-learned automated decision
issues should be investigated everywhere in the world and in making”. Technical report, New York, NY, USA, 2018.
particular in Europe to allow the appropriation by everyone, https://dl.acm.org/citation.cfm?id=3185595/.
person, organization, and company of the digital advances
and transformations that are arising. We are indeed in a situa- Please contact:
tion similar to bioethics in the 1980’s with the necessity to set Claude Kirchner, Inria, France
up ethical committees to consider the consequences of the claude.kirchner@inria.fr
digital innovations and transformations. Initiatives like the
Montreal Declaration [L1], addressing “the responsible James Larus, EPFL, Switzerland
development of artificial intelligence, whether it is to con- james.larus@epfl.ch
The tuning of algorithms with (often heavily biased) real-life Software environment control and reconstruction:
data involves lengthy and computationally intensive adjust- Notebooks and workflows only track limited information
ments of hyper parameters and the classical training/testing regarding the software environment they build on. Without
procedures are often too costly to be rigorously followed and rigorously preserving this environment, the notebook will
statistically rigorous. Furthermore, publishing
results never requires a full disclosure of all the
work, especially when trade secrets or confiden-
tial data is at stake. As a consequence many
research results that may have worked in a spe-
cific restricted context turn out to be very difficult
to reproduce by other independent researchers.
In recent years, many researchers, scientific institutions, Accelerating Reproducible Research with a Massive
funding agencies, and publishers, have started to initiate a Open Online Course
change in our research and publication practices. Many con- Carrying out reproducible research is obviously not a one-
ferences have set up artifact evaluation procedures and the size-fits-all effort. Mastering all these technologies and inte-
ACM has proposed some reproducibility badges. The grating them in a daily methodology requires time. To get a
European Research Council, the National Science majority of researchers started, we have designed with the
and Integrity Training After more than two years’ experience and over 1000 trained
doctoral candidates, the following points are worth high-
for Doctoral Candidates: lighting:
• It is not necessary for a trainer to be an ethics “expert”.
face-to-face is better! Nevertheless they must become familiar with the con-
cepts, ask themselves about their own research practice
by Catherine Tessier (Université de Toulouse) and pay attention to scientific news and associated issues.
Indeed a trainer enriches the course with their own experi-
The University of Toulouse and Inria have set up face-to- ence and thinking.
face training in research ethics and integrity for doctoral • A trainer should not be afraid of disturbing issues, difficult
candidates based on debates about their own theses. situations, conflicting points of view, and where appropri-
ate should be able to cope with emotion. Indeed sensitive
In order to meet the requirements of 2016 French legislation issues are likely to be raised by doctoral candidates.
about doctoral studies, face-to-face training in research • Classes must be composed exclusively of doctoral candi-
ethics and integrity for doctoral students has been conducted dates so that the issues that they raise can be addressed
at Toulouse University since 2016 and at Inria since 2017. without the pressure of “senior” colleagues.
The training programme is based on the work [1] of CERNA, • In order for doctoral candidates to talk freely, it should be
the French ethics committee for research in ICTs. announced at the beginning of the session that everyone
(both the trainers and the doctoral candidates) should com-
Based on the conviction that doctoral candidates should be mit to confidentiality; furthermore a doctoral candidate
able to express themselves and question and discuss issues of may remain anonymous when raising an issue about their
research ethics and integrity that directly concern themselves
Please contact:
Judith ter Schure, CWI, The Netherlands
+31 20 592 4086
Judith.ter.Schure@cwi.nl
Large study series are more likely when they include initial
promising results within the series than when they include very
disappointing ones, just like the availability of the big fish in the
cartoon depends on specific smaller fish available.
Image: Bobb Klissourski (Shutterstock)
The past decade has seen increasing deployment of powerful automated decision-making systems in set-
tings ranging from smile detection on mobile phone cameras to control of safety-critical systems. While
evidently powerful in solving complex tasks, these systems are typically completely opaque, i.e. they pro-
vide hardly any mechanisms to explore and understand their behaviour and the reasons underlying their
decisions. This opaqueness raises numerous legal, ethical and practical concerns, which have led to initia-
tives and recommendations on how to address these problems, calling for greater scrutiny in the deploy-
ment of automated decision-making systems. Clearly, joint efforts are required across technical, legal,
sociological and ethical domains to address these increasingly pressing issues.
Machines are becoming increasingly Expert Group on Artificial Intelligence in principle, fully transparent individu-
responsible for decisions within soci- (HLEG-AI) has been to prepare a draft ally, their complexity frequently ren-
ety. Various groups, including profes- report on ethics guidelines for trust- ders them unusable for understanding
sional societies in the area of informa- worthy AI [4]. These documents pave the decision making complexity.
tion technologies, data analytics the way towards a more transparent and
researchers, industry, and the general sustainable deployment of machine- We should consider the main question
public are realising the power and driven decision making systems. On in the field to be: what is an explana-
potential of these technologies, accom- the other hand, recent studies have tion? This question in itself illustrates
panied by a sense of unease and an shown that models learning from data how new this research topic is. As yet
awareness of their potential dangers. can be attacked to intentionally provide there is no formalism for an explana-
Algorithms acting as black boxes make wrong decisions via generated adver- tion and nor is there a way to quantify
decisions that we are keen to follow as sarial data. In spite of a surge in R&D the grade of comprehensibility of an
they frequently prove to be correct. Yet, activities in this domain [5], massive explanation for humans. The following
no system is fool proof. Errors do and challenges thus remain to ensure auto- works are pioneers in this area, creating
will occur no matter how much the mated decision making can be account- fertile ground for innovation.
underlying systems mature and ably deployed, and the resulting sys-
improve due ever more data being tems can be trusted. Ricardo Guidotti and his colleagues
available to them and ever more power- provide a very good introduction into
ful algorithms learning from them. The articles collated in this special this black-box problem and
theme provide an overview of the range approaches to mitigate it. They also
This awareness has given rise to many of activities in this domain. They dis- propose that explanations be broken
initiatives aiming at mitigating this cuss the steps taken and methods being into two levels, namely a local level
black-box problem, trying to under- explored to make AI systems more explanation on a data instance level,
stand the reasons for decisions taken by comprehensible. They also show the which is subsequently combined on a
systems. These include the "ACM limitations of current approaches, global level by synthesising the local
Statement on Algorithmic Transparency specifically as we leave the domain of explanations, and optimising them for
and Accountability" [1], Informatics analysing visual information. While simplicity and fidelity.
Europe's "European Recommendations visualisation of deep learning networks
on Machine-Learned Automated for image analysis in the form of heat On the other hand, as research has fre-
Decision Making" [2] and the EU's maps, attention maps and the like [6,7] quently pointed out over decades, even
GDPR regulation [3] which introduces, has helped drastically in understanding human experts find it hard or impossi-
to some extent, a right for all individu- and interpreting the regions most rele- ble to provide clear, transparent expla-
als to obtain "meaningful explanations vant in image classification, other nations for their decisions. Fabrizio
of the logic involved" when automated domains are frequently reverting to Falchi in his paper highlights the
decision making takes place. One of the extracting rules as surrogates for or importance of lessons to be learned
first activities of the newly established explanations of more complex machine from the field of psychology in under-
European Commission’s High Level learning models. While such rules are, standing intuition and how these can
Explainable AI is an essential component of a “Human AI”, i.e., an AI that expands human experience,
instead of replacing it. It will be impossible to gain the trust of people in AI tools that make crucial
decisions in an opaque way without explaining the rationale followed, especially in areas where we do
not want to completely delegate decisions to machines.
On the contrary, the last decade has wit- understand and validate its decision fact, only the decision behaviour of the
nessed the rise of a black box society rationale. This limitation impacts not black box can be observed. As dis-
[1]. Black box AI systems for automated only information ethics, but also played in Figure 1, the BBX problem
decision making, often based on accountability, safety and industrial lia- can be further decomposed into:
machine learning over big data, map a bility [3]. Companies increasingly • model explanation when the explana-
user’s features into a class predicting the market services and products with tion involves the whole (global) logic
behavioural traits of individuals, such as embedded machine learning compo- of the black box classifier;
credit risk, health status, etc., without nents, often in safety-critical industries • outcome explanation when the target
exposing the reasons why. This is prob- such as self-driving cars, robotic assis- is to (locally) understand the reasons
lematic not only for lack of trans- tants, and personalised medicine. How for the decision of a given record;
parency, but also for possible biases
inherited by the algorithms from human
prejudices and collection artifacts
hidden in the training data, which may
lead to unfair or wrong decisions [2].
assumptions suggest a two-step, local- variety of alternative technical solutions Please contact:
first approach to the BBX problem: along different dimensions: the variety Riccardo Guidotti, ISTI-CNR, Italy
• Local Step: for any record x in the set of data sources (relational, text, images, +39 377 9933326,
of instances to explain, query the etc.), the variety of learning problems riccardo.guidotti@isti.cnr.it
black box to label a set of synthetic (binary and multi-label classification,
examples in the neighbourhood of x regression, scoring, etc.), the variety of Anna Monreale, University of Pisa,
which are then used to derive a local languages for expressing meaningful Italy
explanation rule using an inter- explanations. With the caveat that +39 328 2598903,
pretable classifier (Figure 2 top). impactful, widely adopted solutions to anna.monreale@unipi.it
• Local-to-Global Step: consider the the explainable AI problem will be only
set of local explanations constructed made possible by truly interdisciplinary Dino Pedreschi, University of Pisa,
at the local step and synthesise a research, bridging data science and AI Italy
smaller set by iteratively composing with human sciences, including philos- +39 348 6544616,
and generalising together similar ophy and cognitive psychology. dino.pedreschi@unipi.it
explanations, optimising for simplici-
ty and fidelity (Figure 2 bottom). This article is coauthored with Fosca
Giannotti, Salvatore Ruggieri, Mattia
The most innovative part is the Local-to- Setzu, and Franco Turini (KDDLab,
Global (L2G) Step. At each iteration, ISTI-CNR Pisa and University of Pisa).
L2G merges the two closest explana-
tions e1, e2 by using a notion of simi- Links:
larity defined as the normalized intersec- [L1] https://kwz.me/hd9
tion of the coverages of e1, e2 on a given [L2] https://kwz.me/hdf
record set X. An explanation e covers a [L3] http://ec.europa.eu/justice/data-
record x if all the requirements of e are protection/
satisfied by x, i.e., boundary constraints,
e.g. age > 26. L2G stops merging expla- References:
nations by considering the relative trade- [1] F. Pasquale: “The black box
off gain between model simplicity and society: The secret algorithms that
fidelity in mimicking the black box. The control money and information”,
result is a hierarchy of explanations that Harvard University Press, 2015.
can be represented by using a dendro- [2] R. Guidotti, et al.: “A survey of
gram (a tree-like diagram, Figure 2 methods for explaining black box
bottom right). models”, ACM Computing Surveys
(CSUR), 51(5), 93, 2018.
We argue that the L2G approach has the [3] J. Kroll, et al.: “Accountable
potential to advance the state of art sig- algorithms”, U. Pa. L. Rev., 165,
nificantly, opening the door to a wide 633, 2016.
In recent years, expert intuition has been a hot topic within the discipline of psychology and decision
making. The results of this research can help in understanding deep learning; the driving force behind
the AI renaissance, which started in 2012.
“Intuition is nothing more and nothing skeptics. However, after six years of • System 1: fast, automatic frequent,
less than recognition” [1], is a famous discussion, a believer, Gary Klein, and a emotional, stereotypic, unconscious.
quote by Herbert Simon, who received skeptic, Daniel Kahneman, wrote an • System 2: slow, effortful, infrequent,
the Turing Award in 1975 and the Nobel important paper in 2009 whose subtitle logical, calculating, conscious.
Prize in 1978. As explained by Daniel was “A failure to disagree”. Trying to
Kahneman, another Nobel Prize winner, answer the question When can we trust It is not by chance that DeepMind
in his book Thinking, Fast and Slow [2], intuition? they agreed on a set of condi- AlphaGo, the program that in 2016
and during his talk at Google in 2011 tions for trustable intuitive expertise. defeated South Korean professional Go
[L1]: “There is really no difference Among these conditions, the most player Lee Sedol, combines DL with
between the physician recognising a important ones are: Monte Carlo tree search. As Michael
particular disease from a facial expres- • an environment that is sufficiently Wooldridge, chair of the IJCAI Awards
sion and a little child learning, pointing regular to be predictable; Committee said, “AlphaGo achieves
to something and saying doggie. The • an opportunity to learn these regular- what it does through a brilliant combi-
little child has no idea what the clues are ities through prolonged practice. nation of classic AI techniques as well
but he just said, he just knows this is a as the state-of-the-art machine learning
dog without knowing why he knows”. I believe most of the researchers techniques that DeepMind is so closely
These milestones should be used as a working on DL would agree that those associated with”. Following our
guideline to help understanding deci- are also good conditions for the ques- metaphor, AlphaGo is a good example
sion making in recent AI algorithms and tion: When can we trust deep learning? of collaboration between System 1 and
thus their transparency. In fact, in order for a DL method to System 2. AlphaGo uses DL to provide
learn, we need a large training set (pro- an intuitive estimation of the likelihood
Most of the recent progress in artificial longed practice) and this set must be that the next stone will be placed in a
intelligence (AI) has been on recognition representative of the application sce- specific place and of the final outcome
tasks, and this progress has been nario in which the environment must be of the game given the current status.
achieved through the adoption of deep sufficiently regular to be predictable. However, the final decision about
learning (DL) methods. The AI renais- where to put a stone is made using the
sance started in 2012 when a deep neural What degree of transparency can we ask Monte Carlo tree search (AlphaGo
network, built by Hinton’s team, won the for from DL methods? Following the System 2). In other words, AlphaGo
ImageNet Large Scale Visual metaphor between DL and intuition, we uses the outcome of artificial intuition
Recognition Challenge. Deep learning can look at what Simon said about implemented using DL methods (its
methods have been, and still are, the driv- human recognition capabilities: “we do System 1) but takes decisions with log-
ing force behind this renaissance. Like not have access to the processes that ical reasoning (its System 2).
the little child mentioned by Kahneman, allow us to recognise a familiar object
a state-of-the-art deep neural network is or person”. I believe the same is true for The few examples discussed here show
able to look at something and say “dog- DL. Even if we can monitor the flow of that psychology can help in under-
gie”, without knowing why it knows. In information in a deep neural network, standing AI. When Simon was con-
other words, the task of recognition, we don’t understand the “process”. ducting his research, psychology and AI
especially in computer vision, has been were closely linked. This is a link that
solved by DL methods with a form of Nevertheless, DL methods can be we need to revisit.
artificial intuition. And this is not a sur- transparent in some terms: knowledge
prise given that important researchers about the used training set and an in- Link: [L1] https://kwz.me/hdj
such as Simon have accepted the equiva- depth analysis of the statistical out-
lence between intuition and recognition. comes can help in making them References:
trustable for a specific task, in a specific [1] H. A. Simon: “What is an ‘explana-
Even if many people feel a sense of context at a specific time. tion’ of behavior?” Psychological
magic talking about DL, the research science 3.3, 150-161, 1992.
conducted in recent years has proven As humans should not rely on intuition [2] D. Kahnemann: “Thinking, Fast
that there is no magic at all in intuition for all decisions, DL methods should be and Slow”, Farrar, Straus and
and the same holds for DL. used as part of more complex AI sys- Giroux, 2011.
tems that also involve non-intuitive
Within the discipline of psychology and processes. Kahneman has used the Please contact:
decision making, expert intuition has metaphor of two systems in his research Fabrizio Falchi, ISTI-CNR, Italy
been discussed a lot in recent years, about human thinking: +39 050 315 29 11
dividing researchers into believers and Fabrizio.falchi@cnr.it
Does the use of online platforms to share opinions contribute to the polarization of the public
debate? An answer from a modelling perspective.
The process of formation of opinions in ences from the design of the social answer some of these questions by
a society is complex and depends on a media platforms in use. These plat- building models of opinion dynamics
multitude of factors. Some relate to per- forms are built with a marketing target that mimic formation of opinions in
sonal preferences, culture or education. in mind: to maximise the number of society. Based on the evolution of opin-
Interaction with peers is also relevant: users and the time they spend on the ions, a group of people can reach con-
we discuss important issues with friends platform. To achieve this, the informa- sensus, i.e. agreement on a certain topic,
and change or reinforce our opinions tion that reaches the users is not ran- or fragmentation and polarisation of
daily. Another significant effect is that domly selected. A so called ‘algo- society can emerge. This process can be
from the media: external information rithmic bias’ exists: we see the news modelled by representing opinions with
reaches and influences us constantly. related to topics that we like and the continuous numbers, and simulating
The choice of the people we interact opinions of friends that are close to our interactions with specific rules to
with, and of the news we read, is thus opinions, so that we are driven to read change opinions at regular intervals.
crucial in the formation of our opinions. them and come back to the platform. The population typically starts from a
In the last decade, the patterns of inter- However, a question arises: does that random configuration and evolves in
action with peers, and of news con- interfere in any way with the formation time to either consensus or a frag-
sumption, have changed dramatically. of our opinions? Do we ever change mented state. One very popular model
While previously one would read the our minds any more, or we just keep for such simulations is the ‘bounded
local newspaper and discuss with close reinforcing our positions? Do we ever confidence’ model, where peers interact
friends and neighbours, nowadays see things in a different perspective, or only if their opinion is close enough. In
people can interact at large distances we are now closed in our little informa- this model, clusters of opinion appear if
and read news across the world through tion bubbles? the confidence is low, while for large
online media. Social media in particular confidence consensus emerges. This
is increasingly used to share opinions, A group of researchers from the model has been modified to include
but also, as the Reuters 2018 Digital Knowledge Discovery and Data Mining algorithmic bias. Instead of selecting
News Report shows, to read and share Laboratory in Pisa, and the Department peers to interact with in a random way,
news. This means that the peer and of Network and Data Science of the they are selected with a bias: a person is
external effects in the dynamics of opin- Central European University, funded by more likely to choose to interact with a
ions are becoming susceptible to influ- the European Commission, are trying to peer that has an opinion close to their
Figure 1: Simulation of opinion formation with and without the bias: The bias slows down the process and leads to the formation of two clusters
instead of one.
own, while it will have a low proba- time to reach it becomes very long. In Reference:
bility of interaction with opinions far practice, this means that it could take [1] Alina Sîrbu, et al.: “Algorithmic
from their own. years for people to agree on an issue, bias amplifies opinion polarization:
being in a highly fragmented state while A bounded confidence model”,
Simulations of the algorithmic bias this occurs. arXiv preprint arXiv:1803.02111,
model show several results that suggest 2018.
that online platforms can have impor- These results bring important evidence https://arxiv.org/abs/1803.02111
tant effect on opinion formation and that algorithmic bias may affect out-
consensus in society. First, the number comes of public debates and consensus Please contact:
of opinion clusters grows when algo- in society. Thus, we believe measures Alina Sîrbu,
rithmic bias grows (see illustration). are required to at least stop its effects, if University of Pisa, Italy
This means that online platforms can not reverse them. Researchers are alina.sirbu@unipi.it
favour fragmentation of opinions. investigating means of promoting con-
Second, this leads also to polarisation, sensus to counteract for the algorithmic
where the distance between the opin- bias effects. In the meantime, users
ions of the people is larger compared to could be informed of the way platforms
the situation without algorithmic bias. feed information and the fact that this
Third, the changes in opinion are much could affect their opinions, and maybe
slower when the bias is in operation. the mechanisms implemented by the
Even when consensus is obtained, the platforms could be slowly withdrawn.
The astonishing and cryptic effectiveness of Deep Neural Networks comes with the critical
vulnerability to adversarial inputs — samples maliciously crafted to confuse and hinder machine
learning models. Insights into the internal representations learned by deep models can help to
explain their decisions and estimate their confidence, which can enable us to trace, characterise,
and filter out adversarial attacks.
Machine learning and deep learning are thus the name) that look authentic to mechanisms for adversarial inputs and
pervading the application space in many human inspection, but cause the tar- machine learning security in general.
directions. The ability of Deep Neural geted model to misbehave (see Figure
Network (DNN) to learn an optimised 1). Although they resemble legitimate To this end, in a joint effort between the
hierarchy of representations of the input inputs, the high non-linearity of DNNs AIMIR Research Group of ISTI-CNR
has been proven in many sophisticated permits maliciously added perturba- and the CNIT Research Unit at MICC
tasks, such as computer vision, natural tions to steer at will the decisions the (University of Florence), we analysed
language processing and automatic model takes without being noticed. the internal representations learned by
speech recognition. As a consequence, Moreover, the generation of these mali- deep neural networks and their evolu-
deep learning methodologies are cious samples does not require a com- tion throughout the network when
increasingly tested in security- (e.g. plete knowledge of the attacked system adversarial attacks are performed.
malware detection, content moderation, and is often efficient. This exposes sys- Opening the “black box” permitted us
biometric access control) and safety- tems with machine learning technolo- to characterise the trace left in the acti-
aware (e.g. autonomous driving vehi- gies to potential security threats. vations throughout the layers of the net-
cles, medical diagnostics) applications work and discern adversarial inputs
in which their performance plays a crit- Many techniques for increasing the among authentic ones.
ical role. model’s robustness or removing the
adversarial perturbations have been We recently proposed solutions for the
However, one of the main roadblocks to developed, but unfortunately, only a detection of adversarial inputs in the
their adoption in these stringent contexts few provide effective countermeasures context of large-scale image recognition
is the diffuse difficulty to ground the for specific attacks, while no or mar- with deep neural networks. The
decision the model is taking. The phe- ginal mitigations exist for stronger rationale of our approaches is to attach
nomenon of adversarial inputs is a attack models. Improving the explain- to each prediction of the model an
striking example of this problem. ability of models and getting deeper authenticity score estimating how much
Adversarial inputs are carefully crafted insights into their internals are funda- the internal representations differ from
samples (generated by an adversary — mental steps toward effective defensive expected ones (represented by the
Figure 1:Example of a common adversarial attack on image classifiers. The adversarial Links:
perturbation added (magnified for visualization purposes) fools the network to predict a wrong [L1]
class with high confidence. http://deepfeatures.org/adversarials/
References:
model’s training set). In [1], such a input occupies in the feature spaces of [1] Carrara et al.: “Adversarial image
score is obtained by analysing the each layer with respect to most common detection in deep neural networks”,
neighbourhood of the input with a reference points (identified by looking Multimedia Tools and
nearest-neighbour search in the activa- to training set inputs). Experiments Applications, 1-21, 2018
tion space of a particular layer. Our showed that adversarial inputs usually [2] Carrara et al.: “Adversarial
experiments on adversarial detection tend to deviate from reference points examples detection in features
permitted us to identify the internal acti- leading to different activation traces in distance spaces”, ECCV 2018
vations which are influenced the most the network with respect to authentic Workshops, 2018.
by common adversarial attacks and to inputs (see Figure 2). Thus, condi-
filter out most of the spurious predic- tioning our detector on such informa- Please contact:
tions in the basic zero-knowledge attack tion permitted us to obtain remarkable Fabio Carrara, ISTI-CNR, Italy
model (see [L1]). detection performance under commonly fabio.carrara@isti.cnr.it
used attacks.
Building on this idea, in [2] we Roberto Caldelli, CNIT Research Unit
improved our detection scheme consid- We plan to extend our analysis in order at MICC ‒ University of Florence,
ering the entire evolution of activations to fully characterise the effect of adver- Italy
throughout the network. An evolution sarial attacks on internal activations roberto.caldelli@unifi.it
map is built by tracing the positions an even in stricter attack models, i.e. when
Figure 2: Conceptualisation of the evolution of features while traversing the network. Each plane represents a feature space defined by the
activations of a particular layer of the deep neural network. Circles on the features space represent clusters of features belonging to a specific
class. Blue trajectories represent authentic inputs belonging to three different classes, and the red trajectory represent an adversarial input. We
rely on the distances in the feature space (red dashed lines) between the input and some reference points representatives of the classes to encode
the evolution of the activations.
With the desire and need to be able to trust decision making systems, understanding the inner
workings of complex deep learning neural network architectures may soon replace qualitative or
quantitative performance as the primary focus of investigation and measure of success. We
report on a study investigating a complex deep learning neural network architecture aimed at
detecting causality relations between pairs of statements. It demonstrates the need to obtain a
better understanding of what actually constitutes sufficient and useful insights into the behaviour
of such architectures that go beyond mere transformation into rule-based representations.
Recently there has been increased pres- out to be useless as they do not provide the unintentional focus on logos
sure by legislators as well as ethics any rationale guiding the process of embedded in images as a clear separator
boards and the public at large to ensure how to repeat such optimisations for between positive and negative class
that algorithmic decision making sys- specific tasks in different settings. images due to the construction of the
tems also provide means for inspecting Inspection of deep learning (DL) net- training data base.
their behaviour and are able to explain works to understand what they are
how they arrive at any specific decision. doing, which parts of the input are most Image analysis settings offer themselves
This is basically well-aligned with influential in the final decision making, for inspection as the data processed can
researchers’ desire to understand the provide valuable insights both in under- conveniently be displayed in visual
workings of an algorithm as this usually standing a specific model as well as form, thus becoming early candidates
constitutes the most structured and only guiding the targeted design of improved for identifying areas of attention [1].
viable approach to improving an algo- architectures, data representations and Discovering the structure learned by
rithm’s overall performance. learning routines. Examples such as various network layers, such as the
attention highlighting in image pro- focus on edges and orientation of these
Naïve approaches such as untargeted cessing revealed both insights into the in subsequent layers of a convolutional
parameter sweeps and architecture vari- characteristics learned by a network, as neural network (NN) architecture [2],
ations leading to higher performance in well as uncovered errors and bias in the provide intuitive insights into their
some specific benchmark settings turn resulting systems, such as, for example, behaviour. Settings that do not offer
themselves for direct visual inspection
provide way harder challenges. Even
more so, tasks that go beyond mere clas-
sification of data based on individual,
independent attributes increase the chal-
lenge in devising interpretable represen-
tations of the inner workings of such
complex DL architectures. Here we
review the challenges of trying to devise
such an inspection in a setting of a
neural language inference model, where
the goal is to correctly classify the log-
ical relationship between a pair of sen-
tences into one of three categories:
entailment, neutral or contradiction.
Figure 1: Activation differences for the first four layers of a neural natural language inference Our approach to understanding com-
model trained on the SNLI corpus [3]. Layer 0 is an embedding lookup, layer 1 and 2 are plex neural network architectures is
bidirectional RNNs, layer 3 is a combination of an attention and a feed forward layer, layer 4 is based on the analysis of the interaction
a bidirectional RNN layer processing the two prior layers. patterns of a single input word with all
Neurodegenerative diseases such as Our algorithm performs by extracting a comparable to the original set, but
Alzheimer’s and Parkinson’s impact rule set from a random forest, which is much easier to comprehend. A clear
millions of people worldwide. Early then minimised within several steps. understanding of the process behind a
diagnosis has proven to greatly increase The algorithm can be divided into three diagnosis is crucial for both doctor and
the chances of slowing the diseases’ major steps. (1) Firstly, a random patient, and it is hoped that systems like
progression [2]. Correct diagnosis often forest, an ensemble of decision trees, is this one will become increasingly
relies on the analysis of large amounts created as the foundation of the algo- prevalent as we continue to improve the
of patient data, and thus lends itself well rithm. (2) Secondly, a set of rules is state-of-the art in predictive medicine.
to support from machine learning algo- extracted from the random forest. (3)
rithms, which are able to learn from past Thirdly, this rule set is reduced signifi- Mashayekhi and Gras [1] introduced
diagnosis and see clearly through the cantly. The user can influence step (3) two methods called RF+HC and
complex interactions of a patient’s by preferring as important considered RF+HC_CMPR which allow to extract
symptoms. Unfortunately, many con- features. The algorithm is implemented a rule set from random forests. The
temporary machine learning techniques in Python and we trained it to predict main idea of their work is to reduce the
fail to reveal details about how they whether patients are likely to suffer number of rules dramatically and there-
reach their conclusions, a property con- from Alzheimer’s or Parkinson’s dis- fore increase the comprehensibility of
sidered fundamental when providing a ease, but it is a generic algorithm that the underlying model. We expanded
diagnosis. This is one reason why we can be applied to any kind of disease. this idea and added another personalis-
introduce our personalisable CDSS that Three main factors are taken into able layer to it to allow the user to add
provides a clear insight into the decision account during the reduction process: data driven and personal evidence to
making process on top of the diagnosis. (a) the performance of individual rules, the algorithm.
Our algorithm enriches the fundamental (b) the rules’ transparency, and (c) the
work of Masheyekhi and Gras [1] in personal preferences of the user. The We applied our algorithm to different
data integration, personal medicine, reduction leads to a new, smaller set of data sets with a variety of data types
usability, visualisation and interactivity. rules, with predictive performance like magnetic resonance images, bio-
In healthcare, trust is the basis of the adversely affect the fabric of trust in actual reasoning patterns remain opaque
doctor-patient relationship. A patient these systems and consequently impact and leave a lot of room for guesswork.
expects the doctor to act reliably and the doctor-patient relationship. Here, deep generative networks show
with precision and to explain options and promise by generating visual clues as to
decisions. The same accuracy and trans- Current solutions for transparency in why a decision was made [1].
parency should be expected of computa- deep learning (used synonymously with
tional systems redefining the workflow AI) centre around the generation of heat- We believe transparency in image based
in healthcare. Since such systems have maps. These highlight high-impact algorithmic decision making can only
inherent uncertainties, it is imperative to image regions on deep learning deci- be achieved if expert computer scien-
understand a) the reasoning behind such sions. While informative in nature, tists and healthcare professionals (radi-
decisions and b) why mistakes occur. direct adaptation of such methods into ologists, pathologists etc.) closely col-
Anything short of this transparency will healthcare is insufficient, because the laborate in an equal-share environment.
Figure 2: With perturbations beyond human perception, deep learning often leads to both wrong decisions and wrong corresponding
interpretation.
and incidents, this black-box behaviour Transparency in decision making has [3] D. Kügler, et al.: “Exploring
of obscuring the reasoning is a signifi- developed to be a key component we Adversarial Examples”,
cant obstacle for the clinical implemen- expect from our doctors. As deep Understanding and Interpreting
tation of AI in healthcare. As (closely) learning is introduced into healthcare, we Machine Learning in Medical
collaborating experts, we are developing need standardised methods to guarantee Image Computing Applications,
courses and guidelines to make doctors results and transparency to doctors. Springer, Cham, 2018.
AI-ready.
References: Please contact:
In retrospective in-silico studies, deep [1] J. C. Y. Seah, et al.: “Chest Anirban Mukhopadhyay
learning-based algorithms are already radiographs in congestive heart Informatik, Technische Universität
showing promise to be the single most failure: visualizing neural network Darmstadt, Germany
successful concept in medical image learning”, Radiology, 2018. anirban.mukhopadhyay@gris.tu-
analysis. In order to realise the potential [2] M. Sahu, et al.: “Addressing multi- darmstadt.de
impact these algorithms can have on our label imbalance problem of
healthcare, we need to introduce trust- surgical tool detection using CNN”,
focused prospective clinical studies. IJCARS, 2017.
Artificial Intelligence (AI) applications may have different ethical and legal implications depending on the
domain. One application of AI is analysis of video-interviews during the recruitment process. There are
pros and cons to using AI in this context, and potential ethical and legal consequences for candidates,
companies and states. There is a deficit of regulation of these systems, and a need for external and
neutral auditing of the types of analysis made in interviews. We propose a multi-agent system
architecture for further control and neutral auditing to guarantee a fair, inclusive and accurate AI and to
reduce the potential for discrimination, for example on the basis of race or gender, in the job market.
Image analysis in human resources: • Attention to detail (eye contact time, according to Siyao et all [1]. MIT
pros and cons emotions-intonation and body lan- researchers acknowledged race-bias in
There has been a recent trend towards guage) and lack of interviewer bias learning algorithms mainly trained with
video-interview analysis in HR depart- (physical appearance, tattoos, etc.). data from white people.
ments. Traditionally, AI played no more
than an assistant role in HR, e.g. But there are several problems that As an illustration of the advances in
resume and CV scanning. But lately, accompany the use of these technologies: sexual orientation recognition both in
apps and systems like HireVue [L1], • Candidates are unfamiliar with video images and sound, one study [2] needed
Montage [L2], SparkHire [L3] and interview analysis (for example, ethical supervision due to the opaque
WePow [L4] have been changing how lighting, settings), which could affect invasive nature of the research and the
recruitment is carried out. An AI-based global performance. use of real user data from dating appli-
video interview system could be pro- • Gender and racial bias: traditionally, cations. The study argues that there is a
grammed to check, during an interview, machine learning algorithms were relationship between homosexuality
features such as age, lighting, tone of trained with data from white people. and exposure to particular concentra-
voice, cadence, keywords used (sub- • Imprecisions of technology. Training tions of hormones in the womb, and that
stantial conversation), mood, behaviour classifiers with biased datasets. For sexual orientation can be determined by
(eccentric, movement or quite calm and instance, Affectiva [L5] dataset of morphological features (for example,
not talkative), eye contact and, above human emotions was fed with data the jawline and forehead).
all, emotions. AI targets the specific from Superbowl viewers, and could
traits of a customer-oriented role that presumably have culture-bias. Ethical and legal aspects of AI
employers want in their teams. Whilst the use of AI in this context
Controversial characteristics may have its benefits, it also strips
AI has produced benefits for HR so far, We studied several potential controver- away aspects of humanity, reducing a
including: sial characteristics, among them, facial human recruit to a set of descriptors.
• Reduction of interview time per can- symmetry, race, gender, sexual orienta- The automation of HR processes could
didate, thus recruiting time. tions in voice and image recordings. lead to potential ethical and legal
• Customised candidate experience The problem of racial-bias in AI is not implications that cannot be ignored. In
and customised questions and new, just like the detection of mixed some countries, companies are not
answers. race in bad lighting conditions allowed to ask a candidate’s age
Automated decision-making has the potential to increase both productivity and competitiveness as
well as compensate for well-known human biases and cognitive flaws [1]. But today’s powerful
machine-learning based technical solutions also bring about problems of their own – not least in
terms of being uncomfortably black-box like. A new research project at RISE Research Institutes of
Sweden, in collaboration with KTH Royal Institute of Technology, has recently been set up to study
transparency in the insurance industry, a sector that is poised to undergo technological disruption.
At the Danish Insurance 2018 confer- keted Watson with the tagline ance claims 25% faster”. The fact that
ence in Copenhagen in February, an “Insurance company employees are such efforts are underway is no sur-
IBM partner in the exhibition area mar- working with Watson to assess insur- prise to the informatics or mathematics
communities. Yet, it serves as an excel- holders. In other words, meaningful self-defeating to be transparent about
lent illustration of both the promise and algorithmic transparency probably algorithms designed to detect insurance
the perils that the insurance industry looks different to a CEO, to an actuary, fraud. Thus, the project will investigate
faces in the digital world. to a software engineer, and to a con- needs and techniques for selective
sumer looking for a car insurance. transparency.
Start with the promise. Insurance still
largely relies on manual labor Another important strand is ethics. The project is still in its infancy, but in
throughout its value-chain. Though dig- Ethicists have argued that the imple- two years the research team will have
itally supported, assessing risk, setting mentation of algorithms entails ethical taken first steps in designing the
premiums, and adjusting claims most judgements and trade-offs, though not modern software quality assurance
often involves human decision-making. always explicit. The project will engage processes that will enable
With smarter tools, whether based on with ongoing software projects with the Länsförsäkringar and the rest of the
rules or machine-learning, more work funder and work with development insurance industry to make more
could certainly be automated, cutting teams to make explicit the ethically rel- informed decisions about how to best
costs for companies and premiums for evant choices made. A promising tech- make use of the advances in machine-
customers. nique is the paradigm of value-sensitive learning. As “insurers face unprece-
design [2], which has been successfully dented competitive pressure owing to
However, the road to digital insurance employed in other areas. technological change” [3], it will cer-
can be rocky. In the past few years, tainly be an interesting road ahead.
there have been abundant reports of A third project cornerstone will be to
flawed machine-learning applications, explore the consequences of increased Link:
e.g., the infamous misclassification of transparency in insurance. Some are [L1] https://www.ri.se/en/what-we-
people as monkeys in image-recogni- expected to be decidedly positive, do/projects/transparent-algorithms-
tion systems, signs of racial bias in sys- ranging from better requirements engi- insurance
tems predicting the probability of crim- neering in software projects to better
inal recidivism, and sexism in natural customer value and even growing References:
language processing. How can an insur- market shares. Indeed, insurance [1] A. Tversky, D. Kahneman:
ance company that starts employing startups and would-be disruptors such “Judgment under uncertainty:
machine-learning at larger scale be sure as Lemonade in the US and Hedvig in Heuristics and biases”, Science
that risk is properly assessed, that pre- Sweden use transparency, in various 185.4157, 1974, 1124–1131. DOI:
miums levels are sustainable, and that forms, as selling points. 10.1126/science.185.4157.1124
customers are fairly treated when [2] B. Friedman, et al.: “Value
claims are adjusted? However, the insurance industry is par- sensitive design and information
ticularly interesting because there is systems”, Early engagement and
Such questions prompted Läns- also a significant potential downside to new technologies: Opening up the
försäkringars forskningsfond, the transparency. Insurance companies laboratory. Springer, Dordrecht,
research arm of a major Swedish need to design their products so that 2013. 55–95. DOI: 10.1007/978-
insurer, to issue a call for research pro- customers’ incentives don’t change for 94-007-7844-3_4
posals related to digitalization and the worse, e.g., a car owner driving [3] “The future of insurance: Counsel
insurance. The resulting project, recklessly because the insurance cov- of protection”, The Economist,
Transparent algorithms in insurance erage will take care of the conse- March 11, 2017, 67–68.
[L1], started in October 2018 and aims quences. Mechanisms such as
to explore issues of algorithmic trans- deductibles are designed precisely to Please contact:
parency, broadly construed, in the insur- prevent this moral hazard. Similarly, the Ulrik Franke
ance industry. During the project’s two- fact that the insured typically knows RISE ICT/SICS, Sweden
year stint, several research questions more than the insurer, e.g., experiencing +46 72 549 92 64
will be studied, some technical, some an ache that a physician cannot detect, ulrik.franke@ri.se
more of a social science nature. might lead to more risk among insureds
than expected. Again, mechanisms such
One important part of the project will be as indemnity limits are designed to pre-
to study conditions for algorithmic vent such adverse selection. While these
transparency in insurance. Many tech- conceptual problems inherent to insur-
niques for transparency in machine- ance have been known for a long time,
learning systems exist, but they are an increased dependence on algorithms
mostly designed to fit technologies such paired with greater transparency might
as deep neural networks or support invite insureds to novel forms of unde-
vector machines. However, meaningful sirable strategic behavior, which ulti-
transparency also needs to account for mately leads to higher premiums for
industry, e.g., insurance, and stake- customers. For example, it is probably
“Cyber threat intelligence” is security-relevant information, often directly derived from cyber
incidents that enables comprehensive protection against upcoming cyber-attacks. However,
collecting and transforming the available low-level data into high-level threat intelligence is usually
time-consuming and requires extensive manual work as well as in-depth domain knowledge.
INDICÆTING supports this procedure by developing and applying machine learning algorithms that
automatically detect anomalies in the monitored system behaviour, correlate affected events to
generate multi-step attack models and aggregate them to generate usable threat intelligence.
Today’s omnipresence of computer sys- compare the observed events with pre- time passes until blacklists are updated
tems and networks provides great ben- defined signatures specified in the with information on immanent attacks.
efit to the economy and society, but also blacklists to raise alerts if certain
opens the door to digital threats, such as thresholds are exceeded. INDICÆTING aims to solve this
data theft. Hackers and cyber-criminal problem by automatically generating
groups carry out attacks by exploiting However, this basic approach suffers complex “threat intelligence”, i.e., more
vulnerabilities of the deployed systems, from a serious shortcoming: simple expressive than simple IoCs, but rather
with little or no time for the victims to indicators of compromise (IoC) such as complex TTPs. INDICÆTING thereby
react. malicious IP addresses are highly pursues anomaly detection rather than
volatile and only valid for a short period blacklisting, i.e., instead of relying on
The growing interconnectedness of dig- of time, since it is easy for attackers to an existing knowledge base,
ital services helps to enable cyber- circumvent their detection. Tactics, INDICÆTING makes use of self-
attacks. The Internet of Things and techniques and procedures (TTP) on the learning algorithms that capture the
Industry 4.0 entail the emergence of other hand are valid for a longer time, normal system behaviour over time and
highly complex system landscapes that because it is difficult to change the detect deviations from the expected pat-
offer numerous entry points for infiltra- modus operandi of attacks [1]. terns [2]. This way, continuously gener-
tion. As a simple measure of protection, However, compiling threat intelligence ated low-level log data that documents
most modern systems are equipped with on TTPs is difficult and requires manu- almost all events occurring in the
blacklists containing indicators of com- ally analysis of complex attack patterns. observed system is continuously moni-
promise, such as IP addresses, known to All manual analyses are tedious and tored as soon as it is generated.
correspond to adversarial entities. time-consuming, and since attacks are
Detection systems monitor infrastruc- often carried out on a large-scale and Log data contains semantically expres-
ture states and outbound connections, affect multiple organisations, too much sive parameters describing the current
Log Parser
Log Parser
data Generator
Anomaly Detection
System Behavior Analysis
Event Classification Log
data
Alert Alert
Static Dynamic
Aggregation Correlation
Anomaly Anomaly
Rule Check
Detection Detection
(e.g.,
SNORT)
IoC Extraction
Parameter/
TTP Rules
Signature
Extraction (e.g., STIX,
Extraction
YARA)
Detected Attacks
system state and is thus suitable for with a corresponding model that was INDICÆTING is financially supported
analysing the roots of system failures in trained over a long time, i.e., a baseline by the Austrian Research Promotion
hindsight after incidents occurred – a model. Thereby, efficiency of the pro- Agency (FFG) under grant number
task that has been carried out by soft- posed algorithms is a key feature, 868306. The project is carried out in
ware engineers for decades. Only because log data is typically produced in course of an industry-related PhD thesis
recently, have system logs been enormous amounts and fast rates. at the Austrian Institute of Technology
analysed in real-time in order to indicate (AIT) in cooperation with the Vienna
system problems almost as they occur. Since anomalies reported on individual University of Technology (TU WIEN).
However, this is a highly non-trivial parameters or events are not sufficient During the runtime of the project, AIT’s
activity: the main issue with processing for describing complex attack patterns, Automatic Event Correlation for
system logs is that they are unstructured a subsequent step is required that Incident Detection (AECID) [L2] tool
and different on every system, and it is analyses the overall system behaviour. will be further developed and used for
challenging to automatically extract For this purpose, the identified anom- evaluating the proposed concepts.
parameters and map individual log lines alies are correlated over multiple data
to more abstract event classes without channels and architectural layers in References:
human intervention. INDICÆTING order to model multi-step attacks and [1] D. Chismon, M. Ruks: “Threat
achieves this by parsing the data, i.e., derive abstract TTPs that affect several Intelligence: Collecting, Analysing,
determining which parts of the log lines components in a narrow time window. Evaluating”, MWR InfoSecurity, 2015.
correspond to constant (textual) parts, For example, consider an employee [2] V. Chandola, et al., “Anomaly
and which correspond to parameters entering login credentials after Detection: A Survey”, in ACM
such as usernames, IDs, IP addresses, receiving a phishing email that contains Comput. Surv., 2009.
etc. As shown in Figure 1, a parser gen- a URL to a malicious website. Using the
erator learns the structure of log data credentials, the attacker then infiltrates Links:
collected from a honeynet, i.e., a system the network from a remote connection. [L1] http://misp-project.org/
specifically set up to attract attackers. This is a multi-step attack that can be [L2] https://aecid.ait.ac.at/
Once the parsing model is established, detected by correlating URLs in mails,
INDICÆTING is able to retrieve param- DNS, and web proxy logs. After appro- Please contact:
eter values and reason on their static dis- priately modelling all involved steps Max Landauer
tributions, discover dynamic dependen- and additional information on parame- AIT Austrian Institute of Technology,
cies between events and construct ters, timing and context, the resulting Austria
process models. Based on these values, high-level threat intelligence is suitable +43 664 88256012
anomalies are then detected by com- to be shared with other parties in cyber- max.landauer@ait.ac.at
paring each newly incoming log line security communities [L1].
In complex production environments, understanding the results of a scheduling algorithm is a challenging task. To
avoid tardy work orders, proALPHA and Fraunhofer ITWM developed a component for identifying practical tardiness
reasons and appropriate countermeasures.
The proALPHA group is the third tion of the whole production process work orders with respect to current
largest ERP provider for small and mid- (from prefabrication to final products) of resource and material availabilities.
sized manufacturing and trading compa- a company by providing decision sup- The considered problem is a (multi-)
nies in Germany, Austria, and port for production planners. To this end, resource constrained multi-project
Switzerland. For more than 25 years, work orders (consisting of structured scheduling problem (see [1]) with sev-
proALPHA has offered a powerful ERP process steps), calendars for production eral additional features: resource and
solution as well as consulting, support, resources (workforces, machines, tools), overload decisions, producer/consumer
training, and maintenance services from material levels (with reorder times and assignments and time bounds among
one source. The flexible and scalable known in- and outflows) and resource as others. The objectives are the mini-
ERP solution features a wide range of well as material requirements can be mization of tardiness, earliness and
functions that allow all processes along managed. In addition, the module con- throughput time of the work orders. To
the value-added chain to be controlled. tains various production planning and model and solve the problem a
detailed scheduling algorithms. Constraint Programming approach (see
One of the modules available for the [2]) is used. For the implementation,
proALPHA ERP solution is Production. These algorithms can be used to exe- the CP-library IBM ILOG CP
The goal of this module is the organiza- cute a complete optimization of all Optimizer (see [3]) is employed.
Since the complete optimization takes simple and can be divided in different Tests with practical data show that far
limited resource and material capacities reason-types: more than 90% of the delays can be
into account, tardy processes can typi- • Structure: Processing times together explained, even if not all possible rea-
cally not be avoided. In many cases, this with precedence constraints lead to a sons are covered. The component helps
is not acceptable, as tardy processes delay. production planners to identify tardi-
violate contracts or delivery promises • Time Bound: A time bound (release ness reasons as well as bottlenecks
made to important customers. dates) leads to a delay. (materials, resources). As a result, it
Therefore, one of the main tasks of a • Single Material: A missing material leads to a higher acceptance of the
production planner is to identify why a leads to a delay. schedules provided by the complete
particular process is scheduled too late • Single Resource: A resource with optimization.
and what he or she can do to make a missing capacity leads to a delay.
production in time possible. This task, • Combined: A missing material References:
however, is not easy as, for example, together with a resource or two [1] S. Hartmann, D. Briskorn: “A
customer orders often require many resources together lead to a delay. survey of variants and extensions
production resources, raw materials and of the resource constrained project
prefabrication steps. Based on these reason-types, the tardi- scheduling problem”, European
ness-reasoning component works as Journal of Operational Research,
To support the production planner in follows. First, all potential reasons for a 207(1), 1-14, 2010.
this critical task, the complete optimiza- tardy process are identified (for [2] P. Baptiste, C. Le Pape, W. Nuijten:
tion has been extended by a tardiness- example, each consumed material is a “Constraint-Based Scheduling –
reasoning component, which provides potential material reason). Then, for Applying Constraint Programming
tardiness reasons with respect to the each potential reason it is tested, if it is to Scheduling Problems”, Kluwer,
optimized schedule. From a mathemat- indeed responsible for a delay. For each 2001.
ical point of view, a tardiness reason of reason-type, special constraint based [3] P. Laborie, J. Rogerie, P. Shaw, P.
a process is a minimal set of constraints, models are used for this test. Finally, if a Vilím: “IBM ILOG CP optimizer
which, together with the constraint that reason is verified, further information is for scheduling, 20+ years of
the process ends in time, is infeasible. calculated that help the production scheduling with constraints at
However, the determination of such planner to evaluate the importance of IBM/ILOG”, Constraints, 23, 210 –
minimal infeasible sets is, in general, a the reason and give hints on how the 250, 2018.
hard problem and minimal sets alone problem can be fixed. Examples are the
are of limited use. delay caused by a reason, the amount of Please contact:
a material and the latest point in time Sebastian Velten
From a practical point of view, tardi- when it is needed as well as the amount Fraunhofer ITWM, Germany
ness reasons are nearly always rather of free time that is missing on a resource +49 (0)631 31600 4260
(see Figure 1). sebastian.velten@itwm.fhg.de
Figure 1: A full non-modular ANN with a random 10% selection of inter-modular synapses, i.e., the synapses of the modular TS and CR
processing network (light grey lines) are augmented with those between the TS and CR processing subnetworks (black lines). The ANN
simulations and the diagram was made with the Simbrain 3.0. Neural Networks framework.
data and how they may be motivated to share them, and on the
Science2Society Project other hand, what needs to be done to enable industry to easily
use these data. The overarching objective is then to identify
unveils the Effective use what type of sustainable business models can best support
both these issues: (i) the sharing of data and (ii) how to enable
of big Research Data industry to take advantage of this data.
As final recommendations for the management of open big Deep neural networks have pushed the boundaries of
research data, researchers should develop a research data artificial intelligence but their training requires vast
management plan prior to the start of the project. The decision amounts of data and high performance hardware. While
to open the data should be based on a careful assessment of truly digitised companies easily cope with these
the underlying opportunities, barriers and alternative business prerequisites, traditional industries still often lack the
models. We provide a generic process model for performing kind of data or infrastructures the current generation of
that assessment. Based on the GCAT case, lessons learnt and end-to-end machine learning depends on. The Fraunhofer
best practices for how to operate a big research database are Center for Machine Learning therefore develops novel
provided. The key for a sustainable business model is to focus solutions which are informed by expert knowledge.
on developing a valuable proposition for data owners and These typically require less training data and are more
users that benefits both parties. transparent in their decision-making processes.
For more information please refer to the Science2Society Big data based machine learning (ML) plays a key role in the
Knowledge Database, which contains case studies, methods recent success of artificial intelligence (AI). In particular,
and tools related to this pilot and the other pilots from the deep neural networks trained with vast amounts of data and
project. Additionally, the document “D3.2 Report on the high performance hardware can now solve demanding cogni-
implementation and evaluation of the UIS interface scheme tive tasks in computer vision, speech recognition, text under-
pilots”, will be available soon in the downloads section of the standing, or planning and decision-making. Nevertheless,
project website. practitioners in industries other than IT are increasingly
skeptical of deep learning, mainly because of:
The knowledge and studies on this collaborative schema and • Lack of training data. Vapnik-Chervonenkis (VC) theory
others will be maintained after the project ends with the set- establishes that supervised training of complex ML sys-
ting of the Learning and Implementation Alliance. tems requires substantial amounts of representative data in
order to learn reliably. Since details depend on a system’s
Links: VC dimension, which is usually hard to come by,
[L1] http://www.science2society.eu/ Widrow’s rule of thumb is to train with at least ten times
[L2] http://www.science2society.eu/kd-front more data than there are system parameters. Modern deep
[L3] http://www.gcatbiobank.org/en_index/ networks with their millions of adjustable parameters thus
need many more millions of examples in order to learn
Please contact: well. However, large amounts of annotated data are rarely
Ricard Munné Caldés, Atos, Spain available in industries that are not yet fully digitised. Even
+34 935485741, ricard.munne@atos.net in contexts such as the internet of things or industry 4.0
where data accumulate quickly, we still often face thin
data scenarios where labeled data for end-to-end machine
learning are inaccessible.
• Lack of traceability. Trained connectionist architectures
are black boxes whose inner computations are abstracted
away from conceptual information processing. As their
decision-making processes may thus be unaccountable,
data scientists in industries where regulatory guidelines
demand automated decision making to be comprehensible
are wary of deep learning. Indeed, recent research shows
that silly mistakes made by deep networks might be avoid-
able if they had “common sense”. Even more alarmingly,
recent research also shows that silly mistakes can be pro-
voked using adversarial inputs.
Link:
[L1] https://www.cit.fraunhofer.de
References:
[1] C. Bauckhage, et al.: “Informed Machine Learning
through Functional Composition”, Proc. KDML, 2018.
[2] N. Aspiron and M. Bortz: “Process Modeling, Simula- About ERCIM
tion and Optimization: From Single Solutions to a Multi- ERCIM – the European Research Consortium for
tude of Solutions to Support Decision Making”, Chemie Informatics and Mathematics – aims to foster collabora-
Ingenieur Technik, 90(11), 2018. tive work within the European research community and to
increase cooperation with European industry. Founded in
Please contact: 1989, ERCIM currently includes 16 leading research
Daniel Schulz establishments. ERCIM is able to undertake consultancy,
Fraunhofer IAIS, Germany development and educational projects on any subject
+49 2241 142401 related to its field of activity.
daniel.schulz@iais.fraunhofer.de
ERCIM members are centres of excellence across
Europe. ERCIM is internationally recognized as a major
representative organization in its field. ERCIM provides
access to all major Information Communication
Technology research groups in Europe and has estab-
lished an extensive program in the fields of science,
strategy, human capital and outreach. ERCIM publishes
ERCIM News, a quarterly high quality magazine and
delivers annually the Cor Baayen Award to outstanding
young researchers in computer science or applied mathe-
matics. ERCIM also hosts the European branch of the
World Wide Web Consortium (W3C).
contact@ercim.eu
Monalisa MANDAL
Former ERCIM Fellow
The fellowship scheme also helps young scientists to Maximiliano Bueno López
Former ERCIM Fellow
improve their knowledge of European research structures
and networks and to gain more insight into the working con-
ditions of leading European research institutions.
The Innovative Training Network POEMA is hiring 15 Doctoral Students starting from September 2019. The proposed proj-
ects will be investigating the development of new algebraic and geometric methods combined with computer algebra tech-
niques for global non-linear optimization problems. Applications will focus on smarter cities challenges, urban traffic man-
agement, water network management, energy flow control, or environmental monitoring.
Available positions
1. Algebraic tools for exact SDP and its variants 9. Structure of moment problems and applications
Advisor: M. Safey el Din (Sorbonne University, Paris, to polynomial optimisation
France) Advisor: B. Mourrain (INRIA, Sophia Antipolis,
France)
2. Exact algorithms for structured polynomial
optimisation 10. Alternative polynomial bases for global
Advisor: M. Safey el Din (Sorbonne University, Paris, optimization
France) Advisor: E. Hubert (INRIA, Sophia Antipolis, France)
3. Polynomial optimization problems with symmetry 11. Numerical cubature with symmetry and applications
Advisor: C. Scheiderer (Univ. of Konstanz, Germany) to polynomial optimisation
Advisor: C. Riener (Univ. of Tromsoe, Norway)
4. Hyperbolic polynomials and the Generalized Lax
Conjecture 12. Algorithms and software for nonlinear convex conic
Advisor: M. Schweighofer (Univ. of Konstanz, optimisation
Germany) Advisor: M. Stingl (Friedrich Alexander Univ.
Erlangen, Germany)
5. Tensor Decomposition by Vector Bundles tools
Advisor: G. Ottaviani (Firenze, Italy) 13. Algorithms and software for structured SDP
Advisor: M. Kocvara (Univ. of Birmingham, UK)
6. Approximation hierarchies for (non-)commutative
polynomial optimisation 14. Polynomial Optimization: Some challenges from
Advisor: M. Laurent (CWI, Amsterdam, the applications
Nerthelands) Advisor: M. Korda (CNRS, LAAS, Toulouse, France)
7. Approximation hierarchies for graph parameters 15. Polynomial Optimization Techniques for Energy
Advisor: M. Laurent (CWI, Amsterdam, the Network Operation and Design
Nerthelands) Advisors: J.-H. Hours, M. Gabay (ARTELYS, Paris,
France)
8. Polynomial Optimization Problems in Operations
Research and Finance
Advisor: E. de Klerk (Univ. of Tilburg, the
Netherlands)
The positions have usually a duration of three years. Applicants must have a Master’s degree in Computer Science,
Mathematics or Engineering (or any equivalent diploma) at the date of recruitment.
POEMA is a Marie Skłodowska-Curie Innovative Training Network funded by the European Union. Its goal is to train sci-
entists at the interplay of algebra, geometry and computer science for polynomial optimization problems and to foster scien-
tific and technological advances, stimulating interdisciplinary and intersectoriality knowledge exchange between alge-
braists, geometers, computer scientists and industrial actors facing real-life optimization problems.
New Book
Editorial Information
Cinderella’s Stick ‒ A fairy
ERCIM News is the magazine of ERCIM. Published quarterly, it
reports on joint actions of the ERCIM partners, and aims to Tale for Digital Preservation
reflect the contribution made by ERCIM to the European
Community in Information Technology and Applied Mathematics. Authors: Yannis Tzitzikas and Yannis, Marketakis
Through short articles and news items, it provides a forum for the (ICS-FORTH)
exchange of information between the institutes and also with the
wider scientific community. This issue has a circulation of about This book explains the main problems related to dig-
6,000 printed copies and is also available online. ital preservation using examples based on a modern
version of the well-known Cinderella fairy tale.
ERCIM News is published by ERCIM EEIG Digital preservation is the endeavor to protect digital
BP 93, F-06902 Sophia Antipolis Cedex, France material against loss, corruption, hardware/software
Tel: +33 4 9238 5010, E-mail: contact@ercim.eu technology changes, and changes in the knowledge of the community.
Director: Philipp Hoschka, ISSN 0926-4981
Τhe structure of the book is modular, with each chapter consisting of two parts: the
Contributions episode and the technical background. The episodes narrate the story in chronolog-
Contributions should be submitted to the local editor of your ical order, exactly as in a fairy tale. In addition to the story itself, each episode is
country related to one or more digital preservation problems, which are discussed in the tech-
nical background section of the chapter. To reveal a more general and abstract formu-
Copyright notice lation of these problems, the notion of pattern is used. Each pattern has a name, a
All authors, as identified in each article, retain copyright of their summary of the problem, a narrative describing an attempt to solve the problem, an
work. ERCIM News is licensed under a Creative Commons explanation of what could have been done to avoid or alleviate this problem, some
Attribution 4.0 International License (CC-BY). lessons learned, and lastly, links to related patterns discussed in other chapters.
Advertising The book is intended for anyone wanting to understand the problems related to digital
For current advertising rates and conditions, see preservation, even if they lack the technical background. It explains the technical
http://ercim-news.ercim.eu/ or contact peter.kunz@ercim.eu details at an introductory level, provides references to the main approaches (or solu-
tions) currently available for tackling related problems, and is rounded out by ques-
ERCIM News online edition: ercim-news.ercim.eu/ tions and exercises appropriate for computer engineers and scientists. In addition, the
book’s website, maintained by the authors, presents the contents of Cinderella’s “real
Next issue: April 2019, Special theme: 5G USB stick,” and includes links to various tools and updates.
Editorial Board:
Central editor: hoRIZoN 2020 Project Management
Peter Kunz, ERCIM office (peter.kunz@ercim.eu)
A European project can be a richly rewarding tool for pushing your research or
Local Editors: innovation activities to the state-of-the-art and beyond. Through ERCIM, our
Austria: Erwin Schoitsch (erwin.schoitsch@ait.ac.at) member institutes have participated in more than 90 projects funded by the
Cyprus: Georgia Kapitsaki (gkapi@cs.ucy.ac.cy European Commission in the ICT domain, by carrying out joint research activities
France: Christine Azevedo Coste (christine.azevedo@inria.fr) while the ERCIM Office successfully manages the complexity of the project
Germany: Alexander Nouak (alexander.nouak@iuk.fraunhofer.de) administration, finances and outreach. The ERCIM Office has recognized
Greece: Lida Harami (lida@ics.forth.gr), expertise in a full range of services, including identification of funding opportuni-
Athanasios Kalogeras (kalogeras@isi.gr) ties, recruitment of project partners, proposal writing and project negotiation, con-
Hungary: Andras Benczur (benczur@info.ilab.sztaki.hu) tractual and consortium management, communications and systems support,
Italy: Maurice ter Beek (maurice.terbeek@isti.cnr.it) organization of events, from team meetings to large-scale workshops and confer-
Luxembourg: Thomas Tamisier (thomas.tamisier@list.lu) ences, support for the dissemination of results.
Norway: Monica Divitini, (divitini@ntnu.no)
Poland: Hung Son Nguyen (son@mimuw.edu.pl) How does it work in practice?
Portugal: José Borbinha (jlb@ist.utl.pt) Contact the ERCIM Office to present your project idea and a panel of experts will
Sweden: Maria Rudenschöld (maria.rudenschold@ri.se) review your idea and provide recommendations. If the ERCIM Office expresses
Switzerland: Harry Rudin (hrudin@smile.ch) its interest to participate, it will assist the project consortium either as project
The Netherlands: Annette Kik (Annette.Kik@cwi.nl) coordinator or project partner.
W3C: Marie-Claire Forgue (mcf@w3.org)
Please contact:
Cover illustrtation: source: Shutterstock. Peter Kunz, ERCIM Office, peter.kunz@ercim.eu
New Eu Project
Celebrate
Data Market Services
the Web@30
ERCIM participates in DataMarketServices (DMS), a new
H2020 project whose objective is to overcome the barriers In 1989, CERN was a
of data-centric European SMEs and start-ups by providing hive of ideas and infor-
free support services around data skills, entrepreneurial mation stored on mul-
opportunities, legal issues and standardization. The expected tiple incompatible computers. Tim Berners-Lee envisioned a
project deliverables are a 100-data-based companies port- unifying structure for linking information across different
folio, twelve free support services, webinars and training in computers, and wrote a proposal in March 1989 called
topics such as GDPR, IPR, etc. “Information Management: A Proposal”. By 1991, this
vision of universal connectivity had become the World Wide
The DMS consortium is composed of Zabala (ES); the Web! To celebrate 30 years since Tim Berners-Lee’s pro-
University of Southampton (UK); Spinlab (DE); Spherik posal and to kick-start a series of celebrations worldwide,
(RO); Bright Pixel (PT); The Next Web (NL); Ogilvy (ES); CERN will host a 30th anniversary event on the morning of
IPTector (DK), and ERCIM which hosts the European 12 March 2019 in partnership with the World Wide Web
branch of the World Wide Web Consortium (W3C). Consortium (W3C) and the World Wide Web Foundation.
http://www.datamarketservices.eu/ This anniversary event will be webcast.
@datamarketsvc https://cern.ch/web30
I.S.I. – Industrial Systems Institute VTT Technical Research Centre of Finland Ltd
Patras Science Park building PO Box 1000
Platani, Patras, Greece, GR-26504 FIN-02044 VTT, Finland
www.isi.gr www.vttresearch.com