Professional Documents
Culture Documents
Solution Documentation - AI FOR ENERGY
Solution Documentation - AI FOR ENERGY
SOLUTION TITLE
STATEMENT OF PURPOSE
The questions in this section are aimed at providing an overview of the solution owner, the documentation team, the problem, the solution and the intended uses for the solution.
INTRODUCTION
PROBLEM DEFINITION
SOLUTION
USAGE
What are the key procedures followed while using the solution?
How is the input provided? By whom? How is the output returned?
What are the domains and applications the solution was tested on or used
for?
Were domain experts involved in the development, testing, and deployment? Please
elaborate
List applications that the solution has been used for in the past
Please provide information about these applications or relevant pointers. Please
provide key performance results for those applications.
Any other comments?
DATASET
MOTIVATION
The questions in this section are primarily intended to encourage dataset creators to clearly articulate their reasons for creating the dataset and to promote
transparency about funding interests
Who created this dataset (e.g., which team, research group) and on behalf
of which entity (e.g., company, institution, organization)?
COMPOSITION
Dataset creators should read through the questions in this section prior to any data collection and then provide answers once collection is complete. Most of
these questions are intended to provide dataset consumers with the information they need to make informed decisions about using the dataset for specific
tasks.The answers to some of these questions reveal information about compliance with the EU’s General Data Protection Regulation(GDPR) or comparable
regulations in other jurisdictions
How many instances are there in total (of each type, if appropriate)?
Does the dataset contain all possible instances or is it a sample (not
necessarily random) of instances from a larger set?
If the dataset is a sample, then what is the larger set? Is the sample representative of
the larger set (e.g., geographic coverage)? If so, please describe how this
representativeness was validated/verified. If it is not representative of the larger set,
please describe why not (e.g., to cover a more diverse range of instances, because
instances were withheld or unavailable).
What data does each instance consist of? “Raw” data (e.g., unprocessed
text or images)or features?
In either case, please provide a description
Does the dataset contain data that might be considered confidential (e.g.,
data that is protected by legal privilege or by doctorpatient confidentiality,
data that includes the content of individuals non-public communications)?
If so, please provide a description
Does the dataset contain data that, if viewed directly, might be offensive,
insulting, threatening, or might otherwise cause anxiety?
If so, please describe why.
Does the dataset relate to people?
If not, you may skip the remaining questions in this section.
Does the dataset contain data that might be considered sensitive in any
way (e.g., data that reveals racial or ethnic origins, sexual orientations,
religious beliefs, political opinions or union memberships, or locations;
financial or health data; biometric or genetic data; forms of government
identification, such as social security numbers; criminal history)?
If so, please provide a description
COLLECTION PROCESS
As with the previous section, dataset creators should read through these questionspriortoanydatacollectiontoflagpotentialissuesandthenprovide answers once
collection is complete. In addition to the goals of the prior section, the answers to questions here may provide information that allow
otherstoreconstructthedatasetwithoutaccesstoit.
If the dataset is a sample from a larger set, what was the sampling strategy
(e.g., deterministic, probabilistic with specific sampling probabilities)?
Did you collect the data from the individuals in question directly, or obtain
it via third parties or other sources (e.g., websites)?
Did the individuals in question consent to the collection and use of their
data?
If so, please describe (or show with screenshots or other information) how consent
was requested and provided, and provide a link or other access point to, or otherwise
reproduce, the exact language to which the individuals consented.
Has an analysis of the potential impact of the dataset and its use on data
subjects (e.g., a data protection impact analysis)been conducted?
If so, please provide a description of this analysis, including the outcomes, as well as a
link or other access point to any supporting documentation.
PREPROCESSING/CLEANING/LABELLING
Dataset creators should read through these questions prior to any preprocessing, cleaning, or labeling and then provide answers once these tasks are
complete. The questions in this section are intended to provide dataset consumers with the information they need to determine whether the “raw” data has
been processed in ways that are compatible with their chosen tasks. For example, text that has been converted into a “bag-of-words” is not suitable for tasks
involving word order
USES
These questions are intended to encourage dataset creators to reflect on the tasks for which the dataset should and should not be used. By explicitly
highlighting these tasks, dataset creators can help dataset consumers to make informed decisions, thereby avoiding potential risks or harms.
Is there a repository that links to any or all papers or systems that use the
dataset?
If so, please provide a link or other access point
Is there anything about the composition of the dataset or the way it was
collected and preprocessed/cleaned/labeled that might impact future
uses?
For example, is there anything that a future user might need to know to avoid uses
that could result in unfair treatment of individuals or groups (e.g., stereotyping,
quality of service issues) or other undesirable harms (e.g., financial harms, legal risks)
If so, please provide a description. Is there anything a future user could do to mitigate
these undesirable harms?
Are there tasks for which the dataset should not be used?
If so, please provide a description.
MAINTENANCE
As with the previous section, dataset creators should provide answers to these questions prior to distributing the dataset. These questions are intended to
encourage dataset creators to plan for dataset maintenance and communicate this plan with dataset consumers.
Is there an erratum?
If so, please provide a link or other access point.
Will the dataset be updated (e.g., to correct labeling errors, add new
instances, delete instances)?
If so, please describe how often, by whom, and how updates will be communicated
to users (e.g., mailing list, GitHub)?
DETAILS
POSTER CONCEPT
RESULT
A description of results with central tendency (e.g. mean) & variation (e.g.
error bars)
SAFETY
The following questions aim to offer insights about potential unintentional harms, and m
GENERAL
Are you aware of possible examples of bias, ethical issues, or other safety
risks as a result of using the solution?
Were the possible sources of bias or unfairness analyzed? Where do they arise from:
the data? the particular techniques being implemented? other sources? Is there any
mechanism for redress if individuals are negatively affected?
Do you use data from or make inferences about individuals or groups of
individuals. Have you obtained their consent?
How was it decided whose data to use or about whom to make inferences? Do these
individuals know that their data is being used or that inferences are being made
about them? What were they told? When were they made aware? What kind of
consent was needed from them? What were the procedures for gathering consent?
Please attach the consent form to this declaration. What are the potential risks to
these individuals or groups? Might the service output interfere with individual rights?
How are these risks being handled or minimized? What trade-offs were made
between the rights of these individuals and business interests? Do they have the
option to withdraw their data? Can they opt out from inferences being made about
them? What is the withdrawal procedure?
EXPLAINABILITY
FAIRNESS
Does the solution implement and perform any bias detection and
remediation?
Please describe model bias policies that were checked, bias checking methods, and
results (e.g., disparate error rates across different groups). What procedures were
used to perform the remediation? Please provide links or references to
corresponding technical papers. Please provide details about the value of any bias
estimates before and after such remediation. How did the value of other
performance metrics change as a result?
SECURITY
The following questions aim to assess the susceptibility to deliberate ha
LINKS
Link to training and evaluation code
Link to dataset(s)
MISCELLANEOUS
DETAILS
POSTER CONCEPT
SAFETY
potential unintentional harms, and mitigation efforts to eliminate or minimize those harms.
GENERAL
EXPLAINABILITY
FAIRNESS
SECURITY
ssess the susceptibility to deliberate harms such as attacks by adversaries
LINKS
MISCELLANEOUS
PROOF OF CONCEPT AND MOCKS
DETAILS
POSTER CONCEPT
RESULT
A description of results with central tendency (e.g. mean) & variation (e.g.
error bars)
SAFETY
The following questions aim to offer insights about potential unintentional harms, and m
GENERAL
Are you aware of possible examples of bias, ethical issues, or other safety
risks as a result of using the solution?
Were the possible sources of bias or unfairness analyzed? Where do they arise from:
the data? the particular techniques being implemented? other sources? Is there any
mechanism for redress if individuals are negatively affected?
Do you use data from or make inferences about individuals or groups of
individuals. Have you obtained their consent?
How was it decided whose data to use or about whom to make inferences? Do these
individuals know that their data is being used or that inferences are being made
about them? What were they told? When were they made aware? What kind of
consent was needed from them? What were the procedures for gathering consent?
Please attach the consent form to this declaration. What are the potential risks to
these individuals or groups? Might the service output interfere with individual rights?
How are these risks being handled or minimized? What trade-offs were made
between the rights of these individuals and business interests? Do they have the
option to withdraw their data? Can they opt out from inferences being made about
them? What is the withdrawal procedure?
EXPLAINABILITY
FAIRNESS
Does the solution implement and perform any bias detection and
remediation?
Please describe model bias policies that were checked, bias checking methods, and
results (e.g., disparate error rates across different groups). What procedures were
used to perform the remediation? Please provide links or references to
corresponding technical papers. Please provide details about the value of any bias
estimates before and after such remediation. How did the value of other
performance metrics change as a result?
SECURITY
The following questions aim to assess the susceptibility to deliberate ha
LINKS
Link to training and evaluation code
Link to dataset(s)
MISCELLANEOUS
DETAILS
POSTER CONCEPT
SAFETY
potential unintentional harms, and mitigation efforts to eliminate or minimize those harms.
GENERAL
EXPLAINABILITY
FAIRNESS
SECURITY
ssess the susceptibility to deliberate harms such as attacks by adversaries
LINKS
MISCELLANEOUS
ANALYTICAL MODEL
MODEL DETAILS
EVALUATION
In addition to the solution provider, was this service tested by any third
party?
Please list all third parties that performed the testing. Also, please include
information about the tests and test results.
RESULT
A description of results with central tendency (e.g. mean) & variation (e.g.
error bars)
ENVIRONMENT
SAFETY
The following questions aim to offer insights about potential unintentional harms, and m
GENERAL
Are you aware of possible examples of bias, ethical issues, or other safety
risks as a result of using the solution?
Were the possible sources of bias or unfairness analyzed? Where do they arise from:
the data? the particular techniques being implemented? other sources? Is there any
mechanism for redress if individuals are negatively affected?
EXPLAINABILITY
FAIRNESS
Does the solution implement and perform any bias detection and
remediation?
Please describe model bias policies that were checked, bias checking methods, and
results (e.g., disparate error rates across different groups). What procedures were
used to perform the remediation? Please provide links or references to
corresponding technical papers. Please provide details about the value of any bias
estimates before and after such remediation. How did the value of other
performance metrics change as a result?
CONCEPT DRIFT
Does your solution make updates to its behavior based on newly ingested
data?
Is the new data uploaded by users? Is it generated by an automated
process? Are the patterns in the data largely static or do they change over
time? Are there any performance guarantees/bounds? Does the service
have an automatic feedback/retraining loop, or is there a human in the
loop?
How is the solution tested and monitored for model or performance drift
over time?
If applicable, describe any relevant testing along with test results.
How can the solution be checked for correct, expected output when new
data is added?
Does the solution allow for checking for differences between training and
usage data?
Does it deploy mechanisms to alert the user of the difference?
SECURITY
The following questions aim to assess the susceptibility to deliberate ha
How could this solution be attacked or abused?
Please describe.
LINKS
Link to training and evaluation code
Link to dataset(s)
MISCELLANEOUS
MODEL DETAILS
EVALUATION
ENVIRONMENT
GENERAL
EXPLAINABILITY
FAIRNESS
CONCEPT DRIFT
SECURITY
ssess the susceptibility to deliberate harms such as attacks by adversaries
LINKS
MISCELLANEOUS
MODEL
MODEL DETAILS
EVALUATION
In addition to the solution provider, was this service tested by any third
party?
Please list all third parties that performed the testing. Also, please include
information about the tests and test results.
RESULT
A description of results with central tendency (e.g. mean) & variation (e.g.
error bars)
ENVIRONMENT
SAFETY
The following questions aim to offer insights about potential unintentional harms, and m
GENERAL
Are you aware of possible examples of bias, ethical issues, or other safety
risks as a result of using the solution?
Were the possible sources of bias or unfairness analyzed? Where do they arise from:
the data? the particular techniques being implemented? other sources? Is there any
mechanism for redress if individuals are negatively affected?
EXPLAINABILITY
FAIRNESS
Does the solution implement and perform any bias detection and
remediation?
Please describe model bias policies that were checked, bias checking methods, and
results (e.g., disparate error rates across different groups). What procedures were
used to perform the remediation? Please provide links or references to
corresponding technical papers. Please provide details about the value of any bias
estimates before and after such remediation. How did the value of other
performance metrics change as a result?
CONCEPT DRIFT
Does your solution make updates to its behavior based on newly ingested
data?
Is the new data uploaded by users? Is it generated by an automated
process? Are the patterns in the data largely static or do they change over
time? Are there any performance guarantees/bounds? Does the service
have an automatic feedback/retraining loop, or is there a human in the
loop?
How is the solution tested and monitored for model or performance drift
over time?
If applicable, describe any relevant testing along with test results.
How can the solution be checked for correct, expected output when new
data is added?
Does the solution allow for checking for differences between training and
usage data?
Does it deploy mechanisms to alert the user of the difference?
SECURITY
The following questions aim to assess the susceptibility to deliberate ha
How could this solution be attacked or abused?
Please describe.
LINKS
Link to training and evaluation code
Link to dataset(s)
MISCELLANEOUS
MODEL DETAILS
EVALUATION
ENVIRONMENT
GENERAL
EXPLAINABILITY
FAIRNESS
CONCEPT DRIFT
SECURITY
ssess the susceptibility to deliberate harms such as attacks by adversaries
LINKS
MISCELLANEOUS
HARDWARE SOLUTION
MODEL DETAILS
EVALUATION
Is there a way to verify the performance metrics (e.g., via a service API )?
Briefly describe how a third party could independently verify the performance of the
service. Are there benchmarks publicly available and adequate for testing the service
In addition to the solution provider, was this service tested by any third
party?
Please list all third parties that performed the testing. Also, please include
information about the tests and test results.
RESULT
A description of results with central tendency (e.g. mean) & variation (e.g.
error bars)
ENVIRONMENT
SAFETY
The following questions aim to offer insights about potential unintentional harms, and m
GENERAL
Are you aware of possible examples of bias, ethical issues, or other safety
risks as a result of using the solution?
Were the possible sources of bias or unfairness analyzed? Where do they arise from:
the data? the particular techniques being implemented? other sources? Is there any
mechanism for redress if individuals are negatively affected?
EXPLAINABILITY
FAIRNESS
Does the solution implement and perform any bias detection and
remediation?
Please describe model bias policies that were checked, bias checking methods, and
results (e.g., disparate error rates across different groups). What procedures were
used to perform the remediation? Please provide links or references to
corresponding technical papers. Please provide details about the value of any bias
estimates before and after such remediation. How did the value of other
performance metrics change as a result?
CONCEPT DRIFT
Does your solution make updates to its behavior based on newly ingested
data?
Is the new data uploaded by users? Is it generated by an automated
process? Are the patterns in the data largely static or do they change over
time? Are there any performance guarantees/bounds? Does the service
have an automatic feedback/retraining loop, or is there a human in the
loop?
How is the solution tested and monitored for model or performance drift
over time?
If applicable, describe any relevant testing along with test results.
How can the solution be checked for correct, expected output when new
data is added?
Does the solution allow for checking for differences between training and
usage data?
Does it deploy mechanisms to alert the user of the difference?
SECURITY
The following questions aim to assess the susceptibility to deliberate ha
How could this solution be attacked or abused?
Please describe.
LINKS
Link to training and evaluation code
Link to dataset(s)
MISCELLANEOUS
MODEL DETAILS
EVALUATION
ENVIRONMENT
GENERAL
EXPLAINABILITY
FAIRNESS
CONCEPT DRIFT
SECURITY
ssess the susceptibility to deliberate harms such as attacks by adversaries
LINKS
MISCELLANEOUS