Professional Documents
Culture Documents
KM Notes Unit-2
KM Notes Unit-2
2
Page
Detailed data – EIS provides absolute data from its existing database.
Integrate external and internal data – EIS integrates integrate external and internal data. The
external data collected from various sources.
Presenting information – EIS represents available data in graphical form which helps to
analyze it easily.
Trend analysis – EIS helps executives of the organizations to data prediction based on trend
data.
These information found to be often external, unstructured and even uncertain. The context of
such information provided cannot be known beforehand
4
Page
2.2 Advantages of ESS
It is easy for senior most executive to use
Ability to analyze trends
Augmentation of managers' leadership capabilities
Enhancement of personal thinking and decision-making
Contribution to strategic control flexibility
Broadened the organizational competitiveness in the market place
Instruments of change
Increased executive time horizons.
Better reporting system
Improved mental model of business executive
Help improve consensus building and communication
Improve office automation
Reduction in time to find information
Early identification of company performance
Detailed examination of critical success factor
Better understanding
Time management
Increased communication capacity and quality
Knowledge base contains the facts about the specific expert area and heuristics that describe the
reasoning procedures of an expert on the subject.
The inference engine contains the logic of reaching an inference from the stored data and from
the knowledge base.
Expert systems may be developed by using either the programming language like LISP,
PROLOG or C or by using the expert system packages. Using the expert system packages, one
6
Page
can design an expert system that combines the features of DSS and expert systems. This
integration makes the application development process easier and faster for the end users.
4. Expert System
An expert system is a computer program that is designed to solve complex problems and to
provide decision-making ability like a human expert. It performs this by extracting knowledge
from its knowledge base using the reasoning and inference rules according to the user queries.
The expert system is a part of AI, and the first ES was developed in the year 1970, which was the
first successful approach of artificial intelligence. It solves the most complex issue as an expert
by extracting the knowledge stored in its knowledge base. The system helps in decision making
for compsex problems using both facts and heuristics like a human expert. It is called so
because it contains the expert knowledge of a specific domain and can solve any complex
problem of that particular domain. These systems are designed for a specific domain, such
as medicine, science, etc.
The performance of an expert system is based on the expert's knowledge stored in its knowledge
base. The more knowledge stored in the KB, the more that system improves its performance.
One of the common examples of an ES is a suggestion of spelling errors while typing in the
Google search box.
Below is the block diagram that represents the working of an expert system:
7
Page
Examples of the Expert System:
DENDRAL: It was an artificial intelligence project that was made as a chemical analysis
expert system. It was used in organic chemistry to detect unknown organic molecules
with the help of their mass spectra and knowledge base of chemistry.
MYCIN: It was one of the earliest backward chaining expert systems that was designed
to find the bacteria causing infections like bacteraemia and meningitis. It was also used
for the recommendation of antibiotics and the diagnosis of blood clotting diseases.
PXDES: It is an expert system that is used to determine the type and level of lung cancer.
To determine the disease, it takes a picture from the upper body, which looks like the
shadow. This shadow identifies the type and degree of harm.
CaDeT: The CaDet expert system is a diagnostic support system that can detect cancer at
early stages.
3. Knowledge Base
o The knowledgebase is a type of storage that stores knowledge acquired from the different
experts of the particular domain. It is considered as big storage of knowledge. The more
the knowledge base, the more precise will be the Expert System.
o It is similar to a database that contains information and rules of a particular domain or
subject.
o One can also view the knowledge base as collections of objects and their attributes. Such
as a Lion is an object and its attributes are it is a mammal, it is not a domestic animal, etc.
Knowledge Representation: It is used to formalize the knowledge stored in the knowledge base
using the If-else rules.
Knowledge Acquisitions: It is the process of extracting, organizing, and structuring the domain
knowledge, specifying the rules to acquire the knowledge from various experts, and store that
knowledge into the knowledge base.
o The response of the expert system may get wrong if the knowledge base contains the
wrong information.
o Like a human being, it cannot produce a creative output for different scenarios.
o Its maintenance and development costs are very high.
o Knowledge acquisition for designing is much difficult.
o For each domain, we require a specific ES, which is one of the big limitations.
o It cannot learn from itself and hence requires manual updates.
6. Data Warehouse
A data warehouse is a relational database that is designed for query and analysis rather than for
transaction processing. It usually contains historical data derived from transaction data, but it can
include data from other sources. It separates analysis workload from transaction workload and
enables an organization to consolidate data from several sources.
In addition to a relational database, a data warehouse environment includes an extraction,
13
1. Subject-Oriented
A data warehouse is subject oriented as it offers information regarding a theme instead of
companies' ongoing operations. These subjects can be sales, marketing, distributions, etc.
A data warehouse never focuses on the ongoing operations. Instead, it put emphasis on modeling
and analysis of data for decision making. It also provides a simple and concise view around the
specific subject by excluding data which not helpful to support the decision process.
2. Integrated
In Data Warehouse, integration means the establishment of a common unit of measure for all
similar data from the dissimilar database. The data also needs to be stored in the Datawarehouse
in common and universally acceptable manner.
A data warehouse is developed by integrating data from varied sources like a mainframe,
relational databases, flat files, etc. Moreover, it must keep consistent naming conventions,
format, and coding.
14
This integration helps in effective analysis of data. Consistency in naming conventions, attribute
measures, encoding structure etc. have to be ensured. Consider the following example:
Page
In the above example, there are three different application labeled A, B and C. Information stored in these appl
In Application A gender field store logical values like M or F
In Application B gender field is a numerical value,
In Application C application, gender field stored in the form of a character value.
3. Time-Variant
The time horizon for data warehouse is quite extensive compared with operational systems. The
data collected in a data warehouse is recognized with a particular period and offers information
from the historical point of view. It contains an element of time, explicitly or implicitly.
One such place where Datawarehouse data display time variance is in in the structure of the
record key. Every primary key contained with the DW should have either implicitly or explicitly
15
4. Non-volatile
Data warehouse is also non-volatile means the previous data is not erased when new data is
entered in it.
Data is read-only and periodically refreshed. This also helps to analyze historical data and
understand what & when happened. It does not require transaction process, recovery and
concurrency control mechanisms.
Activities like delete, update, and insert which are performed in an operational application
environment are omitted in Data warehouse environment. Only two types of data operations
performed in the Data Warehousing are
1. Data loading
2. Data access
Complex program must be coded to This kind of issues does not happen because
make sure that data upgrade processes data update is not performed.
maintain high integrity of the final
product.
Data is placed in a normalized form to ensure Data is not stored in normalized form.
minimal redundancy.
Data Warehouse Architecture is complex as it’s an information system that contains historical
and commutative data from multiple sources. There are 3 approaches for constructing Data
Warehouse layers: Single Tier, Two tier and Three tier. This 3 tier architecture of Data
Warehouse is explained as below.
Single-tier architecture
The objective of a single layer is to minimize the amount of data stored. This goal is to remove data redundanc
Two-tier architecture
Two-layer architecture is one of the Data Warehouse layers which separates physically available
sources and data warehouse. This architecture is not expandable and also not supporting a large
number of end-users. It also has connectivity problems because of network limitations.
Three-Tier Data Warehouse Architecture
This is the most widely used Architecture of Data
Warehouse. It consists of the Top, Middle and Bottom Tier.
1. Bottom Tier: The database of the Datawarehouse servers as the bottom tier. It is usually
a relational database system. Data is cleansed, transformed, and loaded into this layer
using back-end tools.
2. Middle Tier: The middle tier in Data warehouse is an OLAP server which is
17
implemented using either ROLAP or MOLAP model. For a user, this application tier
Page
presents an abstracted view of the database. This layer also acts as a mediator between
the end-user and the database.
3. Top-Tier: The top tier is a front-end client layer. Top tier is the tools and API that you
connect and get data out from the data warehouse. It could be Query tools, reporting
tools, managed query tools, Analysis tools and Data mining tools.
The Data Warehouse is based on an RDBMS server which is a central information repository
that is surrounded by some key Data Warehousing components to make the entire environment
functional, manageable and accessible.
by the fact that traditional RDBMS system is optimized for transactional database processing and
not for data warehousing. For instance, ad-hoc query, multi-table joins, aggregates are resource
intensive and slow down performance.
Hence, alternative approaches to Database are used as listed below-
In a datawarehouse, relational databases are deployed in parallel to allow for scalability.
Parallel relational databases also allow shared memory or shared nothing model on
various multiprocessor configurations or massively parallel processors.
New index structures are used to bypass relational table scan and improve speed.
Use of multidimensional database (MDDBs) to overcome any limitations which are
placed because of the relational Data Warehouse Models. Example: Essbase from Oracle.
3. Metadata
The name Meta Data suggests some high-level technological Data Warehousing Concepts.
19
However, it is quite simple. Metadata is data about data which defines the data warehouse. It
Page
Analysts frequently need to group, aggregate and join data. These operations in relational
databases are resource intensive. With OLAP data can be pre-calculated and pre-aggregated,
making analysis faster.
20
22
Page
OLAP databases are divided into one or more cubes. The cubes are designed in such a way that
creating and viewing reports become easy. OLAP stands for Online Analytical Processing.
Usually, data operations and analysis are performed using the simple spreadsheet, where data
values are arranged in row and column format. This is ideal for two-dimensional data. However,
OLAP contains multidimensional data, with data usually obtained from a different and unrelated
source. Using a spreadsheet is not an optimal option. The cube can store and analyze
multidimensional data in a logical and orderly manner.
The extracted data is cleaned and transformed. Data is loaded into an OLAP server (or OLAP
cube) where information is pre-calculated in advance for further analysis.
23
Page
9. Difference between OLTP and OLAP
OLTP and OLAP both are the online processing systems. OLTP is a transactional processing
while OLAP is an analytical processing system. OLTP is a system that manages transaction-
oriented applications on the internet for example, ATM. OLAP is an online system that reports to
multidimensional analytical queries like financial reporting, forecasting, etc.
The basic difference between OLTP and OLAP is that OLTP is an online database modifying
system, whereas, OLAP is an online database query answering system.
the tables in OLAP database may not be normalized. The example for OLAP is to view a
financial report, or budgeting, marketing management, sales report, etc
Page
Comparison Chart
BASIS FOR OLTP OLAP
COMPARISON
Basic It is an online transactional It is an online data retrieving and data
system and manages database analysis system.
modification.
Focus Insert, Update, Delete Extract data for analyzing that helps in
information from the database. decision making.
Data OLTP and its transactions are the Different OLTPs database becomes the
original source of data. source of data for OLAP.
29
Page