Professional Documents
Culture Documents
Project Report
Project Report
Five7 is a Gulbarga based technology and IT consulting firm focused on delivering the best
quality and cost effective solutions to clients. The company considers IT as the backbone of a
business ecosystem and provides complete solutions to address all aspects of the business. The
company’s approach focuses on new ways of business, combining IT innovation and adoption
while also an organizations current IT assets which is the business and technology strategies in
today’s environment. The company provides highly customized application development and
integration services.
Five7 is a company with core strength in technology training and software development.
We design, build and deliver solutions to suit the business requirements; be it in application
development or training, with commitment to time, quality, cost effectiveness and integrity.
We see our success in the transformation of our client’s business requirements to best practices
in software, with high ROI (Return on Investment). Our ERP Software solutions span across
mid segment manufacturing industries to Government enterprises, educational institutions and
retail marketing. Our Java based solutions cater to developing of Products in the web / mobile
space.
Project Development
Applications Group
This team is well equipped in terms of web based development, development of portals,
interactive sites; web enabled applications, and well versed with latest technologies.
Five7 caters to the needs of the IT industry in the following segments:
Training
Application Software (ERP & e-Business Applications)
Web based / Mobile based product development
Android App development
MOHAMMED AMAAN - With more than 4 years of experience in the industry, leads the
operations as Director and also as the Senior Manager – Business.
Five7 company, operations are in line to adhere to ISO 9001-2000 standards and our processes,
procedures and methodologies are continuously tuned towards achieving 100% customer
satisfaction.
The entire SDLC (Software Development Life Cycle) right from the initiation of the project to
the delivery and support is well defined as having the following milestones
Feasibility study
The Five7 Pvt limited has various departments like R&D , Education & Training Department,
Finance Department.
In Five7 company, I had a good place to work and good staff to help me out in all situations it
is very quite and pleasant so I can concentrate on the work. Even though I had a lot to travel on
the daily basis to reach my company but it was a good experience to work in company and
learnt many things.
1. Introduction...................................................................................4
Image Segmentation…......................................................................5
Region Growing Approach............................................................................4
Clustering 3
K-means............................................................................................4
Hierarchical Segmentation................................................................9
Thresholding..................................................................................................10
Design Steps..................................................................................................11
System Analysis............................................................................................12
2. Proposed System.........................................................................13
Fast discrete Curvelet Transformation..............................................14
Block Diagram..................................................................................17
Texture Analysis.............................................................................................19
Co-Occurrence Matrixs..................................................................................20
3. Neural Networks.........................................................................21
Neural network...............................................................................................24
The Used Neural Network..............................................................................28
Types of Neural Networks..............................................................................32
radical Basis Layer…...................................................................................36
PNN 38
5. Simulated Results........................................................................48
MRI Brain With Benign Case.......................................................................49
MRI Brain With Malignant Case...................................................................51
MRI Brain With Normal Case.......................................................................53
The Automatic Support Intelligent System is used to detect Brain Tumor through the
combination of neural network and fuzzy logic system. It helps in the diagnostic and aid in the
treatment of the brain tumor. The detection of the Brain Tumor is a challenging problem, due to
the structure of the Tumor cells in the brain. This project presents an analytical method that
enhances the detection of brain tumor cells in its early stages and to analyze anatomical
structures by training and classification of the samples in neural network system and tumor cell
segmentation of the sample using fuzzy clustering algorithm. The artificial neural network will
be used to train and classify the stage of Brain Tumor that would be benign, malignant or
normal.
The Fast discrete curvelet transformation is used to analysis texture of an image. In brain
structure analysis, the tissues which are WM and GM are extracted. Probabilistic Neural
Network with radial basis function is employed to implement an automated Brain Tumor
classification. Decision making is performed in two stages: feature extraction using GLCM and
the classification using PNN-RBF network. The segmentation is performed by fuzzy logic
system and its result would be used as a base for early detection of Brain Tumor which would
improves the chances of survival for the patient. The performance of this automated intelligent
system evaluates in terms of training performance and classification accuracies to provide the
precise and accurate results. The simulated result enhances and shows that classifier and
segmentation algorithm provides better accuracy than previous methodologies.
1
CHAPTER 1
INTRODUCTION
2
CHAPTER 1
INTRODUCTION
1 INTRODUCTION
3
manually. Thus, fully understanding the recognition, classification or clustering techniques is
essential to the developments of Neural Network systems particularly in medicine problems.
Segmentation of brain tissues in gray matter, white matter and tumor on medical images
is not only of high interest in serial treatment monitoring of “disease burden” in oncologic
imaging, but also gaining popularity with the advance of image guided surgical approaches.
Outlining the brain tumor contour is a major step in planning spatially localized radiotherapy
(e.g., Cyber knife, iMRT ) which is usually done manually on contrast enhanced T1-weighted
magnetic resonance images (MRI) in current clinical practice. On T1 MR Images acquired after
administration of a contrast agent (gadolinium), blood vessels and parts of the tumor, where the
contrast can pass the blood–brain barrier are observed as hyper intense areas. There are various
attempts for brain tumor segmentation in the literature which use a single modality, combine
multi modalities and use priors obtained from population atlases.
IMAGE SEGMENTATION
OVERVIEW
Segmentation problems are the bottleneck to achieve object extraction, object specific
measurements, and fast object rendering from multi-dimensional image data. Simple segmentation
techniques are based on local pixel-neighborhood classification. Such methods fail however to
global objects rather than local appearances and require often intensive operator assistance. The
reason is that the “logic” of a object does not necessarily follow that of its local image
representation. Local properties, such as textures, edginess, and ridgeness etc. do not always
represent connected features of a given object.
4
REGION GROWING APPROACH
Region growing technique segments image pixels that are belong to an object into
regions. Segmentation is performed based on some predefined criteria. Two pixels can be
grouped together if they have the same intensity characteristics or if they are close to each other.
It is assumed that pixels that are closed to each other and have similar intensity values are likely
to belong to the same object. The simplest form of the segmentation can be achieved through
thresholding and component labeling. Another method is to find region boundaries using edge
detection. Segmentation process, then, uses region boundary information to extract the regions.
The main disadvantage of region growing approach is that it often requires a seed point as the
starting point of the segmentation process. This requires user interaction. Due to the variations in
image intensities and noise, region growing can result in holes and over segmentation. Thus, it
sometimes requires post-processing of the segmentation result.
CLUSTERING
5
In the first case data are grouped in an exclusive way, so that if a certain datum belongs to a
definite cluster then it could not be included in another cluster. On the contrary the second type,
the overlapping clustering, uses fuzzy sets to cluster data, so that each point may belong to two
or more clusters with different degrees of membership. In this case, data will be associated to an
appropriate membership value. A hierarchical clustering algorithm is based on the union between
the two nearest clusters. The beginning condition is realized by setting every datum as a cluster.
After a few iterations it reaches the final clusters wanted
K-MEANS SEGMENTATION
K-means is one of the simplest unsupervised learning algorithms that solve the well-
known clustering problem. The procedure follows a simple and easy way to classify a given data
set through a certain number of clusters (assume k clusters) fixed a priori. The main idea is to
define k centroids, one for each cluster. These centroids should be placed in a cunning way
because of different location causes different result. So, the better choice is to place them as
much as possible far away from each other. The next step is to take each point belonging to a
given data set and associate it to the nearest centroids. When no point is pending, the first step is
completed and an early group age is done. At this point we need to re-calculate k new centroids
as centers of the clusters resulting from the previous step. After we have these k new centroids, a
new binding has to be done between the same data set points and the nearest new centroids. A
loop has been generated. As a result of this loop we may notice that the k centroids change their
location step by step until no more changes are done. In other words centroids do not move any
more. Finally, this algorithm aims at minimizing an objective function, in this case a squared
error function.
6
HIERARCHICAL SEGMENTATION
A goal of the subject analysis of the segmentation hierarchy is to identify the hierarchical
level at which the object of interest is represented by a single region segment. The object may
then be identified through its spectral and spatial characteristics. Additional clues for object
identification may be obtained from the behavior of the image segmentations at the hierarchical
segmentation level above and below the level at which the object of interest is represented by a
single region.
7
THRESHOLDING
The simplest method of image segmentation is called the thresholding method. This
method is based on a clip-level (or a threshold value) to turn a gray- scale image into a binary
image. The key of this method is to select the threshold value (or values when multiple-levels are
selected). Several popular methods are used in industry including the maximum entropy method,
Otsu's method (maximum variance), and k-means clustering. Recently, methods have been
developed for thresholding computed tomography (CT) images. The key idea is that, unlike
Otsu's method, the thresholds are derived from the radiographs instead of the (reconstructed)
image.
DESIGN STEPS
(1) Set the initial threshold T= (the maximum value of the image brightness
+ the minimum value of the image brightness)/2.
(2) Using T segment the image to get two sets of pixels B (all the
pixel values are less than T) and N (all the pixel values are greater
than T);
(3) Calculate the average value of B and N separately, mean ub and un.
(5) Repeat Second steps to fourth steps up to iterative conditions are met
and get necessary region from the brain image.
8
SYSTEM ANALYSIS
EXISTING TECHNIQUES
9
DRAWBACKS
DEFINITION
Hilbert basis, that is a complete Ortho normal system, for the Hilbert of
square
integrable functions. The Hilbert basis is constructed as the family of functions for integers
. This family is an Ortho normal system if it is ortho normal under the inner product
10
With convergence of the series understood to be convergence in norm. Such a representation of a
function f is known as a wavelet series. This implies that an ortho normal wavelet is self-dual.
DRAWBACKS
KNN CLASSIFIER
The same method can be used for regression, by simply assigning the property value for
the object to be the average of the values of itsk nearest neighbors. It can be useful to weight the
contributions of the neighbors, so that the nearer neighbors contribute more to the average than
the more distant ones. (A common weighting scheme is to give each neighbor a weight of 1/d,
where d is the distance to the neighbor. This scheme is a generalization of linear interpolation.)
11
DRAWBACKS
PROPOSED METHOD
MRI Brain image Classification and anatomical structure analysis are proposed based on,
(1) PNN-RBF Network for classification
(2) Fuzzy clustering for tumor detection and structural analysis
METHODOLOGIES
ADVANTAGES
(1) It can segment the Brain regions from the image accurately.
(2) It is useful to classify the Brain Tumor images for accurate detection.
(3) Brain Tumor will be detected in an early stages
12
CHAPTER 2
PROPOSED SYSTEM
13
CHAPTER 2
PROPOSED SYSTEM
Curvelets implementations are based on the original construction which uses apre-
processing step involving a special partitioning of phase-space followed by the ridge let
transform which is applied to blocks of data that are well localized in space and frequency.
BLOCK DIAGRAM
In the last two or three years, however, Curvelets have actually been redesigned in a
effort to make them easier to use and understand. As a result, the new construction is
considerably simpler and totally transparent. What is interesting here is that the new
mathematical architecture suggests innovative
14
algorithmic strategies, and provides the opportunity to improve upon earlier implementations.
The two new fast discrete curvelet transforms (FDCTs) which are simpler, faster, and less
redundant than existing proposals:
(1) Curvelets via USFFT, and
(2) Curvelets via Wrapping.
The block size can be changed at each scale level. The wrapping construction is shown is
taken to be a Cartesian array and ˆ f[n1, n2 ] denotes its 2-D discrete Fourier transform, then the
architecture of the FDCT via wrapping is as follows.
4) Apply the inverse 2-D FFT to each ˜ fj,l , hence collecting the
discrete coefficient .
15
TEXTURE ANALYSIS
2.3.1 OVERVIEW
Texture is that innate property of all surfaces that describes visual patterns, each having
properties of homogeneity. It contains important information about the structural arrangement of
the surface, such as; clouds, leaves, bricks, fabric, etc. It also describes the relationship of the
surface to the surrounding environment. In short, it is a feature that describes the distinctive
physical composition of a surface. Texture properties include:
(1) Coarseness
(2) Contrast
(3) Directionality
(4) Line-likeness
(5) Regularity
(6) Roughness
METHODS OF REPRESENTATION
There are three principal approaches used to describe texture; statistical, structural and
spectral…
16
(1) Statistical techniques characterize textures using the statistical properties
of the grey levels of the points/pixels comprising a surface image.
Typically, these properties are computed using: the grey level co-
occurrence matrix of the surface, or the wavelet transformation of the
surface.
(3) Spectral techniques are based on properties of the Fourier spectrum and
describe global periodicity of the grey levels of a surface by identifying
high-energy peaks in the Fourier spectrum.
For optimum classification purposes, what concern us are the statistical techniques of
characterization… This is because it is these techniques that result in computing texture
properties… The most popular statistical representations of texture are:
17
CO-OCCURRENCE MATRIX
Examples for the operator P are: “i above j”, or “i one position to the right and two below j”, etc.
This can also be illustrated as follows… Let t be a translation, then a co-occurrence matrix Ctof a
region is defined for every grey-level (a, b) by [1]:
18
Here, Ct(a, b) is the number of site-couples, denoted by (s, s + t) that are separated by a
translation vector t, with a being the grey-level of s, and b being the grey-level of s + t.
For example; with an 8 grey-level image representation and a vector t that considers
only one neighbor, we would find [1]:
At first the co-occurrence matrix is constructed, based on the orientation and distance
between image pixels. Then meaningful statistics are extracted from the matrix as the texture
representation. Haralick proposed the following texture features:
(1) Energy
(2) Contrast
(3) Correlation
(4) Homogeneity
19
ENERGY
CONTRAST
Contrast is the main diagonal near the moment of inertia, which measure the value of the
matrix is distributed and images of local changes in number, reflecting the image clarity and
texture of shadow depth.
CORRELATION COEFFICIENT
Measures the joint probability occurrence of the specified pixel pairs Correlation: sum
(sum ((x- μx)(y-μy)p(x , y)/σxσy))
HOMOGENEITY
Measures the closeness of the distribution of elements in the GLCM to the GLCM
diagonal.
Homogenity = sum (sum(p(x , y)/(1 + [x-y])))
20
CHAPTER 3
PROBABILISTIC NEURAL NETWORKS
21
CHAPTER 3
PROBABILISTIC NEURAL NETWORKS
Neural networks are predictive models loosely based on the action of biological neurons.
The selection of the name “neural network” was one of the great PR successes of the Twentieth
Century. It certainly sounds more exciting than a technical description such as “A network of
weighted, additive values with nonlinear transfer functions”. However, despite the name, neural
networks are far from “thinking machines” or “artificial brains”. A typical artificial neural
network might have a hundred neurons. In comparison, the human nervous system is believed to
have about 3x1010 neurons. We are still light years from “Data”.
The original “Perception” model was developed by Frank Rosenblatt in 1958. Rosenblatt’s
model consisted of three layers, (1) a “retina” that distributed inputs to the second layer, (2)
“association units” that combine the inputs with weights and trigger a threshold step function
which feeds to the output layer, (3) the output layer which combines the values. Unfortunately,
the use of a step function in the neurons made the perceptions difficult or impossible to train. A
critical analysis of perceptions published in 1969 by Marvin Musky and Seymour Paper pointed
out a number of critical weaknesses of perceptions, and, for a period of time, interest in
perceptions waned.
Interest in neural networks was revived in 1986 when David Rumelhart, Geoffrey Hinton
and Ronald Williams published “Learning Internal Representations by Error Propagation”. They
proposed a multilayer neural network with nonlinear but differentiable transfer functions that
avoided the pitfalls of the original perception’s step functions. They also provided a reasonably
effective training algorithm for neural networks.
22
In this paper, we investigate the possibility of utilizing the detection capability of the
neural networks as a diagnostic tool for early breast cancer. Breast cancer is one of leading
causes of women death in the world. Detection and localization of tumor at early stages is the
only way to decrease the mortality rate. X-ray mammography is the most used diagnostic tool for
breast cancer screening. However, it has important limitations with. For example, it has high
false -negative and -positive rates that may reach up to 34%. Thus, the development of imaging
modalities which enhance, complement, or replace X-ray mammography has been a priority in
the medical imaging research. Recently, several microwave imaging methods have been
developed. Microwave ultra-wideband (UWB) imaging is currently one of promising methods
that are under investigation by many research groups around the world. This method involves
transmitting UWB signals through the breast tissue and records the scattered signals from
different locations surrounding the breast using compact UWB antennas. The basis of this
method is the difference in the electrical properties between the tumors and the healthy tissues.
It has been shown by different research groups worldwide that the healthy fat tissues of
the breast have a dielectric constant that ranges between S and 9 and conductivity between 0.02
Sims and 0.2 S/m. On the hand, the malignant tissues have a dielectric constant of around 60 and
conductivity around 2 Sim. Thus, it is clear that there is a significant contrast between the
electrical characteristics of the healthy and malignant tissues. From the electromagnetic point of
view, this contrast means a significant difference in the scattering properties of those tissues.
From the neural network point of view, it is possible to say that the difference in the scattering
properties of healthy and malignant tissues means that the scattered signals recorded in different
places around the breast have the signature of existence of tumor. The comparison between those
received signals recorded in different places gives an indication about the place of tumor.
23
Reviewing the literature shows that most of the research on using neural networks for
breast cancer detection were aimed at enabling radiologist to make accurate diagnoses of breast
cancer on X-ray mammograms. For example, several neural network methods were used in [IS]
for automated classification of clustered micro-calcifications in mammograms.
Concerning the UWB systems, the neural network has been used extensively to detect the
direction of arrival of ultra wideband signals. It is also utilized in a simple manner to detect and
locate tumor of 2.S mm radius along certain axis
(one dimension) and also in two dimensions but by rotating the place of receiving the signal by
steps of 1 degree around the breast.
This paper presents the use of neural network to detect and locate tumors in breast with
sizes down to 1 mm in radius by using four permanent probes around the breast. A three
dimensional breast model is built for the purpose of this investigation. A pulsed plane
electromagnetic wave is sent towards the breast and the scattered signals are collected by four
antennas surrounding the breast. The transmitted pulse occupies the UWB frequency range from
GHz to 10.6 GHz.
A simple feed-forward back-propagation neural network is then used to detect the tumor and locate
its position.
24
THE USED NEURAL NETWORK
Neural network is the best tool in recognition and discrimination between different sets
of signals. To get best results using the neural network, it is necessary to choose a suitable
architecture and learning algorithm. Unfortunately there is no guaranteed method to do that. The
best way to do that is to choose what is expected to be suitable according to our previous
experience and then to expand or shrink the neural network size until a reasonable output is
obtained. In this work we tried different sizes for the neural network using MA TLAB and we
found that the best in our case is the model shown in Fig. 2. It has an input layer with 2000
inputs, first hidden layer with 11 nodes, and T ANSIG transfer function, second hidden layer
with 7 nodes, and T ANSIG transfer function, and output layer with PURELIN transfer function
and 2 outputs. One of the two outputs is used for the detection of tumor, and the other for the
localization. T ANSIG transfer function is selected to limit the signal between -1 and1.
For the output layer, PURELIN transfer function is chosen to give all the possible cases
for the location of tumor. The proposed neural network has been used to detect and locate the
tumor in two cases. The first case was to detect and locate a tumor in a two-dimensional sector of
the breast model. The location of tumor was considered randomly at the center and in any of the
four quadrature . The second case was to detect and locate tumor anywhere in the three-
dimensional model. The neural network has been trained using 100 sets of inputs using the
25
training function (TRAINSCG). Additional 40sets of inputs were used to test the performance of
each neural network.
26
THE MULTILAYER PERCEPTRON NEURAL NETWORK MODEL
This network has an input layer (on the left) with three neurons, one hidden layer (in the
middle) with three neurons and an output layer (on the right) with three neurons.
There is one neuron in the input layer for each predictor variable. In the case of
categorical variables, N-1 neurons are used to represent the N categories of the variable.
INPUT LAYER
A vector of predictor variable values (x1...xp) is presented to the input layer. The input
layer (or processing before the input layer) standardizes these values so
27
that the range of each variable is -1 to 1. The input layer distributes the values to each of the
neurons in the hidden layer. In addition to the predictor variables, there is a constant input of 1.0,
called the bias that is fed to each of the hidden layers; the bias is multiplied by a weight and
added to the sum going into the neuron.
HIDDEN LAYER
Arriving at a neuron in the hidden layer, the value from each input neuron is multiplied
by a weight (wji), and the resulting weighted values are added together producing a combined
value uj. The weighted sum (uj) is fed into a transfer function, σ, which outputs a value hj. The
outputs from the hidden layer are distributed to the output layer.
OUTPUT LAYER
Arriving at a neuron in the output layer, the value from each hidden layer neuron is
multiplied by a weight (wkj), and the resulting weighted values are added together producing a
combined value vj. The weighted sum (vj) is fed into a transfer function, σ, which outputs a value
yk. The y values are the outputs of the network.
If a regression analysis is being performed with a continuous target variable, then there is
a single neuron in the output layer, and it generates a single y value. For classification problems
with categorical target variables, there are N neurons in the output layer producing N values, one
for each of the N categories of the target variable.
28
PROBABILISTIC NEURAL NETWORKS (PNN)
Probabilistic (PNN) and General Regression Neural Networks (GRNN) have similar
architectures, but there is a fundamental difference: Probabilistic networks perform classification
where the target variable is categorical, whereas general regression neural networks perform
regression where the target variable is continuous. If you select a PNN/GRNN network, DTREG
will automatically select the correct type of network based on the type of target variable.
ARCHITECTURE OF A PNN
Fig 3.6
(1) Input layer - There is one neuron in the input layer for each predictor
variable. In the case of categorical variables, N-1 neurons are used where N
is the number of categories. The input neurons (or processing before the
input layer) standardizes the range of the values by subtracting the median
29
and dividing by the inter quartile range. The input neurons then feed the values to
each of the neurons in the hidden layer.
(2) Hidden layer - This layer has one neuron for each case in the training data
set. The neuron stores the values of the predictor variables for the case along
with the target value. When presented with the x vector of input values from
the input layer, a hidden neuron computes the Euclidean distance of the test
case from the neuron’s center point and then applies the RBF kernel function
using the sigma value(s). The resulting value is passed to the neurons in the
pattern layer.
(3) Pattern layer / Summation layer - The next layer in the network is different
for PNN networks and for GRNN networks. For PNN networks there is one
pattern neuron for each category of the target variable. The actual target
category of each training case is stored with each hidden neuron; the
weighted value coming out of a hidden neuron is fed only to the pattern
neuron that corresponds to the hidden neuron’s category. The pattern
neurons add the values for the class they represent (hence, it is a weighted
vote for that category).
For GRNN networks, there are only two neurons in the pattern layer. One neuron
is the denominator summation unit the other is the numerator summation unit. The
denominator summation unit adds up the weight values coming from each of the hidden
neurons. The numerator summation unit adds up the weight values multiplied by the
actual target value for each hidden neuron.
30
(4) Decision layer - The decision layer is different for PNN and GRNN
networks. For PNN networks, the decision layer compares the weighted
votes for each target category accumulated in the pattern layer and uses the
largest vote to predict the target category.
For GRNN networks, the decision layer divides the value accumulated in the
numerator summation unit by the value in the denominator summation unit and uses the
result as the predicted target value.
The following diagram is actual diagram or propose network used in our project
Fig 3.6b
The input vector, denoted as p, is presented as the black vertical bar. Its dimension is R × 1.
In this paper, R = 3.
31
(2) Radial Basis Layer:
In Radial Basis Layer, the vector distances between input vector p and the weight vector
made of each row of weight matrix W are calculated. Here, the vector distance is defined as the
dot product between two vectors [8]. Assume the dimension of W is Q×R. The dot product
between p and the i-th row of W produces the i-th element of the distance vector ||W-p||, whose
dimension is Q×1. The minus symbol, “-”, indicates that it is the distance between vectors. Then,
the bias vector b is combined with ||W- p|| by an element-by-element multiplication,
.The result is denoted as n = ||W- p|| ..p. The transfer function in PNN has built into a distance
criterion with respect to a center. In this paper, it is defined as radbas (n)
= 2 n e- (1) each element of n is substituted into Eq. 1 and produces corresponding element of a,
the output vector of Radial Basis Layer. The i-th element of a can be represented as ai = radbas
(||Wi - p|| ..bi) (2) where Wi is the vector made of the i- th row of W and bi is the i-th element of
bias vector b.
The i-th element of a equals to 1 if the input p is identical to the ith row of input weight matrix
W. A radial basis neuron with a weight vector close to the input vector p produces a value near 1
and then its output weights in the competitive layer will pass their values to the competitive
function. It is also possible that several elements of a are close to 1 since the input pattern is
close to several training patterns.
32
(2) Competitive Layer:
Fig 3.6.2 a
33
that a predicted target value of an item is likely to be about the same as other items that
have close values of the predictor variables. Consider this figure:
Assume that each case in the training set has two predictor variables, x and
y. The cases are plotted using their x,y coordinates as shown in the figure. Also assume that the
target variable has two categories, positive which is denoted by a square and negative which is
denoted by a dash. Now, suppose we are trying to predict the value of a new case represented by
the triangle with predictor values x=6, y=5.1. Should we predict the target as positive or
negative?
Notice that the triangle is position almost exactly on top of a dash representing a negative
value. But that dash is in a fairly unusual position compared to the other dashes which are
clustered below the squares and left of center. So it could be that the underlying negative value is
an odd case.
The nearest neighbor classification performed for this example depends on how many
neighboring points are considered. If 1-NN is used and only the closest point is considered, then
clearly the new point should be classified as negative since it is on top of a known negative
point. On the other hand, if 9-NN classification is used and the closest 9 points are considered,
then the effect of the surrounding 8 positive points may overbalance the close negative point.
A probabilistic neural network builds on this foundation and generalizes it to consider all
of the other points. The distance is computed from the point being evaluated to each of the other
points, and a radial basis function (RBF) (also called a kernel function) is applied to the distance
to compute the weight (influence) for
34
each point. The radial basis function is so named because the radius distance is the argument to the
function.
The further some other point is from the new point, the less influence it has.
Fig 3.6.2b
Fig 3.7
35
ADVANTAGES AND DISADVANTAGES OF PNN NETWORKS
DTREG provides an option to cause it remove unnecessary neurons from the model after
the model has been constructed.
36
The process of removing unnecessary neurons is an iterative process. Leave- one-out
validation is used to measure the error of the model with each neuron removed. The neuron that
causes the least increase in error (or possibly the largest reduction in error) is then removed from
the model. The process is repeated with the remaining neurons until the stopping criterion is
reached.
When unnecessary neurons are removed, the “Model Size” section of the analysis report
shows how the error changes with different numbers of neurons. You can see a graphical chart of
this by clicking Chart/Model size.
There are three criteria that can be selected to guide the removal of neurons:
(1) Minimize error – If this option is selected, then DTREG removes neurons as
long as the leave-one-out error remains constant or decreases. It stops when
it finds a neuron whose removal would cause the error to increase above the
minimum found.
(2) Minimize neurons – If this option is selected, DTREG removes neurons until
the leave-one-out error would exceed the error for the model with all
neurons.
(3) # of neurons – If this option is selected, DTREG reduces the least significant
neurons until only the specified number of neurons remain.
37
PNN
NEWPNN creates a two layer network. The first layer has RADBAS neurons, and
calculates its weighted inputs with DIST, and its net input with NETPROD. The second layer has
COMPET neurons, and calculates its weighted input with DOTPROD and its net inputs with
NETSUM. Only the first layer has biases. NEWPNN sets the first layer weights to P', and the
first layer biases are all set to 0.8326/SPREAD resulting in radial basis functions that cross 0.5
at weighted inputs% of +/- SPREAD. The second layer weights W2 are set to T.
3.10.1 METHODOLOGY
A description of the derivation of the PNN classifier was given. PNNs had been used for
classification problems. The PNN classifier presented good accuracy, very small training time,
robustness to weight changes, and negligible retraining time. There are 6 stages involved in the
proposed model which are starting from the data input to output. The first stage is should be the
image processing system. Basically in image processing system, image acquisition and image
enhancement are the steps that have to do. In this paper, these two steps are skipped and all the
images are collected from available resource. The proposed model requires converting the image
into a format capable of being manipulated by the computer. The MR images are converted into
matrices form by using MATLAB. Then, the PNN is used to classify the MR images. Lastly,
performance based on the result will be analyzed at the end of the development phase.
38
CHAPTER 4
39
SPATIAL FUZZY CLUSTERING MODEL
Fuzzy clustering plays an important role in solving problems in the areas of pattern
recognition and fuzzy model identification. A variety of fuzzy clustering methods have been
proposed and most of them are based upon distance criteria. One widely used algorithm is the
fuzzy c-means (FCM) algorithm. It uses reciprocal distance to compute fuzzy weights. A more
efficient algorithm is the new FCFM. It computes the cluster center using Gaussian weights, uses
large initial prototypes, and adds processes of eliminating, clustering and merging. In the
following sections we discuss and compare the FCM algorithm and FCFM algorithm.
Spatial Fuzzy C Means method incorporates spatial information, and the membership
weighting of each cluster is altered after the cluster distribution in the neighborhood is
considered. The first pass is the same as that in standard FCM tocalculate the membership
function in the spectral domain. In the second pass, the membership information of each pixel is
mapped to the spatial domain and the spatial function is computed from that. The FCM iteration
proceeds with the new membership that is incorporated with the spatial function. The iteration is
stopped when the maximum difference between cluster centers or membership functions at two
successive iterations is less than a least threshold value.
The fuzzy c-means (FCM) algorithm was introduced by J. C. Bezdek [2]. The idea of
FCM is using the weights that minimize the total weighted mean- square error:
(k=1,K) (wqk) =1
40
2 1/(p-1) 2 1/(p-1)
wqk = (1/(Dqk) ) / (k=1,K) (1/(Dqk) ) , p>1
The FCM allows each feature vector to belong to every cluster with a fuzzy truth value
(between 0 and 1), which is computed using Equation (4). The algorithm assigns a feature vector
to a cluster according to the maximum weight of the feature vector over all clusters.
ALGORITHM FLOW
41
STANDARDIZE THE WEIGHTS OVER Q
During the FCM iteration, the computed cluster centers get closer and closer.
To avoid the rapid convergence and always grouping into one cluster, we use w[q,k] =
(w[q,k] – wmin)/( wmax – wmin)
before standardizing the weights over Q. Where wmax, wmin are maximum or minimum weights
over the weights of all feature vectors for the particular class prototype.
validity :
= {(1/K) 2
}/D 2
(k=1,K) k min
2 (q) (k) 2
k
= (q=1,Q) wqk || x – c ||
42
The variance of each cluster is calculated by summing over only the members of each cluster
rather than over all Q for each cluster, which contrasts with the original Xie-Beni validity
measure.
(q) (k) 2
2
= {q: q is in cluster k} wqk || x – c ||
k
The spatial function is included into membership function as given in Equation
MORPHOLOGICAL PROCESS
Morphological image processing is a collection of non-linear operations related to the
shape or morphology of features in an image. Morphological operations rely only on the relative
ordering of pixel values, not on their numerical values, and therefore are especially suited to the
processing of binary images. Morphological operations can also be applied to grey scale images
such that their light transfer functions are unknown and therefore their absolute pixel values are
of no or minor interest.
Morphological techniques probe an image with a small shape or template called a
structuring element. The structuring element is positioned at all possible locations in the image
and it is compared with the corresponding neighbourhood of pixels. Some operations test
whether the element "fits" within the neighbourhood, while others test whether it "hits" or
intersects the neighbourhood:
A morphological operation on a binary image creates a new binary image in which the pixel has
a non-zero value only if the test is successful at that location in the input image.
43
A morphological operation on a binary image creates a new binary image in which the
pixel has a non-zero value only if the test is successful at that locationin the input image.
The structuring element is a small binary image, i.e. a small matrix of pixels, each with
a value of zero or one:
The pattern of ones and zeros specifies the shape of the structuring element.
An origin of the structuring element is usually one of its pixels, although generally the origin can
be outside the structuring element.
A common practice is to have odd dimensions of the structuring matrix and the origin
defined as the centre of the matrix. Structuring elements play in morphological image processing
the same role as convolution kernels in linear image filtering.
When a structuring element is placed in a binary image, each of its pixels is associated
with the corresponding pixel of the neighbourhood under the structuring element. The structuring
element is said to fit the image if, for each of its pixels set to 1, the corresponding image pixel is
also 1. Similarly, a structuring element is
44
said to hit, or intersect, an image if, at least for one of its pixels set to 1 the corresponding
image pixel is also 1.
fig 4.5a Fitting and hitting of a binary image with structuring elements s1 and s2.
Zero-valued pixels of the structuring element are ignored, i.e. indicate points where the
corresponding image value is irrelevant.
FUNDAMENTAL OPERATIONS
More formal descriptions and examples of how basic morphological operations work are
given in the Hypermedia Image Processing Reference (HIPR) developed by Dr. R. Fisher et al. at
the Department of Artificial Intelligence in the University of Edinburgh, Scotland, UK.
Erosion with small (e.g. 2×2 - 5×5) square structuring elements shrinks an image by
stripping away a layer of pixels from both the inner and outer boundaries of regions. The holes
and gaps between different regions become larger, and small details are eliminated:
45
fig 4.7 Erosion: a 3×3 square structuring element
Larger structuring elements have a more pronounced effect, the result of erosion with a
large structuring element being similar to the result obtained by iterated erosion using a smaller
structuring element of the same shape. If s1 and s2 are a pair of structuring elements identical in
shape, with s2 twice the size of s1, thenf s2 ≈ (f s1) s1.
Erosion removes small-scale details from a binary image but simultaneously reduces the
size of regions of interest, too. By subtracting the eroded image from the original image,
boundaries of each region can be found: b = f − (f s ) where f is an image of the regions, s is a
3×3 structuring element, and b is an image of the region boundaries.
The holes enclosed by a single region and gaps between different regions become
smaller, and small intrusions into boundaries of a region are filled in:
46
fig 4.7a Dilation: a 3×3 square structuring element
Results of dilation or erosion are influenced both by the size and shape of a structuring
element. Dilation and erosion are dual operations in that they have opposite effects. Let f c denote
the complement of an image f, i.e., the image produced by replacing 1 with 0 and vice versa.
Formally, the duality is written as
f s = f c srot
47
CHAPTER 5
SIMULATED RESULTS
48
MRI Brain Image with Benign Case
49
IMAGE SEGMENTATION
50
IMAGE SEGMENTATION
51
SNAPSHOTS OF MODEL
52
TUMOUR DETECTION
53
TUMOUR AREA VIEW
54
CONCLUSION AND FUTURE SCOPE
The project presented that automated brain image classification for early stage
abnormality detection with use of neural network classifier and spotting of
tumor was done with image segmentation.
55
REFERENCES
[1] N. Kwak, and C. H. Choi, “Input Feature Selection for Classification Problems”,
IEEE Transactions on Neural Networks, 13(1), 143–159, 2002.
[2] E. D. Ubeyli and I. Guler, “Feature Extraction from Doppler Ultrasound Signals
for Automated Diagnostic Systems”, Computers in Biology and Medicine, 35(9),
735–764, 2005.
[3] D.F. Specht, “Probabilistic Neural Networks for Classification, mapping, or
associative memory”, Proceedings of IEEE International Conference on Neural
Networks, Vol.1, IEEE Press, New York, pp. 525-532, June 1988.
[4] D.F. Speech, “Probabilistic Neural Networks” Neural Networks, vol. 3, No.1, pp.
109-118, 1990.