Professional Documents
Culture Documents
Fractals Part2
Fractals Part2
Fractals Part2
Now we introduce ways to quantify fractals. We motivate our study of dimensions, we look carefully at some
peculiar aspects of measurement.
We shall see that measuring a shape in the wrong dimension gives results that at first
Ineffective Ways to Measure
seem peculiar.
A familiar method of measuring the length of a curve is to
Measuring in a dimension too low gives infinity: a filled-in square has infinite length. approximate the curve by straight line segments; add the lengths of
the line segments.
Smaller line segments should give a better approximation, and
we look for a limiting value of the lengths of the line segments, as we
Measuring in a dimension too high gives zero: a filled-in square has zero volume.
use smaller and smaller line segments.
So what are we to make of the calculation, that the Koch curve has infinite length
(1-dimensional measure) and zero area (2-dimensional measure)?
Does this mean the dimension of the Koch curve lies between 1 and 2? How is this
possible? What could it mean?
First, we apply this to a situation where it works: finding the length Heartened by this success, we try the same approach to compute the length of
(circumference) of a circle. the Koch curve.
First, we approximate the Koch curve by the straight line segment between its
We know how to measure the length of a straight line.
endpoints. Call the length of this line segment L0.
For a reasonably smooth curve we can try to measure the length of the curve by
For example, we might say L0 = 1 meter. Whatever it is, the length of the Koch
approximating it with straight line segments.
curve is greater than 1.
Using segments of length d, suppose N such segments are needed to
approximate the curve. Then the length of the curve, measured at scale d, is
L(d) = Nd
Now add three points between the original endpoints, obtaining four line
For example, let us approximate a circle of radius 1 by straight line segments. segments, each of length 1/3.
The collection of these segments has length L1 = 4/3.
We see the length of the Koch curve is greater than 4/3
Continue: between each pair of points, add three more points. This gives
a collection of sixteen segments, each of length 1/9
Taken together, the collection has length L2 = 16/9 = (4/3)2.
1
4/22/2013
n length of a segment number of segments Ln Not only does the Koch curve itself have infinite length, but measured along the
curve, the length of any pair of points in the Koch curve is infinite.
0 1 1 L0 = 1
1 1/3 4 L1 = 4/3
2 1/9 = 1/32 16 = 42 L2 = 16/9 = (4/3)2
To make these computations concrete, note that if the original L0 is 1 meter, then
3 1/27 = 1/33 64 = 43 L3 = 64/27 = (4/3)3
... ... ... ...
number of segments length of the curve
length of a segment
n 1/3n 4n Ln = (4/3)n
L24 3.5(10-12) m 281, 474, 976, 710, 656 1 km
Area of the Koch curve Replace each small triangle by four still smaller triangles. We see these triangles
have base and height 1/3 those of the previous triangles.
We'll try to compute the area of the Koch curve by covering it with isosceles
triangles (suggested by the general shape of the Koch curve). Thus these triangles have area 1/9 that of the previous triangles, and so (1/9)2 that
of the original triangle.
Using smaller and smaller triangles should give a better estimate of the area of the
Koch curve. The areas of these triangles sum to
First, cover the Koch cuve with a single triangle. A2 = ((3)/12)(16/81) = ((3)/12)(4/9)2.
This triangle has base length 1 and altitude sqrt(3)/6 (from the Pythagorean
theorem),
hence area
A0 = (3)/12.
Certainly, the area of the Koch curve is less than A0.
General Pattern We have seen that trying to measure the length of the Koch curve gives infinity,
number of while trying to measure the area of the Koch curve gives zero.
n area of a triangle An
triangles
Neither is a useful result. Here we shall introdce a more general measure that
0 (3)/12 1 A0 = (3)/12 leads to the idea of box-counting dimension.
1 ((3)/12)(1/9) 4 A1 = ((3)/12)(4/9)
Instead of approximating the Koch curve with line segments or triangles, we
A2 = ((3)/12)(16/81)
2 ((3)/12)(1/81) = ((3)/12)(1/9)2 16 = 42 could cover it with squares
= ((3)/12)(4/9)2
... ... ... ...
n ((3)/12)(1/9)n 4n An = ((3)/12)(4/9)n
2
4/22/2013
Covering a line segment with smaller and smaller boxes we see the pattern in Covering a filled-in square with smaller and smaller boxes we see the pattern
the table on the right shown in the table on the right.
N(1) = 1
N(1) = 1
N(1/2) = 2 = 1/(1/2)
N(1/2) = 4 = 1/(1/4) = (1/(1/2))2
N(1/4) = 4 = 1/(1/4)
N(1/4) = 16 = 1/(1/16) = (1/(1/4))2
N(1/8) = 8 = 1/(1/8)
N(1/8) = 64 = 1/(1/64) = (1/(1/8))2
and so in general
and so in general
N(r) = 1/r.
N(r) = (1/r)2.
For more complicated shapes, the relation between N(r) and 1/r may be
a power law, N(r) = k(1/r)d.
Definition of Box-Counting Dimension Solving for d and taking the limit as r 0 gives
db = lim r 0Log(N(r))/Log(1/r)
For different side lengths r we count N(r), the smallest number of boxes of side
length r needed to cover the shape. (Note as r 0 we have 1/r , so Log(1/r) and Log(k)/Log(1/r) 0.)
How does N(r) depend on r? If the limit exists, it is called the box-counting dimension, db, of the shape.
If the shape is 1-dimensional, such as the line segment, we have This limit may be slow to converge; an alternate approach
seen N(r) = 1/r. is to notice
If the shape is 2-dimensional, such as the (filled-in) unit square, we Log(N(r)) = dLog(1/r) + Log(k)
have seen N(r) = (1/r)2. is the equation of a straight line with slope d and y-
If the shape is 3-dimensional, such as the (filled-in) unit cube, we intercept Log(k).
expect N(r) = (1/r)3. So plotting Log(N(r)) vs Log(1/r), the points should lie
approximately on a straight line with slope db. This is
For more complicated shapes, the relation between N(r) the log-log approach to finding the box-counting
and 1/r may not be so clear. dimension.
If we suspect that N(r) is approximately k(1/r)d, a power
law relation, how can we find d? Taking Log of both sides
of N(r) = k(1/r)d, we obtain
Log(N(r)) = Log(k) + Log((1/r)d) = dLog(1/r) + Log(k)
with the expectation that the approximation becomes
better for smaller r.
Box-Counting Dimension of a Gasket Here is the Log-Log plot to estimate the box-counting dimension of the gasket.
In this case, the pattern is simple enough that we can find the exact value of the
Covering a Gasket with smaller and smaller boxes we see the pattern illustrated in dimension.
the table on the right
N(1) = 1
N(1/2) = 3
N(1/4) = N((1/2)2) = 9 = 32 Plotting the points in the
table, we obtain the graph. So the box-counting
N(1/8) = N((1/2)3) = 27 = 33 dimension of the
The points lie on a straight
and in general line of slope about 1.59 gasket is about 1.59.
N((1/2)n) = 3n.
db = limrn0Log(N(rn)) / Log(1/rn)
= limnLog(N(rn)) / Log(1/rn)
(Log(1/r0), Log(N(r0))) = (Log(1), Log(1)) = (0, 0) From the relation
(Log(1/r1), Log(N(r1))) = (Log(2), Log(3)) = (0.301, 0.477) N((1/2)n) = 3n we
can compute the = limnLog(N((1/2)n)) / Log(1/((1/2)n))
(Log(1/r2), Log(N(r2))) = (Log(4), Log(9)) = (0.602, 0.954) exact value of db for = limnLog(3n) / Log(2n)
(Log(1/r3), Log(N(r3))) = (Log(8), Log(27)) = (0.903, 1.432) the gasket.
= limn(nLog(3)) / (nLog(2))
(Log(1/r4), Log(N(r4))) = (Log(16), Log(81)) = (1.204, 1.908)
... ... ... = Log(3) / Log(2) 1.58996
3
4/22/2013
Box-Counting Dimension of the Koch Curve Here is the Log-Log plot to estimate the box-counting dimension of the Koch curve.
In this case, the pattern is simple enough that we can find the exact value of the
Covering the Koch curve with smaller and smaller boxes we see the pattern dimension.
illustrated in the table on the right.
we see they lie on a
straight line of slope
N(1/3) = 3 about 1.26
So the box-counting
N(1/9) = N((1/3)2) = 12 = 34 dimension of the
N(1/27) = N((1/3)3) = 48 = 342 Koch curve is about
1.26.
and in general
N((1/3)n) = 34n-1
db = limrn0Log(N(rn)) / Log(1/rn)
= limnLog(N(rn)) / Log(1/rn)
(Log(1/r2),Log(N(r2))) = (Log(9), Log(12)) = (0.954, 1.079) value of db for the = limn((n-1)Log(4) + Log(3)) / (nLog(3))
(Log(1/r3),Log(N(r3))) = (Log(27), Log(48)) = (1.431, 1.681) gasket. = limn(nLog(4) - Log(4) + Log(3)) / (nLog(3))
Box-Counting Dimension of the Cantor Set What happens when we measure an object in the wrong dimension?
The Cantor Middle Thirds set is generated by the IFS The reason for the useless results we got when we tried to measure the area and length of the Koch curve is this.
Trying to measure in a dimension lower than an object gives infinity (Imagine the length of infinitely thin thread needed to cover up a filled-
T1(x) = x/3 T2(x) = x/3 + 2/3 in square.) and
To compute the box-counting dimension of the Cantor set, we cover it with smaller trying to measure in a dimension higher than an object gives zero. (Think of the volume of a filled-in square. How much water can it hold?)
and smaller boxes, taking the box scaling based on the natural size structure of the For the Koch curve, the dimension lies between 1 and 2, so we should not be surprised that its length is infinite and its area zero.
fractal. That is, we use boxes of side length 1/3, 1/32, 1/33, ... . To emphasize this point, recall that using boxes of side length rn 0 as n ,
and in general Here are some graphs, using the Koch curve data, for different values of d.
N((1/3)n) = 2n.
= Log(2)/Log(3) 0.62989
N(1/3) = 6 = 3121
N(1/9) = N((1/3)2) = 36 and in
= 94 = 3222 general
N((1/3)n) = 3n2n.
4
4/22/2013
= limnLog(3n) / Log(2n)
= limn(nLog(3)) / (nLog(2))
The straightforward approach may appear to run into some trouble. = Log(3)/Log(2)
Specifically, how can we simplify
Log(3n + 2n)
Note this is the larger of the box-counting dimension of the
when the Log does not behave nicely with respect to sums? line segment (db = 1) and the Gasket (db = Log(3)/Log(2)).
The trick is to turn the sum into a product:
3n + 2n = 3n(1 + (2/3)n)
A Common Mistake
We could compute the slope by finding the
We want to show the side-elevation view of the building (in blue) is a fractal. best-fitting line, but some examples
(incorrectly) compute slopes for each pair
To do this, we cover the image with boxes of points.
of several sizes, count the boxes, and do a
log-log plot of the number of boxes.
If the points are close to a straight line, the slope of
the line is the box-counting dimension. If the first and second (1.724 - 1.230)/(.301 - 0) = 1.641
dimension is not a whole number, the shape must be second and third (2.262 - 1.724)/(.602 - .301) = 1.787
a fractal, because only fractals have non-integer
first and third (2.262 - 1.230)/(.602 - 0) = 1.714
dimensions.
These numbers are not close to 2, so the
shape must be a fractal.
WRONG WRONG WRONG As we saw in the example of the gasket and line
segment example, if a shape consists of several pieces, the dimension of the
shape is the largest of the dimensions of the pieces.
This shape contains filled-in rectangles, having dimension 2, so the whole shape
rn N(rn) 1/rn Log(1/rn) Log(N(rn)) has dimension 2.
1 17 1 0 1.230 What went wrong with the calculation? Not nearly small enough boxes were used.
1/2 53 2 .301 1.724 Box-counting ratios can be very slow to converge to the dimension.
1/4 183 4 .602 2.262
We cover a shape with boxes of side lengths r1, r2, r3, ..., with r1 > r2 > r3 > ... , with the hope that smaller boxes will
pick up finer details.
Count the numbers N(r1), N(r2), N(r3), ... of these boxes. We hypothesize a power-law scaling
N(ri) = k(1/ri)d
Taking log of both sides and simplifying, we obtain
log(N(ri)) = dlog(1/ri) + log(k)
the equation of a straight line, with log(N(ri)) on the y-axis and log(1/ri) on the x-axis.
If the power-law scaling is correct, the points will lie on a straight line, and the slope of that line id the box-counting
dimension db.
If we can find a formula for N(ri), then we can find db by taking the limit
db = limri 0log(N(ri))/log(1/ri)
5
4/22/2013
Similarity Dimension To emphasize this common pattern, and to give a short-cut for computing the
dimension of a self-similar shape, we define the similarity dimension.
For exactly self-similar fractals, with all pieces scaled by the same factor,
calculation of the box-counting dimension can be simplified greatly. For a self-similar shape made of N copies of itself, each scaled by a similarity with
Later we shall see that the Moran equation generalizes this computation to self- contraction factor r, the similarity dimension is
similar fractals where different pieces are scaled by different factors.
ds = Log(N)/Log(1/r).
For self-similar fractals such as the Koch curve and the Sierpinski gasket, the
This formula can be generalized to account for the posibility that different pieces
box-counting dimension is easiest to compute if the box sizes are taken to be
are scaled by similarities with different contraction factors. We shall see the more
powers of the scaling factor, 1/3 for the Koch curve and 1/2 for the Sierpinski
general formula is the Moran
Note equation.
this fractal and the Sierpinski
gasket.
gasket
The similarity dimension have the box-counting
equals same similaritydimension, but the box-counting
dimension is defineddimension.
for a widerNot surprisingly,
variety of shapes.dimension
For example, we observed the gasket can be covered with 3n boxes of side
alone need not distinguish one fractal
length 1/2n, so the box-counting dimension computation is Then why mention the similarity dimension at all?
from another.
It is much easier to compute than the box-counting dimension.
db = lim n Log(3n)/Log((2n))
= lim n (nLog(3))/(nLog(2))
= Log(3)/Log(2). The shape can be decomposed into
Note n cancels out of the computation. This is not an accident, but a consequence N = 3 pieces, each scaled by a factor
of the self-similar scaling. of r = 1/2, so ds = Log(3)/Log(2).
The similarity dimension equation can be applied only when the all the pieces are
scaled by the same amount.
Yet many self-similar fractals are made of pieces scaled by different amounts. Now Because different r values are involved, the similarity
we learn to compute the similarity dimension of these more general self-similar dimension formula cannot be applied. What can we do?
fractals.
6
4/22/2013
For some values of ri the Moran equation can be solved exactly, but often we this fractal can be decomposed into three
must solve it numerically. pieces with r1 = r2 = r3 = 1/2, and one
0.5d + 0.5d + 0.5d + 0.25d =1 with r4 = 1/4
Interchanging the exponents of the second term of the left side we obtain
This approach can be adopted to any situation in which all the scaling factors are (integer) powers of one of the scaling factors.
7
4/22/2013
Assuming the shape has finite extent, the power relation holds only for a range of
r values.
When r becomes too large, the entire object is contained inside the circle of
radius r and M(r) no longer changes as r increases.
When r becomes too small, we are in the realm of atoms, quarks, superstrings,
who knows what? There is no reason to exect the power law relation to hold on
scales smaller than the forces that sculpt the object.
Power Law Relations Pulse rate is related to metabolic rate per unit mass, so smaller animals should have faster
pulse rates and larger animals slower. Indeed, this is observed, familiar even. A mouse's heart
Relations of the form f(x) = kxh are called power law relations. beats very rapidly, a whale's heart very slowly. Add in the observation that most mammal
hearts beat 1 to 2 billion times during the animal's life and we understand that in the absence
Science is filled with power laws. For instance, of external perturbation (early death due to predation or disease, for example), a mouse has a
Hooke's law for springs: F(x) = -kx shorter life than a person, who in turn has a shorter life than a whale.
Newton's law of gravitation: F(r) = GMmr-2
This makes perfect sense, but careful measurements by Kleiber in 1932 revealed something
the allometry of animal metabolic rates: different: for most animals, over a range of sizes spanning 21 orders of magnitude, the
metabolic rate = k(weight)3/4. It turns out this is a bit of a surprise. metabolic rate per unit mass varies as M-1/4 rather than as M-1/3.
Allometric rates Adding to the interest of this question, plants exhibit this M-1/4 metabolic scaling law. Also, other
biological variables exhibit power law scalings with mass. Life-span scales as M1/4, age of first
Very roughly, the surface area, A, of an animal scales as its linear size, L, squared: reproduction as M3/4, the time of embryonic development as M-1/4, and the diameters of tree
trunks and of aortas as M3/8, for example.
A L2
by which we mean there is a constant k1 with A = k1L2 The reason for the observed M-1/4 scaling is not understood, but several explanations have
Similarly, the mass, M, of an animal scales as M = k2L3 been proposed. One is based on the observation that the diffusion does not occur across a
smooth 2-dimensional surface, but across the fractal boundary of the lungs. Because the
Heat dissipation occurs across the surface, so the total metabolic rate of an animal is dimension of the lungs is d > 2, the metabolic rate scales as Ld, so as Md/3. Then the metabolic
proportional to L2, hence to M2/3. rate per unit mass scales as Md/3 - 1. Because d > 2, d/3 - 1 > -1/3. Measurements of the
The metabolic rate per unit mass then is proportional to M-1/3, or so argued Rubner in dimension of the lungs suggest a metabolic rate/unit mass of about -1/4.
1883.
By themselves, power laws do not imply fractal structure. Crumple pieces of paper of
For example, the Stellpflug formula relating weight and radius of pumpkins is different sizes and measure
the diameters of the paper
weight = kradius2.78
balls.
yet no one would say a pumpkin is a fractal.
One experiment gave the
A pumpkin is a roughly spherical shell enclosing an empty cavity; a pumpkin results on the left.
certainly isn't made up of smaller pumpkins.
Plotting Log(mass) vs
Rather, this is the scaling relation between the thickness of the shell and the size of Log(diam), we see the
the pumpkin. points fall almost along a
straight line of slope 2.5.
Here are two examples where a power law does give a dimension.
8
4/22/2013
Plotting Log(mass) vs Log(diam), we see scaled by r = 1/2), of dimension 2. In this family of fractals, we keep N = 4, and step r from
the points fall almost along a straight line of 0.25 to 0.5 in increments of 0.025. As their dimension increases, the Koch curve relatives
slope 3.0.
come closer to filling in a solid region of the plane.
Although the proof is quite complicated, the intersection result extends to fractal shapes:
How is the dimension of the union of A and
if A and B are subsets of n-dimensional space, then
B related to the dimensions A and
B? Typically, the dimension of the union is
the maximum of the dimensions. dim(A B) = dim(A) + dim(B) - n
Example Suppose A is a gasket with dimension d(A) = log(3)/log(2) = 1.58496, and Compute the box-counting dimension of the product of a Sierpinski gasket and a
B is a Cantor set with dimension d(B) = log(2)/log(3) = 0.63093. unit line segment.
What cube side lengths should be used? What is the pattern relating N(r) and r?
We select cubes of side lengths 1/2, 1/4, 1/8, ... , 1/2n, ...
noting the scaling of the gasket.
d = lim nEach square
Log(N(1/2 n)) covering part
/ Log(1/(1/2n))of the gasket is the top face of
9
4/22/2013
Some examples of calculating dimensions of natural objects. For example, the graph below illustrates the number N of earthquakes per year of
magnitude exceeding m in southern California for the period 1932-1972. The circle
Can dimension express the evolutionary complexity of sea shells, the roughness of represents the expected rate of great earthquakes in southern California. The data
coastlines and mountains, the distribution of earthquakes? predict about one per century.
Earthquakes
The Gutenberg-Richter Law is an empirical relation
log(N) = -bm + a
between the magnitude m and the number N of earthquakes of magnitude
exceeding m.
Straightforward calculations convert this to a power law.
N = Br-2b
where the length of the fault break is r = A and A is the area of the fault break.
Then 2b = D, the dimension of seismic activity for the region from which the data
were collected.
number of boxes
side length s
N(s)
1/4 52
1/8 115
1/16 275
Fractal dimension can measure roughness. Many objects have varying degrees The pictures below were generated with these probabilities
of roughness, so a single dimension is not an adequate measure. p1 = 0.1, p2 = p3 = p4 = 0.3.
Multifractals are a way to quantify this variability. Successive pictures show increments of 25000 points. With enough patience, the
whole square will fill in, but some regions fill in more quickly than others.
As a first mathematical example, we see that by adjusting the probabilities,
we can make different parts of the fractal fill in at different rates.
For example: The IFS of an example with probabilities generates the unit
square.
However, the square fills up in a non-uniform way, revealing many fractals.
With equal probabilities, the Random Algorithm for the IFS with these rules fills
in the unit square uniformly. The probabilities of applying each transformation represent the fraction of the
total number of iterates in the region determined by the transformation.
T3(x, y) = (x/2, y/2) + (0, 1/2) T4(x, y) = (x/2, y/2) + (1/2, 1/2)
T1(x, y) = (x/2, y/2) T2(x, y) = (x/2, y/2) + (1/2, 0)
10
4/22/2013
Variable Probability Histograms Here we show the first four generations, with the height of the box in a region
In a typical picture about 0.1 of the points will lie in the square with address representing the fraction of the points in that region.
1, and about 0.3 of the points will lie in each of the squares with address 2, 3, All the pictures have been adjusted to have the same height, whereas square 4
and 4. has 0.3 of the points, square 44 has 0.09 of the points, square 444 has 0.027 of
the points, and so on.
Arguing in the same way, about 0.01 = 0.1*0.1 of the points will lie in the square with
address 11, about 0.03 = 0.1*0.3 of the points will lie in the square with address 12, and so
on.
The walls of the boxes in the pictures obscure the intermediate values.
The exponents (ijk) are coarse Holder exponents. They are easily computed from Here we show the same data (with two more iterates), but just the tops of the
the probabilities: boxes.
(i1 ... in) = Log(Prob(i1 ... in))/Log(2-n) So again the height represents the fraction of the points landing in that region.
For this example, the length 2 addresses have probabilities .01, .03, and .09. The
corresponding coarse Holder exponents are
Log(.01)/Log(.25) 3.32, Log(.03)/Log(.25) 2.53, and Log(.09)/Log(.25) 1.74
Note the smaller probability gives the larger Holder exponent.
Computing the
dimension of the second-
highest section requires
some work, because at
each iteration sections of
We see the limiting some squares are
The highest regions converge to a gasket, not a surprise as we've seen. shape is filled withremoved,
a and sections of
cascade of smaller previously
and empty
What about the second highest probability regions? smaller gaskets. squares are filled.
11
4/22/2013
Example B, with p1 = 0.2, p2 = p3 = 0.25, and p4 = 0.3. For each region we expect that
Now structures more complicated than gaskets will prob scales as (side length)some power
appear. So instead of letting the height of the graph represent the probability of the region,
now we assign height Log(prob)/Log(side length) to the region.
Among other things, the number of values
of the probabilities of regions increases Because the probability measures the fraction of the points that occupy a region,
more rapidly we think of this ratio as a dimension.
Being viewed at the resolution of the side length of the region, this is a coarse
Holder exponent; it is also called the coarse dimension.
Dimensions of strata Here are approximations of E ( = height) for the example with p1 = 0.1 and p2 = p3 = p4 =
0.3. We picture the coarse Holder exponents for length 5 address squares. With these
Now we stratify the square into regions having the same Holder exponent .
probabilities, there are six values of .
Computing the dimensions of these strata is how multifractals are quantified.
A plot of dimension as a function of is the f() curve.
Taking limits as the side length of the regions go to zero, the coarse Holder
exponent can be refined to the local Holder exponent (or roughness) at any
point (x, y) in address square i1...in is
dloc(x,y) = lim n Log(Prob(i1...in))/Log(2-n)
where Prob(i1...in) is the probability pr(i1)...pr(in).
The value for a square of finite length address is called the coarse Holder 1.737 2.054 2.371
exponent. So the local Holder exponent of a point (x, y) is the limit as n
of the coarse Holder exponents of the length n address squares
containing (x, y).
Now define
E = {(x, y): dloc(x, y) = },
the collection of all points of the fractal having local Holder exponent .
As takes on all values of the local Holder exponent, we decompose the 2.688 3.005 3.322
fractal into these strata E.
12
4/22/2013
When was fractals and chaos first researched? When was fractals and chaos first researched?
When was chaos theory relating to biology first When was chaos theory relating to biology first
researched? researched?
The father of fractals is often considered to be a man by the name of Chaos theory relating to biology was first researched in the early 70s.
Gaston Julia. In the early 1900s, Julia did much research on iterated Researchers were looking at how chaos theory could be used to model
functions, and even drew some of his famous Julia sets by hand. True, population trends. Several researchers, such as George Oster, Robert
there were some other works out there, such as Sierpinskis triangle and May, and Jim Yorke, looked at equations such as this one in their effort
Kochs curve, but Julias work was a major breakthrough. Until the to model population: xt+l = l xt(1 xt). As for human biology, shortly after
1960s much of the work with fractals was abandoned due to lack of the Mandelbrot Set was discovered this also took off. Dynamical
technology. That changed in the 1970s when Mandelbrot used diseases, a term coined in this era, described diseases that show
computers to create what we now know today as the Mandelbrot Set. chaotic systems. Researchers such as Leon Glass and Michael Mackey
did research in this field. Now, there are organizations dedicated to
research with chaos theory and fractals in the field of biology.
Gaston Julia,
who sadly, lost
Pictures from:
http://www.fractalus.ch/index_pic/J his nose in Leon Glass
ulia.jpg
The Mandelbrot Set WWI
http://www.spsu.edu/math/edward
s/mandel/bigman.jpg Picture from:
www.physionet.org/ Onward
contributors/
13
4/22/2013
As you can see the top graph shows a As you can see the heart rate is very complex
normal healthy heart. The graph of the and does show signs of self-similarity. What
healthy heart has more complexity then the is more surprising is how the heart-rate seems
bottom graph. Complexity = healthy in to lose its long range correlation as the heart
many physiological aspects. The bottom becomes diseased or break down.
time scale graph shows a heart with CHF
(congestive heart failure). CHF is just one (Click to continue)
of the many diseases that causes the heart
rate to lose its chaotic property. Click on
the image for a further view of the normal
heart-rate.
Picture from:http://www.physionet.org/tutorials/ndc/
http://www.physionet.org/tutorials/ndc/
Picture from:http://www.physionet.org/tutorials/ndc
http://www.physionet.org/tutorials/ndc
14
4/22/2013
-These again show bifurcation along with other fractal - The fractal geometry of the heart www.medtronic.com/cardsurgery/images/fix4.jpg
15
4/22/2013
Fractals and Chaos in the Brain, Fractals and Chaos in the Brain,
DNA DNA
By modeling electrical signals in the brain and nerves, scientists have been able to
model the *alpha rhythm using chaos. Alpha rhythm, along with other wave
frequencies, can often be modeled using time series graphs. Beta: above 12
Hz
Images
from:http://www.crossroadsinstitute.org/eeg.html
If alpha rhythm is
between 8 -12 Hz and Delta: .1 3 Hz
has the graph shown
try to guess the
following.
Match the graph to the
type of rhythm (frequency
Alpha rhythm Does this look band). Theta: 4 8 Hz
familiar to any time plots seen in Delta: .1 3 Hz
the study of fractals and chaos?
Theta: 4 8 Hz
Images
Beta: above 12 Hz from:http://www.crossroadsinstitute.org/eeg.html
*Alpha rhythm- frequency of brainwaves between 8 -12 Hz.
Click to see answer
16
4/22/2013
Fractal Landscapes
FRACTAL IMAGE COMPRESSION
Fractals are now used in many forms to create textured landscapes and
other intricate models. It is possible to create all sorts of realistic fractal What is Fractal Image Compression ?
forgeries, images of natural scenes, such as lunar landscapes, mountain
ranges and coastlines. This is seen in many special effects within
Hollywood movies and also in television advertisements.
How can we describe behavior of the machine ? Transformations of the Iterated Function Systems ( IFS )
form as follows will help us.
An iterated function system consists of a collection of contractive
x a bi x ei affine transformations.
wi = i +
d i y fi
n
y ci W (*) = U wi (*)
i =1
Such transformations are called affine transformations. Affine For an input set S, we can compute wi for each i, take the union of these
transformations are able to skew, stretch, rotate, scale and translate an sets, and get a new set W(S).
input image.
Hutchinson proved that in IFS, if the wi are contractive, then W is
M.Barnsley suggested that perhaps storing images as collections of contractive, thus the map W will have a unique fixed point in the space of all
transformations could lead to image compression. images. That means, whatever image we start with, we can repeatedly
apply W to it and our initial image will converge to a fixed image. Thus W
completely determine a unique image.
| W | f = lim W ( n ) ( f o )
x
17
4/22/2013
where f and g are value of the level of grey of pixel, P is the space x ai bi 0 x ei
of the image wi y = ci di 0 y + fi
z 0 0 si z oi
Suppose we have an image f that we want to encode. On the other words, The decoding step is very simple. We start with any image and
we want to find a IFS W which f to be the fixed point of the map W apply the stored affine transformations repeatedly till the image
We seek a partition of f into N non-overlapped pieces of the image to which no longer changes or changes very little. This is the decoded
we apply the transforms wi and get back f. image.
We should find pieces Di and maps wi, so that when we apply a wi to the
part of the image over Di , we should get something that is very close to the
any other part of the image over Ri .
Finding the pieces Ri and corresponding Di by minimizing distances
between them is the goal of the problem.
First three iterations of the decompression of the image of an
eye from an initial solid grey image
Several years ago, one of the London tabloids printed this picture
Resolution Independence
18
4/22/2013
Discovered around 1980, the Mandelbrot set may well be the most familiar
image produced by the mathematics of the last century. Its status as a cultural First, we reformulate the process without using complex numbers
icon needs no support. Think of the complex number z as a pair (x, y) of real numbers, and think of the
complex number c as a pair of real numbers (a, b).
From a philosophical perspective, the Mandelbrot set challenges familiar In these terms, z z2 + c becomes
notions of simplicity and complexity: how could such a simple formula, x x2 - y2 + a
involving only multiplication and addition, produce a shape of great organic and
beauty and infinite subtle variation?
y 2xy + b
First we define the filled-in Julia set, Kc, for each complex number c.
Julia Sets. For a complex number c, the filled-in Julia set of c is the set of all z for
which the iteration z z2 + c does not diverge to infinity. The Julia set is the boundary of For each point z0 of the plane, generate a sequence z1, z2, z3, ... by the basic
the filled-in Julia set. For almost all c, these sets are fractals. iteration rule
zn+1 = zn2 + c
If the sequence does not run away to infinity, then the point z0 belongs to Kc;
if the sequence does run away to infinity, then z0 does not belong to Kc.
The Mandelbrot set. The Mandelbrot set is the set of all c for which
the iteration z z2 + c, starting from z = 0, does not diverge to infinity. Julia
sets are either connected (one piece) or a dust of infinitely many points. The
Mandelbrot set is those c for which the Julia set is connected.
First we note three computational aspects
of the definition of Kc:
19
4/22/2013
Unfortunately not: maybe z200 will be farther than 2 from the origin.
If all of z0, ..., zN lie within a distance of 2 from the origin, we assert the sequence
will never run away to infinity and so z0 belongs to Kc.
This leaves open the possibility that we incorrectly conclude some points belong
to Kc.
Generally, the larger N, the fewer mistakes
we make.
If the iterates of z0 fail to satisfy the escape criterion, we assume z0 belongs to the The Mandelbrot set, M, is defined by the same iteration process used to define
filled-in Julia set Kc and paint black the pixel containing z0. Julia sets, but applied in a different way.
If the iterates of z0 do satisfy the escape criterion, we usually paint the pixel Instead of using the complex plane to represent the different choices of
containing z0 with a color indicating how many iterations were needed to first get z0 (the Dynamical Plane), it represents different values of c (the Parameter
farther than 2 from the origin. Plane).
To compute an image of the Mandelbrot set, for each c start with z0 = 0 and
For example, generate the sequence z1, z2, z3, ... by the basic iteration rule
paint the pixel containing z0 white if z1 is farther than 2 from the origin. zn zn+1 = zn2 + c
Paint the pixel light to dark pink if a few more iterates are needed to get farther than If the sequence does not run away to infinity, then the point c belongs to M; if the
2 from the origin. sequence does run away to infinity, then c does not belong to M.
Paint the pixel orange to yellow if a few more iterates are needed to get farther than
2 from the origin. As with Julia sets, we paint the pixel black if the sequence produced by the c
at its center does not run away to infinity, that is, if this c belongs to M, and
Paint the pixel green to blue if a few more iterates are needed to get farther than 2
otherwise we paint the pixel a color determined by how quickly the sequence
from the origin.
gets farther than 2 from the origin (and hence runs away to infinity).
and so on.
For reference, here is a picture of the Mandelbrot set. The pink circle is the circle of The Mandelbrot Set - Maximum Iterations
radius 2 centered at the origin.
Recall the maximum number of iterations used in drawing Julia sets.
Here is an illustration of the effect of the maximum number of iterations on
rendering the Mandelbrot set. In the animation we step through from max =1 to
max = 25
20
4/22/2013
Magnifications, such as those below, certainly suggest that the small copies of the
Mandelbrot set are islands, well-separated from the main body of the set and from one
another.
That is, the intricately branched decorations are filled with strings of tiny Mandelbrot sets,
invisible at this level of magnification, but all linked back to the main body of the set.
This is another lesson on the need to be careful when interpreting computer pictures.
One of the early surprises of the Mandelbrot set is that its periphery is filled with
a halo of tiny copies of the entire set, each of which is surrounded by its own
halo of still tinier copies, and so on, on smaller and smaller scales, without end.
Despite appearances, these small copies are attached to the main body of the
set, through a sequence of still smaller copies. That is, the Mandelbrot set
is connected.
21