Professional Documents
Culture Documents
Hardware Book
Hardware Book
1
SYLLABUS
Unit – 1
Unit – 2
Unit – 3
Unit – 4
Unit – 5
Unit – 6
2
Unit – 1
BASIC OF ELECTRONIC :
Electronic components are Resistors, Capacitor, Diode, PCB board, Transistors, IC’s and
socket’s.
Resistor’s : A resistors is a two-terminal electronic component that produce a voltage across
its terminals that is proportional to the electric current passing through it in accordance with
ohm’s law V=IR. Refer fig. 1 Resistor diagram
Fig 1 Resistor
Capacitor: A capacitor is a passive electronic components consisting of a pair of conductors
separated by a dielectric. The capacitors charge a temporary current using circuit. Refer fig
1.1
3
Transistor: A transistor is a semiconductor device used to amplify and switch electronic
signals. It is made of solid piece of semiconductor material, which at least three terminals for
connection to external circuit. A Voltage or current applied to one pair of the transistor’s
terminals change the current flowing through another pair of terminals. Because the
controlled power can be much more than the controlling i/p power, the transistor provides of
amplification a signal. Some transistors are packaged individually but many more are found
embedded in integrated circuits. Refer fig 1.3
Fig 1.3
Introduction to computer
Fig. 1.4
Computers were initially large machines that could fill entire room. Some were operated
using large vacuum tubes that formed the basis of today’s transistors. In order to operator
such machine, punch card were used.1833 Charles Babbage invented his difference engine an
early calculator. Together with the punch card design, he created the analytical engine.
Regrettably the engine never saw completion due to political issues.
Here are some computers that came and went in the history of computing. Some modern
examples are also here.
ENIAC: ENIAC stood for Electrical Numerical Integrator and Computer. The ENIAC used
thousands of vacuum tubes and a punch card mechanism.
4
Fig 1.5
Commodore 64: An 8 bit computer introduced in January 1982, the commodore rose to
become the best selling personal computer of all times. Utilising the commodore BASIC
programming language licensed from Microsoft, the commodore was able to host over
10,000 commercial programs.
Macintosh: First introduced by apple 1984, the Macintosh was the first computer to use a
mouse and graphical user interface (gui) rather than a command line interface. Until the
dominance of the IBM Pc, the Macintosh saw use primarily as a desktop publishing tool.
Types of computer
1. Super computer
2. Main frame
3. Work station
4. The personal computer or PC
5. Micro controller
6. Server
Supercomputer:
Super computer are fast because they are really many computers working together. Super
computer were introduced in the 1960’s as the world’s most advanced computer. As of
November 2008, the fastest super computer is the IBM roadrunner. It has theoretical
processing peak of 1.71 petaflops and has currently peaked at 1.456 petaflops.
Mainframe: They are computer where all the processing is done centrally and the user
terminals are called “dumb terminals” since they only input and outputs.
Mainframes are computers used mainly by large organizations for critical applications,
typically bulk data processing such as census. Examples: banks, airlines, insurances
companies, and colleges.
5
Workstation: Workstation are high-end expensive computer that are made for complex
producers and are intended for one user at a time. Some of the complex producers consist of
science, math and engineering calculations and are useful for computer design and
manufacturing. Workstations are some times improperly named for marketing reasons. Real
workstations are not usually sold in retail.
Microcontroller: They are mini computers that enable the user to store data, do simple
commands and tasks, with title or no user interaction with the processor. These single circuit
devices have minimal memory and program length but can be integrated with other processor
for more complex functionality.
Generation of computer’s
Processor Introduction
If example Pentium IV processor contains a million type of transistor in that processor &
designed. First generation processor contain using 80XXX series, Second generation
802XXX series used, second generation processor’s using a expansion slots type processors.
Refer slide
6
UNIT – 2
INTRODUCTION TO MOTHER BOARD
A mother board made up of PCB board. If mother board have double circuits Front of
component and back side Circuits can link. Mother board consists of
Also referred to system board and main board. The motherboard is the foundation of the
pc system. Without the motherboard there would be no computer. Now there are some major
parts of the motherboard that you need to understand and get very familiar. As always I will
try to make this is as simply to understand as possible.
So you know what the main idea is of the motherboard now you need to know what its
functions are and how it works. Take a look at functions.
MOTHER FUNCTIONS
The mother takes care of the entire system task in one way or another. It is a go-between
of the system. You will find that almost all component are attached to motherboard in one
way or another way. Without the motherboard these system components would be hard
pressed to work.
Don’t go cheap on this item and get a good one. When deciding on a case and processors
we need to be concerned with designs.
Motherboard formats
When it comes to format we all need to pay attention not only to motherboard format but
also case format. The formats used today are mainly the ATX, and the AT. These are the only
styles you should worry about until things change again and there is a better alternative.
AT- this is a design that is fathered after IBM and very common. Though the style is old
and not really recommended today. There are many motherboard manufacturers that still
make AT boards for those trying to save a buck.
7
Baby AT- The baby AT is a smaller than the AT and is found in many cloned IBM
machines today. Like the AT it is an old style and is slowly fading out to the ATX.
ATX- this is becoming the standard and a nice motherboard. The board is modeled after the
baby AT design except the board is turned 90 degrees in the case allowing much room for big
expansion cards.
-For those of you that go ahead and throw in a few more dollars and get a ATX motherboard
and the case you will be very pleased. The transition is an easy one from the old school style
of the AT format to the new ATX.
-So what it the Core parts of the motherboard and what do they do for you? Check out some
of the components and learn more.
INTRODUCTION TO ATX
8
PROCESSOR & SOCKET’S
1. CPU (central processing unit) an electronic circuit that can execute computer programs
2. Microprocessor, a CPU on one chip as part of a microcomputer
3. Graphics processing unit (GPU / VPU), a dedicated graphics rendering device for a
personal computer or game console
4. Physics processing unit (PPU), a dedicated microprocessor designed to handle the
calculations of physics
5. Digital signal processor, a specialized microprocessor designed specifically for digital
signal processing
6. Network processor, a microprocessor specifically targeted at the networking application
domain
7. Front end processor, a helper processor for communication between a host computer and
other devices
8. Coprocessor
9. Floating point unit
10. Data processor, a system that translates or converts between different data formats
11. Word processor, a computer application used for the production of printable material
12. Audio processor, used in studios and radio stations
8080
Introduced April 1, 1974
Clock rate 2 MHz
0.64 MIPS
Bus Width 8 bits data, 16 bits address
Enhancement load NMOS logic
Number of Transistors 6,000
Assembly language downwards compatible with 8008.
Addressable memory 64 KB
Up to 10X the performance of the 8008
Used in the Altair 8800, Traffic light controller, cruise missile
Required six support chips versus 20 for the 8008
8085
Introduced March 1976
Clock rate 3 MHz
0.37 MIPS
Bus Width 8 bits data, 16 bits address
Depletion load NMOS logic
Number of Transistors 6,500 at 3 µm
Binary compatible downwards with the 8080.
Used in Toledo scales. Also was used as a computer peripheral controller – modems,
harddisks,printers, etc...
CMOS 80C85 in Mars Sojourner, Radio Shack Model 100 portable.
High level of integration, operating for the first time on a single 5 volt power supply, from 12
volts previously. Also featured serial I/O,3 maskable interrupts,1 Non-maskable interrupt,1
externally expandable interrupt w/[8259],status,DMA
9
Pentium (chronological entry)
Original Pentium
10
Number of transistors 3.2 million
Variants
120 MHz Introduced March 27, 1995
P54CS – 0.35 µm process technology
Number of transistors 3.3 million
90 mm² die size
Family 5 model 2
Variants
Socket 5 296/320 pin PGA package
133 MHz Introduced June 12, 1995
150, 166 MHz Introduced January 4, 1996
Socket 7 296/321 pin PGA package
200 MHz Introduced June 10, 1996
[edit] Pentium with MMX Technology
P55C – 0.35 µm process technology
Introduced January 8, 1997
Pentium Pro
11
0.35 µm process technology, or 0.35 µm CPU with 0.6 µm L2 cache
Number of transistors 5.5 million
512 KB or 256 KB integrated L2 cache
60 or 66 MHz system bus clock rate
Variants
166 MHz (66 MHz bus clock rate, 512 KB 0.35 µm cache) Introduced November 1, 1995
180 MHz (60 MHz bus clock rate, 256 KB 0.6 µm cache) Introduced November 1, 1995
200 MHz (66 MHz bus clock rate, 256 KB 0.6 µm cache) Introduced November 1, 1995
200 MHz (66 MHz bus clock rate, 512 KB 0.35 µm cache) Introduced November 1, 1995
200 MHz (66 MHz bus clock rate, 1 MB 0.35 µm cache) Introduced August 18, 1997
Pentium II
12
32 KB L1 cache
No L2 cache
Variants
266 MHz Introduced April 15, 1998
300 MHz Introduced June 9, 1998
Mendocino – 0.25 µm process technology
Introduced August 24, 1998
242-pin Slot 1 SEPP (Single Edge Processor Package), Socket 370 PPGA package
Number of transistors 19 million
66 MHz system bus clock rate
Slot 1, Socket 370
32 KB L1 cache
128 KB integrated cache
Family 6 model 6
Variants
300, 333 MHz Introduced August 24, 1998
366, 400 MHz Introduced January 4, 1999
433 MHz Introduced March 22, 1999
466 MHz
500 MHz Introduced August 2, 1999
533 MHz Introduced January 4, 2000
266 MHz (Mobile)
300 MHz (Mobile)
333 MHz (Mobile) Introduced April 5, 1999
366 MHz (Mobile)
400 MHz (Mobile)
433 MHz (Mobile)
450 MHz (Mobile) Introduced February 14, 2000
466 MHz (Mobile)
500 MHz (Mobile) Introduced February 14, 2000
13
Introduced October 25, 1999
Number of transistors 28.1 million
256 KB Advanced Transfer L2 Cache (Integrated)
242-pin Slot-1 SECC2 (Single Edge Contact cartridge 2) processor package, 370-pin FC-
PGA (Flip-chip pin grid array) package
System Bus clock rate 100 MHz (E-models), 133 MHz (EB models)
Slot 1, Socket 370
Family 6 model 8
Variants
500 MHz (100 MHz bus clock rate)
533 MHz
550 MHz (100 MHz bus clock rate)
600 MHz
600 MHz (100 MHz bus clock rate)
650 MHz (100 MHz bus clock rate) Introduced October 25, 1999
667 MHz Introduced October 25, 1999
700 MHz (100 MHz bus clock rate) Introduced October 25, 1999
733 MHz Introduced October 25, 1999
750, 800 MHz (100 MHz bus clock rate) Introduced December 20, 1999
850 MHz (100 MHz bus clock rate) Introduced March 20, 2000
866 MHz Introduced March 20, 2000
933 MHz Introduced May 24, 2000
1000 MHz Introduced March 8, 2000 (Not widely available at time of release)
1100 MHz
1133 MHz (first version recalled, later re-released)
400, 450, 500 MHz (Mobile) Introduced October 25, 1999
600, 650 MHz (Mobile) Introduced January 18, 2000
700 MHz (Mobile) Introduced April 24, 2000
750 MHz (Mobile) Introduced June 19, 2000
800, 850 MHz (Mobile) Introduced September 25, 2000
900, 1000 MHz (Mobile) Introduced March 19, 2001
Intel Core
14
Intel Core Duo L2500 1.83 GHz (Low voltage, 15W TDP)
Intel Core Duo L2400 1.66 GHz (Low voltage, 15W TDP)
Intel Core Duo L2300 1.5 GHz (Low voltage, 15W TDP)
Intel Core Duo U2500 1.2 GHz (Ultra low voltage, 9W TDP)
Intel Core Solo T1350 1.86 GHz (533 FSB)
Intel Core Solo T1300 1.66 GHz
Intel Core Solo T1200 1.5 GHz [35]
Dual-Core Xeon LV
Pentium 4
15
0.13 µm process technology Northwood A (1.7, 1.8, 1.9, 2, 2.2, 2.4, 2.5, 2.6,
2.8(OEM),3.0(OEM) GHz)
Improved branch prediction and other microcodes tweaks
512 KB integrated L2 cache
Number of transistors 55 million
400 MHz system bus.
Family 15 model 2
0.13 µm process technology Northwood B (2.26, 2.4, 2.53, 2.66, 2.8, 3.06 GHz)
533 MHz system bus. (3.06 includes Intel's hyper threading technology).
0.13 µm process technology Northwood C (2.4, 2.6, 2.8, 3.0, 3.2, 3.4 GHz)
800 MHz system bus (all versions include Hyper Threading)
6500 to 10000 MIPS
Dual-core microprocessor
Enabled Hyper-Threading
800(4x200) MHz front side bus
Smithfield – 90 nm process technology (3.2 GHz)
Variants
Pentium 840 EE – 3.20 GHz (2 x 1 MB L2)
Presler – 65 nm process technology (3.46, 3.73)
2 MB x 2 (non-shared, 4 MB total) L2 cache
Variants
Pentium 955 EE – 3.46 GHz, 1066 MHz front side bus
Pentium 965 EE – 3.73 GHz, 1066 MHz front side bus
16
INTRODUCTION TO BIOS
BIOS
The Basic Input-Output System (BIOS) is an essential set of routines in a PC, which is
stored on a chip on the motherboard. It acts as an intermediary between a computer's
hardware and its operating system. Without the BIOS, the PC's operating system would have
no way to communicate with, or take control of, the hardware.
In other words, the BIOS is a crucial component of any computer. If its options are set
incorrectly, the BIOS could slow your computer down by as much as 40%. Unfortunately, as
new processors and motherboard chipsets are released, BIOS options continue to get even
more confusing. As a result, many seasoned technicians are still baffled by the jargon-laced
and confusing options available in a modern computer's BIOS setup program.
Many large PC manufacturers such as Dell, HP, Gateway and Micron limit the options
available to the end-user in the BIOS, in order to reduce ill-advised "tinkering" and the
resulting support calls. As a result, you may not be able to take advantage of some of the
advanced settings mentioned here on PCs from these major vendors.
It is recommended that you reboot after each individual BIOS setting change to ensure
that your system functions normally. If you make numerous changes before rebooting, and
your system will no longer boot, you won't know which change is responsible for the failure.
When the computer starts, it starts the operating system that takes the control of the
machine. An Operating System is a set of programs that help in controlling and managing the
Hardware and the Software resources of a computer system. A good operating system should
have the following features:
1. Help in the loading of programs and data from external sources into the internal memory
before they are executed.
17
As the name suggests, the operating System is used for operating the system or the
computer. It is a set of computer programs and also known as DOS (Disk Operating System).
The main functions of DOS are to manage disk files, allocate system resources according to
the requirement. DOS provides features essential to control hardware devices such as
Keyboard, Screen, Disk Devices, Printers, Modems and programs.
Basically, DOS is the medium through which the user and external devices attached to the
system communicate with the system. DOS translate the command issued by the user in the
format that is understandable by the computer and instruct computer to work accordingly. It
also translates the result and any error message in the format for the user to understand.
LOADING DOS
The BOOT Record into the computer memory loads DOS. BOOT Record in turn is
triggered by ROM program already there in the computer.
The system start-up routine of ROM runs a reliability test called Power On Self Test (POST)
which initializes the chips and the standard equipment attached to the PC, and check
whether peripherals connected to the computer are working or not. Then it tests the RAM
memory. Once this process is over, the ROM bootstrap loader attempts to read the Boot
record and if successful, passes the control on to it. The instructions/programs in the boot
record then load the rest of the program. After the ROM boot strap loader turns the
control over to boot record, the boot tries to load the DOS into the memory by reading the
two hidden files IBMBIO.COM and IBMDOS.COM. If these two are found, they are
loaded along with the DOS command interpreter COMMAND.COM. COMMAND.COM
contains routines that interpret what is typed in through the keyboard in the DOS command
mode. By comparing the input with the list of command, it acts by executing the required
routines/commands or by searching for the required routine utility and loads it into the
memory.
BOOTING
In computing, booting (also known as "booting up") is a bootstrapping process that starts
operating systems when the user turns on a computer system. A boot sequence is the initial
set of operations that the computer performs when power is switched on.
A computer's central processor can only execute program code found in Read-Only
Memory (ROM), Random Access Memory (RAM) or an operator's console. Modern
operating systems and application program code and data are stored on nonvolatile data
storage devices, such as hard disk drives, CD, DVD, flash memory cards (like an SD card),
USB flash drive, and floppy disk. When a computer is first powered on, it does not have an
operating system in ROM or RAM. The computer must initially execute a small program
stored in ROM along with the bare minimum of data needed to access the nonvolatile devices
from which the operating system programs and data are loaded into RAM.
18
The small program that starts this sequence of loading into RAM, is known as a bootstrap
loader, bootstrap or boot loader. This small boot loader program's only job is to load other
data and programs which are then executed from RAM. Often, multiple-stage boot loaders
are used, during which several programs of increasing complexity sequentially load one after
the other in a process of chain loading.
BOOT LOADERS
BIOS
OpenBIOS
EFI
OpenBoot
SLOF
CD To Go Inner Directory
CD.. To go outer directory
DIR List the directory
RD To remove directory
COPY copy any file in commands
MOVE move any file in commands
DIR/AH diplay hidden files
DIR/AS display system files
DIR/AR display read-only file
DEL delete files
RENAME (REN) rename files
FORMAT format drives
ATTRIB To change file permission
CLS to clear screen
DATE to display date
TIME To display time
EXIT exit from dos
LABLE to display volume lable
MKDIR to make new folder or directory
MD ””
PROMPT $$ linux
VER To display msdos version
XCOPY copy entire directory
19
INTRODUCTION TO FLOPPY
A floppy disk is a data storage medium that is composed of a disk of thin, flexible
("floppy") magnetic storage medium encased in a square or rectangular plastic shell.
Floppy disks are read and written by a floppy disk drive or FDD, [2] Invented by the
American information technology company IBM, floppy disks in 8 inch, 5¼ inch and 3½
inch forms enjoyed nearly three decades as a popular and ubiquitous form of data storage and
exchange, from the mid-1970s to the late 1990s. While floppy disk drives still have some
limited uses, especially with legacy industrial computer equipment, they have now been
superseded by USB flash drives, external hard disk drives, CDs, DVDs, Blu Ray discs,
memory cards and computer networks.
A small motor in the drive rotates the diskette at a regulated speed, while a second motor-
operated mechanism moves the magnetic read/write head,(or heads, if a double-sided drive)
along the surface of the disk. To write data onto the disk, current is sent through a coil in the
head. The magnetic field of the coil magnetizes spots on the disk as it rotates; the change in
magnetization encodes the digital data. To read data, the tiny voltages induced in the head
coil by the magnetization on the disk are detected, amplified by the disk drive electronics,
and sent to the Floppy disk controller. The controller separates the data from the stream of
pulses coming from the drive, decodes the data, tests for errors, and sends the data on to the
host computer system.
A blank diskette has a uniform featureless coating of magnetic oxide on it. A pattern of
magnetized tracks, each broken up into sectors, is initially written to the diskette so that the
diskette controller can find data on the disk. The tracks are concentric rings around the
diskette, with spaces between the tracks where no data is written. Other gaps, where no user
data is written, are provided between the sectors and at the end of the track to allow for slight
speed variations in the disk drive.
USAGES:
The flexible magnetic disk, commonly called floppy disk, revolutionized computer disk
storage for small systems and became ubiquitous in the 1980s and 1990s in their use with
personal computers and home computers to distribute software, transfer data, and create
backups. In general, different physical sizes of floppy disks are incompatible by definition,
and disks can be loaded only on the correct size of drive. There were some drives available
with both 3½-inch and 5¼-inch slots that were popular in the transition period between the
sizes.
20
HARD DISK
A hard-disk drive is non-volatile device used for storage that is located inside the
computer case. Like the floppy drive, it holds its data on rotating platters with a magnetic
upper exterior which are changed or read by electromagnetic tipped arms that move over the
disk as it spins.
hard disks come in various speeds. An IDE hard disk spins at 4200rpm and is significantly
slower than SATA A particular SATA hard disk has a spindle speed of 7200rpm. A SCSI
hard disk (used in servers and high end computers)have a spindle speed of up to 15000rpm
Currently, the fastest hard drive would be a SSD (solid state drive) which relies on non-
volatile silicon memory chips arranged in arrays to store data. SSDs have nearly no read and
write latency and is capable of speeds of 200-300mbps compared to the standard sata which
is capable of 40-90mbps. A SSD however can cost 10times the price of a mechanical Hard
Disk and store a fraction of the data a mechanical hard disk is capable of. Currently the
record is 256GB for SSDs and 1TB for HDDs
The setup of a hard disk may include: stand alone master, slave with master present, RAID 0,
RAID 1 - 10
RAID Setup: the two most common configurations are RAID 0 and 1
RAID is the abbreviated term for: Redundant Array of Inexpensive Drives. Raid 0 uses the
concept of stripping to evenly split data between 2 or more drives. This allows the computer
to access multiple drives simultaneously to increase data transfer rate and response time. The
disadvantages of this setup is reliability. If one of the drives fail -- your data is gone.
RAID 1, is the setup in which 2 or more disks are used to create copies of each other assuring
no data loss if one or more drives fail in the array. Performance wise, there is no gain.
The most cost effective purchase would be for a 500GB internal HD in the SATA format
which will run about 120$. 25 cent per MB is the current sweat spot for purchase, anything
higher or lower will cost more per MB.
HDD Formatting
Modern HDDs, such as SAS[29] and SATA[30] drives, appear at their interfaces as a
contiguous set of logical blocks; typically 512 bytes long but the industry is in the process of
changing to 4,096 byte logical blocks.[31] The process of relating these logical blocks to their
physical location on the HDD is called low level formatting which is usually performed at the
factory and is not normally changed in the field.[32] High level formatting then writes the file
system structures into selected logical blocks to make the remaining logical blocks available
to the host OS and its applications
21
22
CD & DVD”S
CD-RW discs require a more sensitive laser optics. Also, CD-RWs cannot be read in some
CD-ROM drives built prior to 1997. This is why CD-ROM drives of the age must bear a
"MultiRead" certification to show compatibility. CD-RW discs need to be blanked before
reuse. Different blanking methods can be used, including "full" blanking in which the entire
surface of the disc is cleared, and "fast" blanking in which only meta-data areas are cleared:
PMA, TOC and pregap, comprising a few percent of the disc. Fast blanking is much quicker,
and is usually sufficient to allow rewriting the disc. Full blanking removes traces of the
former data, often for confidentiality. It may be possible to recover data from full-blanked
CD-RWs with specialty data recovery equipment[citation needed]; however, this is generally
not used except by government agencies due to cost.
CD-RW also have a shorter rewriting cycles life (ca. 1,000) compared to virtually all of the
previously exposed types storage of media (typically well above 10,000 or even 100,000),
something which however is less of a drawback considering that CD-RWs are usually written
and erased in their totality, and not with repeated small scale changes, so normally wear
leveling is not an issue.
Their ideal usage field is in the creation of test disks, temporary short or mid-term backups,
and in general, where an intermediate solution between online and offline storage schemes is
required.
WORK:
All CDs and DVDs work by virtue of marks on the disc that appear darker than the
background. These are detected by shining a laser on them and measuring the reflected light.
In the case of molded CDs or DVDs, such as those bought in music or video stores, these
marks are physical “pits” imprinted into the surface of the disc. In CD-Recordable (CD-R)
discs, a computer’s writing laser creates permanent marks in a layer of dye polymer in the
disc. CD-Rewritable (CD-RW) discs are produced in a similar fashion, except that the change
to the recording surface is reversible.
The key is a layer of phase-change material, an alloy composed of silver, indium, antimony
and tellurium. Unlike most solids, this alloy can exist in either of two solid states: crystalline
(with atoms closely packed in a rigid and organized array) or amorphous (with atoms in
random positions). The amorphous state reflects less light than the crystalline one does.
When heated with a laser to about 700 degrees Celsius, the alloy switches from the original
crystalline phase to the amorphous state, which then appears as a dark spot when the disc is
played back. These spots can be erased using the same laser (at a lower power) to heat the
material to a temperature of 200 degrees C or so; this process returns the alloy to its
crystalline state. Most CD-RW makers suggest that one disc can be overwritten up to 1,000
times and will last about 30 years.
23
Prior to the introduction of the CD-RW technology, a standard for magneto-optical
recordable and erasable CDs called CD-MO was introduced in 1988 and set in the Orange
Book, part 1, and was basically a CD with a magneto-optical recording layer. The CD-MO
standard also allowed for an optional non-erasable zone on the disk, which could be read by
normal CD-ROM reader units.
Data recording (and erasing) was achieved by heating the magneto-optical layer's material
(eg. DyFeCo or less often TbFeCo or GdFeCo) up to its Curie point thus erasing all previous
data and then using a magnetic field to write the new data, in a manner essentially identical to
Sony's MiniDisc and other magneto-optical formats. Reading of the discs relied on the Kerr
effect. This was also the first major flaw of this format: it could only be read in special drives
and was physically incompatible with non magneto-optical enabled drives, in a much more
radical way than the later CD-RWs.
24
UNIT-3
A peripheral is a device attached to a host computer but not part of it, and is more or less
dependent on the host. It expands the host's capabilities, but does not form part of the core
computer architecture.
Examples are computer printers, image scanners, tape drives, microphones, loudspeakers,
webcams, and digital cameras.
Storage
Input (a. input devices)
Out put devices (a. Output device b. Display device c. Graphical output device d.
Computer display)
STORAGE
Computer data storage, often called storage or memory, refers to computer components and
recording media that retain digital data used for computing for some interval of time.
Computer data storage provides one of the core functions of the modern computer, that of
information retention. It is one of the fundamental components of all modern computers, and
coupled with a central processing unit (CPU, a processor), implements the basic computer
model used since the 1940s.
The contemporary distinctions are helpful, because they are also fundamental to the
architecture of computers in general. The distinctions also reflect an important and significant
technical difference between memory and mass storage devices, which has been blurred by
the historical usage of the term storage. Nevertheless, this article uses the traditional
nomenclature.
25
Purpose of storage
Many different forms of storage, based on various natural phenomena, have been invented.
So far, no practical universal storage medium exists, and all forms of storage have some
drawbacks. Therefore a computer system usually contains several kinds of storage, each with
an individual purpose.
A digital computer represents data using the binary numeral system. Text, numbers, pictures,
audio, and nearly any other form of information can be converted into a string of bits, or
binary digits, each of which has a value of 1 or 0. The most common unit of storage is the
byte, equal to 8 bits. A piece of information can be handled by any computer whose storage
space is large enough to accommodate the binary representation of the piece of information,
or simply data. For example, using eight million bits, or about one megabyte, a typical
computer could store a short novel.
Traditionally the most important part of every computer is the central processing unit (CPU,
or simply a processor), because it actually operates on data, performs any calculations, and
controls all the other components.
Primary Memory
Primary storage (or main memory or internal memory), often referred to simply as memory,
is the only one directly accessible to the CPU. The CPU continuously reads instructions
stored there and executes them as required. Any data actively operated on is also stored there
in uniform manner.
Historically, early computers used delay lines, Williams tubes, or rotating magnetic drums as
primary storage. By 1954, those unreliable methods were mostly replaced by magnetic core
memory. Core memory remained dominant until the 1970s, when advances in integrated
circuit technology allowed semiconductor memory to become economically competitive.
This led to modern random-access memory (RAM). It is small-sized, light, but quite
expensive at the same time. (The particular types of RAM used for primary storage are also
volatile, i.e. they lose the information when not powered).
As shown in the diagram, traditionally there are two more sub-layers of the primary storage,
besides main large-capacity
26
RAM:
Secondary storage
Secondary storage (or external memory) differs from primary storage in that it is not directly
accessible by the CPU. The computer usually uses its input/output channels to access
secondary storage and transfers the desired data using intermediate area in primary storage.
Secondary storage does not lose the data when the device is powered down—it is non-
volatile. Per unit, it is typically also two orders of magnitude less expensive than primary
storage. Consequently, modern computer systems typically have two orders of magnitude
more secondary storage than primary storage and data is kept for a longer time there.
27
In modern computers, hard disk drives are usually used as secondary storage. The time taken
to access a given byte of information stored on a hard disk is typically a few thousandths of a
second, or milliseconds. By contrast, the time taken to access a given byte of information
stored in random access memory is measured in billionths of a second, or nanoseconds. This
illustrates the very significant access-time difference which distinguishes solid-state memory
from rotating magnetic storage devices: hard disks are typically about a million times slower
than memory. Rotating optical storage devices, such as CD and DVD drives, have even
longer access times. With disk drives, once the disk read/write head reaches the proper
placement and the data of interest rotates under it, subsequent data on the track are very fast
to access. As a result, in order to hide the initial seek time and rotational latency, data are
transferred to and from disks in large contiguous blocks.
The secondary storage is often formatted according to a file system format, which provides
the abstraction necessary to organize data into files and directories, providing also additional
information (called metadata) describing the owner of a certain file, the access time, the
access permissions, and other information.
INPUT DEVICES
An input device is any peripheral (piece of computer hardware equipment) used to provide
data and control signals to an information processing system (such as a computer). Input and
output devices make up the hardware interface between a computer as a scanner or 6DOF
controller.
Whether the input is direct or indirect. With direct input, the input space coincides with the
display space, i.e. pointing is done in the space where visual feedback or the cursor appears.
Touch screens and light pens involve direct input. Examples involving indirect input include
the mouse and trackball.
Whether the positional information is absolute (e.g. on a touch screen) or relative (e.g. with a
mouse that can be lifted and repositioned)
Output devices
An output device is any piece of computer hardware equipment used to communicate the
results of data processing carried out by an information processing system (such as a
computer) to the outside world.
28
In computing, input/output, or I/O, refers to the communication between an information
processing system (such as a computer), and the outside world. Inputs are the signals or data
sent to the system, and outputs are the signals or data sent by the system to the outside.
Conventional PCI (part of the PCI Local Bus standard and often shortened to PCI) is a
computer bus for attaching hardware devices in a computer. These devices can take either the
form of an integrated circuit fitted onto the motherboard itself, called a planar device in the
PCI specification, or an expansion card that fits into a slot. The name PCI is an initialism
formed from Peripheral Component Interconnect. The PCI Local Bus is common in modern
PCs, where it has displaced ISA and VESA Local Bus as the standard expansion bus, and it
also appears in many other computer types. Despite the availability of faster interfaces such
as PCI-X and PCI Express, conventional PCI remains a very common interface.
The PCI specification covers the physical size of the bus (including the size and spacing of
the circuit board edge electrical contacts), electrical characteristics, bus timing, and protocols.
The specification can be purchased from the PCI Special Interest Group (PCI-SIG).
Typical PCI cards used in PCs include: network cards, sound cards, modems, extra ports such
as USB or serial, TV tuner cards and disk controllers. Historically video cards were typically
PCI devices, but growing bandwidth requirements soon outgrew the capabilities of PCI. PCI
video cards remain available for supporting extra monitors and upgrading PCs that do not
have any AGP or PCI Express slots.[1]
Many devices traditionally provided on expansion cards are now commonly integrated onto
the motherboard itself, meaning that modern PCs often have no cards fitted. However, PCI is
still used for certain specialized cards, although many tasks traditionally performed by
expansion cards may now be performed equally well by USB devices.
PCI (Peripheral Component Interconnect) was immediately put to use in servers, replacing
MCA and EISA as the server expansion bus of choice. In mainstream PCs, PCI was slower to
replace VESA Local Bus (VLB), and did not gain significant market penetration until late
1994 in second-generation Pentium PCs. By 1996 VLB was all but extinct, and
manufacturers had adopted PCI even for 486 computers.[2] EISA continued to be used
alongside PCI through 2000. Apple Computer adopted PCI for professional Power Macintosh
computers (replacing NuBus) in mid-1995, and the consumer Performa product line
(replacing LC PDS) in mid-1996.
Later revisions of PCI added new features and performance improvements, including a 66
MHz 3.3 V standard and 133 MHz PCI-X, and the adaptation of PCI signaling to other form
factors. Both PCI-X 1.0b and PCI-X 2.0 are backward compatible with some PCI standards.
The PCI-SIG introduced the serial PCI Express in 2004. At the same time they rechristened
PCI as Conventional PCI. Since then, motherboard manufacturers have included
progressively fewer Conventional PCI slots in favor of the new standard.
29
EXPANSION SLOT
The expansion card (also expansion board, adapter card or accessory card) in computing is a
printed circuit board that can be inserted into an expansion slot of a computer motherboard to
add functionality to a computer system.
One edge of the expansion card holds the contacts (the edge connector) that fit exactly into
the slot. They establish the electrical contact between the electronics (mostly integrated
circuits) on the card and on the motherboard.
Connectors mounted on the bracket allow the connection of external devices to the card.
Depending on the form factor of the motherboard and case, around one to seven expansion
cards can be added to a computer system. In the case of a backplane system, up to 19
expansion cards can be installed. There are also other factors involved in expansion card
capacity. For example, most graphics cards on the market as of 2010 are dual slot graphics
cards, using the second slot as a place to put an active heat sink with a fan.
Some cards are "low-profile" cards, meaning that they are shorter than standard cards and
will fit in a lower height computer chassis. (There is a "low profile PCI card" standard[1] that
specifies a much smaller bracket and board area). The group of expansion cards that are used
for external connectivity, such as a network, SAN or modem card, are commonly referred to
as input/output cards (or I/O cards).
The primary purpose of an expansion card is to provide or expand on features not offered by
the motherboard. For example, the original IBM PC did not provide graphics or hard drive
capability as the technology for providing that on the motherboard did not exist. In that case,
a graphics expansion card and an ST-506 hard disk controller card provided graphics
capability and hard drive interface respectively.
In the case of expansion of on-board capability, a motherboard may provide a single serial
RS232 port or Ethernet port. An expansion card can be installed to offer multiple RS232
ports or multiple and higher bandwidth Ethernet ports. In this case, the motherboard provides
basic functionality but the expansion card offers additional or enhanced ports.
30
expansion slot shown in fig.,
31
RAM
An early type of widespread writable random-access memory was the magnetic core
memory, developed from 1949 to 1952, and subsequently used in most computers up until the
development of the static and dynamic integrated RAM circuits in the late 1960s and early
1970s. Before this, computers used relays, delay line/delay memory, or various kinds of
vacuum tube arrangements to implement "main" memory functions (i.e., hundreds or
thousands of bits), some of which were random access, some not. Latches built out of
vacuum tube triodes, and later, out of discrete transistors, were used for smaller and faster
memories such as random-access register banks and registers. Prior to the development of
integrated ROM circuits, permanent (or read-only) random-access memory was often
constructed using semiconductor diode matrices driven by address decoders.
RAM disks
Software can "partition" a portion of a computer's RAM, allowing it to act as a much faster
hard drive that is called a RAM disk. A RAM disk loses the stored data when the computer is
shut down, unless memory is arranged to have a standby battery power source.
Double data rate synchronous dynamic random access memory (DDR SDRAM) is a class of
memory integrated circuits used in computers.
Compared to single data rate (SDR) SDRAM, the DDR SDRAM interface makes higher
transfer rates possible by more strict control of the timing of the electrical data and clock
signals. Implementations often have to use schemes such as phase-locked loops and self-
calibration to reach the required timing accuracy.[1][2]The interface uses double pumping
(transferring data on both the rising and falling edges of the clock signal) to lower the clock
frequency. One advantage of keeping the clock frequency down is that it reduces the signal
integrity requirements on the circuit board connecting the memory to the controller. The
name "double data rate" refers to the fact that a DDR SDRAM with a certain clock frequency
achieves nearly twice the bandwidth of a single data rate (SDR) SDRAM running at the same
clock frequency, due to this double pumping.
With data being transferred 64 bits at a time, DDR SDRAM gives a transfer rate of (memory
bus clock rate) × 2 (for dual rate) × 64 (number of bits transferred) / 8 (number of bits/byte).
Thus, with a bus frequency of 100 MHz, DDR SDRAM gives a maximum transfer rate of
1600 MB/s.
"Beginning in 1996 and concluding in June 2000, JEDEC developed the DDR (Double Data
Rate) SDRAM specification (JESD79)."[3] JEDEC has set standards for data rates of DDR
SDRAM, divided into two parts. The first specification is for memory chips, and the second
is for memory modules. DDR SDRAM (sometimes referred to as DDR1 SDRAM) has been
superseded by DDR2 SDRAM and DDR3 SDRAM.
32
DDR SERIES
OLD SD RAM
33
MOTHER BOARD
A motherboard is the central printed circuit board (PCB) in many modern computers and
holds many of the crucial components of the system, while providing connectors for other
peripherals. The motherboard is sometimes alternatively known as the main board, system
board, or, on Apple computers, the logic board.[1] It is also sometimes casually shortened to
mobo
During the late 1980s and 1990s, it became economical to move an increasing number of
peripheral functions onto the motherboard (see below). In the late 1980s, motherboards began
to include single ICs (called Super I/O chips) capable of supporting a set of low-speed
peripherals: keyboard, mouse, floppy disk drive, serial ports, and parallel ports. As of the late
1990s, many personal computer motherboards supported a full range of audio, video, storage,
and networking functions without the need for any expansion cards at all; higher-end systems
for 3D gaming and computer graphics typically retained only the graphics card as a separate
component.
Most computer motherboards produced today are designed for IBM-compatible computers,
which currently account for around 90% of global PC sales
A motherboard, like a backplane, provides the electrical connections by which the other
components of the system communicate, but unlike a backplane, it also connects the central
processing unit and hosts other subsystems and devices.
A typical desktop computer has its microprocessor, main memory, and other essential
components connected to the motherboard. Other components such as external storage,
controllers for video display and sound, and peripheral devices may be attached to the
motherboard as plug-in cards or via cables, although in modern computers it is increasingly
common to integrate some of these peripherals into the motherboard itself.
34
Power connectors, which receive electrical power from the computer power supply and
distribute it to the CPU, chipset, main memory, and expansion cards
35
Additionally, nearly all motherboards include logic and connectors to support commonly
used input devices, such as PS/2 connectors for a mouse and keyboard. Early personal
computers such as the Apple II or IBM PC included only this minimal peripheral support on
the motherboard. Occasionally video interface hardware was also integrated into the
motherboard; for example, on the Apple II and rarely on IBM-compatible computers such as
the IBM PC Jr. Additional peripherals such as disk controllers and serial ports were provided
as expansion cards.
Given the high thermal design power of high-speed computer CPUs and components, modern
motherboards nearly always include heat sinks and mounting points for fans to dissipate
excess heat.
CPU SOCKETS
A CPU socket or slot is an electrical component that attaches to a printed circuit board (PCB)
and is designed to house a CPU (also called a microprocessor). It is a special type of
integrated circuit socket designed for very high pin counts. A CPU socket provides many
functions, including a physical structure to support the CPU, support for a heat sink,
facilitating replacement (as well as reducing cost), and most importantly, forming an
electrical interface both with the CPU and the PCB. CPU sockets can most often be found in
most desktop and server computers (laptops typically use surface mount CPUs), particularly
those based on the Intel x86 architecture on the motherboard. A CPU socket type and
motherboard chipset must support the CPU series and speed. Generally, with a newer AMD
microprocessor, you need only select a motherboard that supports the CPU and not be
concerned with the chipset.
INTERGRATED PERIPHERALS
With the steadily declining costs and size of integrated circuits, it is now possible to include
support for many peripherals on the motherboard. By combining many functions on one
PCB, the physical size and total cost of the system may be reduced; highly integrated
motherboards are thus especially popular in small form factor and budget computers.
For example, the ECS RS485M-M, a typical modern budget motherboard for computers
based on AMD processors, has on-board support for a very large range of peripherals
disk controllers for a floppy disk drive, up to 2 PATA drives, and up to 6 SATA
drives (including RAID 0/1 support)
integrated ATI Radeon graphics controller supporting 2D and 3D graphics, with VGA
and TV output
integrated sound card supporting 8-channel (7.1) audio and S/PDIF output
Fast Ethernet network controller for 10/100 Mbit networking
USB 2.0 controller supporting up to 12 USB ports
IrDA controller for infrared data communication (e.g. with an IrDA-enabled cellular
phone or printer)
temperature, voltage, and fan-speed sensors that allow software to monitor the health
of computer components
36
TEMPERATURE AND RELIABILITY
Motherboards are generally air cooled with heat sinks often mounted on larger chips, such as
the Northbridge, in modern motherboards. If the motherboard is not cooled properly, it can
cause the computer to crash. Passive cooling, or a single fan mounted on the power supply,
was sufficient for many desktop computer CPUs until the late 1990s; since then, most have
required CPU fans mounted on their heat sinks, due to rising clock speeds and power
consumption. Most motherboards have connectors for additional case fans as well. Newer
motherboards have integrated temperature sensors to detect motherboard and CPU
temperatures, and controllable fan connectors which the BIOS or operating system can use to
regulate fan speed. Some computers (which typically have high-performance
microprocessors, large amounts of RAM, and high-performance video cards) use a water-
cooling system instead of many fans.
BIOS
Motherboards contain some non-volatile memory to initialize the system and load an
operating system from some external peripheral device. Microcomputers such as the Apple II
and IBM PC used ROM chips, mounted in sockets on the motherboard. At power-up, the
central processor would load its program counter with the address of the boot ROM and start
executing ROM instructions, displaying system information on the screen and running
memory checks, which would in turn start loading memory from an external or peripheral
device (disk drive). If none is available, then the computer can perform tasks from other
memory stores or display an error message, depending on the model and design of the
computer and version of the BIOS.
Most modern motherboard designs use a BIOS, stored in an EEPROM chip soldered to the
motherboard, to bootstrap the motherboard. (Socketed BIOS chips are widely used, also.) By
booting the motherboard, the memory, circuitry, and peripherals are tested and configured.
This process is known as a computer Power-On Self Test (POST) and may include testing
some of the following devices:
floppy drive
network controller
CD-ROM drive
DVD-ROM drive
SCSI hard drive
IDE, EIDE, or SATA hard disk
External USB memory storage device
37
38
Block diagram
39
SMPS
Unlike a linear power supply, the pass transistor of a switching mode supply switches very
quickly (typically between 50 kHz and 1 MHz) between full-on and full-off states, which
minimizes wasted energy. Voltage regulation is provided by varying the ratio of on to off
time. In contrast, a linear power supply must dissipate the excess voltage to regulate the
output. This higher efficiency is the chief advantage of a switch-mode power supply.
Switching regulators are used as replacements for the linear regulators when higher
efficiency, smaller size or lighter weight are required. They are, however, more complicated,
their switching currents can cause electrical noise problems if not carefully suppressed, and
simple designs may have a poor power factor.
A modern computer power supply is a switch with on and off supply designed to convert
110-240 V AC power from the mains supply, to several output both positive (and historically
negative) DC voltages in the range + 12V,-12V,+5V,+5VBs and +3.3V. The first generation
of computers power supplies were linear devices, but as cost became a driving factor, and
weight became important, switched mode supplies are almost universal.
40
ATX shown in fig;
41
UNIT – 4
TROUBLESHOOTING
1. RAM PROBLEM:
System give 3 short beep’s like 1-1-1
No display means ( monitor cannot receive signal from mother board )
BLUE SCREEN ERROR
System alt ( hang )
Solve:
2. PROCESSOR PROBLEM:
Solve:
3. BIOS PROBLEM:
Solved:
42
4. NO DISPLAY PROBLEM:
OPERATING SYSTEM
INSTALLATION XP:
First check the bios booting device’s DVD & HARD DISK
Assign the 1, 2, 3 nd booting device’s.
Insert “WINDOWS XP” in to the DVD or CD Drive
In screen “PRESS ANY KEY TO CONTINUE. . . . .” Press any key in key board.
Enter to “XP INSTALLATION” will start.
Press “Enter” to Welcome screen
Press “F8 - I AGREE” To continue for license agreement.
IF new hard disk they to enter to the partition if no hard disk is old ask this question
“C:\WINDOWS\SYSTEM32\” press “ENTER TO CONTINUE”.
You will enter in partition option’s
After you create partition go to c:\ press ENTER
He ask 4 question’s Like this QUICK NTFS, QUICK FAT, NTFS, FAT, Leave changes
If put NTFS quick format.
After 39minute’s will start
Click NEXT
Click Next for Language and keyboard
Enter “PRODUCT ID” Click Next
Enter date and Time zone. Click Next
Enter Network option click Next
After complete the Setup in Welcome screen in Windows XP.
43
SOFTWARE INSTALLATION
Driver Installation
Insert Mother board CD in to DVD or CD Drive if some CD’s can auto play installation.
I old type installation is .inf installation.
Software are Three types
I - Package
II - Application
III - programming
IV - System software
Various types of computer software are used to simplify the operations and applications of
computer programs. Computer software enables the computer system to perform in
accordance with the given tasks. Computer software involves all types of software related
applications. Electronic media content is also explained by the computer software.
Different types of computer software are available in the global computer market.
Programming software, system software and application software are the three main types of
computer software used in computer networking. System software is the most commonly
used variety types of software. System software offers a protective shield to all software
applications. It also provides support to the physical components of computers. System
software coordinates all external devices of computer system like printer, keyboard, displays
etc.
Application software is used for commercial purpose. The application software is widely
used in educational, business and medical fields. Computer games are the most popular forms
of application software. Industrial automation, databases, business software and medical
software prove to be of great help in the respective fields. Educational software is widely
used in educational institutes across the globe.
Computer software works through computer programming. The whole process runs like a
chain reaction. Transfer of commands initiates the chain. Machine code generated by
computer software ends the entire process.
44
UNIT -5
INTRODUCTION TO NETWORKING
What is a Network?
A computer network is simply two or more computers connected together so they can
exchange information. A small network can be as simple as two computers linked together by
a single cable.
Most networks use hubs to connect computers together. A large network may connect
thousands of computers and other devices together.
45
46
Sharing Files and Drives
If your computers are connected to a network, each computer can make its resources
available to other computers in your office by sharing them over the network. Instead of
working in isolation as you do on a single computer not linked to a network, you can work
collectively, within a system that shares resources among a group of computer users.
Each computer on your network can share folders, entire disk drives, or a CD-ROM drive.
Then other computers on your network can access documents and other files stored in the
folders and on the drives. Instead of copying a document to a diskette and giving it to another
person to view, anyone can open and view the document using the network.
If you want to view the company's annual report stored on a co-worker's computer, you can
use the network to access the document on that computer. If you want to listen to music
stored on a computer in another room, you can use the network to access the music files.
Sharing a Printer
If you have a printer connected to your computer, you can share the printer with other
computers on the network. Then instead of buying a printer for every computer, all the
computers can print across the network to the printer. Suppose you want to print a document
on a color laser printer that is connected to another computer in the office. Instead of copying
your file to a disk, going to the other computer, and interrupting the person using that
computer, you can print directly over the network.
If you already have access to the Internet from one computer on your network, you can share
that Internet connection with other computers on the network. Then all the computers on your
network can browse the Web at the same time, using this single Internet connection.
Networking Components
To network computers together, you need to install networking hardware and software. Every
network includes these three components:
The computers that are connected together. Computers and similar devices are called
nodes when connected to a network.
The networking hardware that connects the computers together, including hardware
installed in your computer, network cables, and devices that connect all the cables together.
Networking software that runs on each computer and enables it to communicate with
other computers on the network.
47
Networking Hardware
Here is the networking hardware you need to set up a small network: Network adapter cards:
expansion cards that provide the physical connection between each computer and the
network. The card installs into a slot on your computer, just like a sound card or modem card.
Some newer computers have a network adapter already built into the system. Laptop
computers often use a card that slides into a PC card slot.
Network cables: special, unshielded twisted-pair (UTP) cables used to connect each
computer to the hub. The cable you need is Category 5 UTP cable with a square plastic RJ-45
connector on each end.
All the networking hardware described here is known as Ethernet. Ethernet is the industry-
wide standard for computer networks. Standard Ethernet networks transmit data at 10 million
bits per second (Mbps). A newer Ethernet standard, called Fast Ethernet, transmits data at
100 Mbps. Computer networks often contain a mixture of 10 Mbps and 100 Mbps devices.
Types of Networks
The type of network described in this book is a simple local network, often called a local area
network or LAN. A LAN connects computers together at one location.
Small Peer-to-Peer Networks
You can build a simple, small network without using the complex and expensive equipment
used in large networks. On such a network, often called apeer-to-peer network, each
computer can communicate with any other computer on the network.
48
You can connect computers together using network cables and a hub, or use wireless
technology to network the computers.
The focus of this book is on building a simple peer-to-peer network. Peer-to-peer networks
are easy to install and maintain, and they give you many of the advantages of a large network.
A peer-to-peer network is the obvious choice for a network in a home or small office. You
can set up this network yourself, without buying an expensive server, and without paying for
the services of a network administrator to install and manage the network.
Peer-to-peer networking has gained recent popularity on the Internet. Computers connected to
the Internet communicate directly with each other and share files. The software to set up a
local peer-to-peer network has been included in Windows since the release of Windows 95.
People have been building simple peer-to-peer networks since that time, using the software
built into Windows.
clients on the network access the servers to log on, access files, and print documents. The
servers may be running networking software from Novell or Microsoft, or they may be
running the UNIX or Linux"!" operating systems.
49
Large networks are usually constructed by connecting several small networks together with
special networking equipment that controls communication between the smaller segments of
subnetworks or Subnets
A ring network is a network topology in which each node connects to exactly two other
nodes, forming a single continuous pathway for signals through each node - a ring. Data
travels from node to node, with each node along the way handling every packet.
Advantages
Very orderly network where every device has access to the token and the opportunity to
transmit
Performs better than a star topology under heavy network load
Can create much larger network using Token Ring
Does not require network server to manage the connectivity between the computers
Disadvantages
One malfunctioning workstation or bad port in the MAU can create problems for the entire
network
Moves, adds and changes of devices can affect the network
Network adapter cards and MAU's are much more expensive than Ethernet cards and hubs
Much slower than an Ethernet network under normal load
REMOTE CONNECTION
50
UNIT-6
A computer virus is a computer program that can copy itself and infect a computer. The term
"virus" is also commonly but erroneously used to refer to other types of malware, including
but not limited to adware and spyware programs that do not have the reproductive ability. A
true virus can spread from one computer to another (in some form of executable code) when
its host is taken to the target computer; for instance because a user sent it over a network or
the Internet, or carried it on a removable medium such as a floppy disk, CD, DVD, or USB
drive.
Viruses can increase their chances of spreading to other computers by infecting files on a
network file system or a file system that is accessed by another computer.
As stated above, the term "computer virus" is sometimes used as a catch-all phrase to include
all types of malware, even those that do not have the reproductive ability. Malware includes
computer viruses, computer worms, Trojan horses, most rootkits, spyware, dishonest adware
and other malicious and unwanted software, including true viruses. Viruses are sometimes
confused with worms and Trojan horses, which are technically different. A worm can exploit
security vulnerabilities to spread itself automatically to other computers through networks,
while a Trojan horse is a program that appears harmless but hides malicious functions.
Worms and Trojan horses, like viruses, may harm a computer system's data or performance.
Some viruses and other malware have symptoms noticeable to the computer user, but many
are surreptitious or simply do nothing to call attention to themselves. Some viruses do
nothing beyond reproducing themselves
The first academic work on the theory of computer viruses (although the term "computer
virus" was not invented at that time) was done by John von Neumann in 1949 who held
lectures at the University of Illinois about the "Theory and Organization of Complicated
Automata". The work of von Neumann was later published as the "Theory of self-
reproducing automata".[5] In his essay von Neumann postulated that a computer program
could reproduce
In order to avoid detection by users, some viruses employ different kinds of deception. Some
old viruses, especially on the MS-DOS platform, make sure that the "last modified" date of a
host file stays the same when the file is infected by the virus. This approach does not fool
anti-virus software, however, especially those which maintain and date Cyclic redundancy
checks on file changes.
Some viruses can infect files without increasing their sizes or damaging the files. They
accomplish this by overwriting unused areas of executable files. These are called cavity
viruses. For example, the CIH virus, or Chernobyl Virus, infects Portable Executable files.
Because those files have many empty gaps, the virus, which was 1 KB in length, did not add
to the size of the file.
Some viruses try to avoid detection by killing the tasks associated with antivirus software
before it can detect them.
As computers and operating systems grow larger and more complex, old hiding techniques
need to be updated or replaced. Defending a computer against viruses may demand that a file
system migrate towards detailed and explicit permission for every kind of file access.
51
Anti-virus software and other preventive measures
Many users install anti-virus software that can detect and eliminate known viruses after the
computer downloads or runs the executable. There are two common methods that an anti-
virus software application uses to detect viruses. The first, and by far the most common
method of virus detection is using a list of virus signature definitions. This works by
examining the content of the computer's memory (its RAM, and boot sectors) and the files
stored on fixed or removable drives (hard drives, floppy drives), and comparing those files
against a database of known virus "signatures". The disadvantage of this detection method is
that users are only protected from viruses that pre-date their last virus definition update. The
second method is to use a heuristic algorithm to find viruses based on common behaviors.
This method has the ability to detect novel viruses that anti-virus security firms have yet to
create a signature for.
Virus removal
One possibility on Windows Me, Windows XP, Windows Vista and Windows 7 is a tool
known as System Restore, which restores the registry and critical system files to a previous
checkpoint. Often a virus will cause a system to hang, and a subsequent hard reboot will
render a system restore point from the same day corrupt. Restore points from previous days
should work provided the virus is not designed to corrupt the restore files or also exists in
previous restore points.[33] Some viruses, however, disable System Restore and other
important tools such as Task Manager and Command Prompt. An example of a virus that
does this is CiaDoor. However, many such viruses can be removed by rebooting the
computer, entering Windows safe mode, and then using system tools.
Administrators have the option to disable such tools from limited users for various reasons
(for example, to reduce potential damage from and the spread of viruses). A virus can modify
the registry to do the same even if the Administrator is controlling the computer; it blocks all
users including the administrator from accessing the tools. The message "Task Manager has
been disabled by your administrator" may be displayed, even to the administrator
Reinstalling the operating system is another approach to virus removal. It involves either
reformatting the computer's hard drive and installing the OS and all programs from original
media, or restoring the entire partition with a clean backup image. User data can be restored
by booting from a Live CD, or putting the hard drive into another computer and booting from
its operating system with great care not to infect the second computer by executing any
infected programs on the original drive; and once the system has been restored precautions
must be taken to avoid reinfection from a restored executable file.
These methods are simple to do, may be faster than disinfecting a computer, and are
guaranteed to remove any malware. If the operating system and programs must be reinstalled
from scratch, the time and effort to reinstall, reconfigure, and restore user preferences must
be taken into account. Restoring from an image is much faster, totally safe, and restores the
exact configuration to the state it was in when the image was made, with no further trouble.
52
53