Professional Documents
Culture Documents
Full Merged
Full Merged
Data is a set of numbers, text, pictures, videos etc. that have no set meaning or context
information is data that have been given meaning and context after it has been processed by
a computer or user.
Direct sources
Questionnaires-Questionaries are set of arranged questions that can be distributed to
people to get answers from them, they are easy to distribute, complete and collect as they
are user friendly.
interview-is a one-to-one meeting between the interviewee and the interviewer, in which
the interviewer would ask questions, to get direct opinions/facts from the interviewer, these
can be open ended and close ended questions, questions can be expanded on by the
interviewer and more answers can be gotten.
Observation- a method of gathering data by collectors by watching what happens in a given
situation
Data logging- using sensors and computers to collect data, which is analysed, saved and
output in graphs, charts, etc., most commonly used in scientific experiments, where human
intervention is not suitable.
Indirect sources
Electoral registers- Registers that hold record of citizens that are eligible to vote in an
election, they have citizen details such as name, address, age etc. which have some details
removed from the open version of the register
Third Party- a method of collecting data is from third party sources, there are companies
that collect an individual’s personal information and sell it to each other and to advertisers.
Business’s often ask customers if personal data could be shared with other organisations
after the customer use/buy their product
Quality of Information
Accuracy
If the data collected is inaccurate, the information after processing would be inaccurate,
therefore data should always be accurate so that it can be considered of good quality
Relevance
Data must be relevant to its purpose, the user must find data that is not required and
remove it, so that after processing only data required would be present
Age
Information must be up to date, as old information might be irrelevant as information tends
to change as time goes on. If data is not up to data it can lead to inaccuracy
Level of detail
Information should have the right amount of detail, it is possible for it to have to much detail
making it hard to extract information, if there is less information than required it can lead to
an inconvenience when using it.
Completeness
Information must be complete and have all required information to be of good quality, if not
it cannot be used properly for a particular purpose.
Encryption
Symmetric
A method of encryption that requires the use of the same private key to encrypt and decrypt
files. Both the Encryptor and recipient needs to have the key, therefore the key must be sent
with the file/or separately.
Asymmetric
A method of encryption that uses a public key to encrypt files and a private key to decrypt
files. The Encryptor encrypts using a public key that is available to everyone and sends the
files to the recipient who decrypts using a private key known only to him/her.
Encryption protocols
Purpose of SSL/TLS
Enable Encryption in order to protect data
Make Sure the exchanging of data is from legit individuals who they say they are
Ensure data is not altered or corrupted while being exchanged
Use of SSL/TLS
Used in Web server to client communication, to establish a secure connection between both
parties, by helping both parties to agree on rules of communication by carrying through a
handshake protocol.
Uses of Encryption
Hard Disks
When a file is created/written on a disk, it is automatically encrypted, when it is read, it is
automatically decrypted while leaving other files encrypted.
The whole disk is encrypted so that data is protected if the disk is stolen or left unattended
Keys need to be secured in a well available location as data can’t be recovered without the
key
Data can be permanently lost if the encrypted disk crashes or the OS gets corrupted.
Email
Email Encryption uses Asymmetric encryption, both the sender and Recipient exchange
messages between each other to add each other’s digital certificates, these are used to
encrypted files specific to each other to be decrypted.
Encrypting an email would encrypt any attachments
Emails are very susceptible to being intercepted by hackers, therefore encrypting all emails
are a good practice.
HTTPS
Hypertext transfer protocol is shown by a URL having HTTPS://, or a Padlock
A session key is encrypted using a public key which is sent to the webserver by a web
browser, then is decrypted using the servers private key, after which all exchange of
information is conducted using encryption using the session key.
HTTPS uses Asymmetric encryption initially to establish a secure session, then uses
symmetric encryption after that point forward.
After the session has ended, the symmetric key is disposed of.
Slower than HTTP, needs to be kept up to date by the host, but is more secure in data
transfer and Sites with HTTPS is Ranked Higher/Given more priority by Search engines.
Validation and Verification
Validation
A process in which data entered is reasonable and follows defined rules, if data
doesn’t conform to the rules, data is rejected and an error message is presented.
Verification
A process in which data entered is verified to be accurate and match the original
source.
Online Processing
Online Processing- processing is done in a computer that has direct communication with a
user. Data is processed almost immediately with a short delay, but isn’t that long for a user.
Each transaction is processed before taking the next transaction. Data is searched using
direct access.
Uses of Supercomputers
Drug research
Modern drug research uses complex models to see how changing the structure of
drugs effect the body
Numerous models have to be run to predict any side effects, which require high
computational power
Genetic Analysis
Finding genes that make humans susceptible to disease has always been complex,
supercomputers reduced the time taken to perform calculations with genes from
months to minutes
Quantum mechanics
Quantum mechanics is the study of matter and light in an atomic and subatomic
level, they require high computational power to process large number of
calculations with great accuracy
Weather forecasting
Data from sensors are weather stations and around the world are used to collect
numerous variables such as Humidity, temperature, rainfall etc. using past weather
data, the computer performs complex calculations that are large in number to
compare and predict the weather.
Climate research
Data collected for many decades which have trends of different variables such as
Pollutions, Ph, Temperature, Water level etc. are put in different complex
mathematical equations and collected to model and mimic the process of climate
change using supercomputers.
Compilers
A Program that translates high level language source code into an executable file called
object code
It produces a lot of error messages after it has translated the whole program
The compiled program will only run on the OS it has been originally compiled on
Interpreters
A Program that translates high level language source code into an intermediate form, line by
line which it executes after each line.
It stops translating when an error is encountered, and will not start till that error is fixed
The interpreted program can be transferred between operating systems because it is in the
form of source code, but it needs to be translated in each computer it is moved too.
Interpreter vs Compiler
Interpreters detect errors line by line, making debugging easier compared to compilers
which require the whole program to be translated to debug
Programmers can correct errors as they are found, while programmers can’t do anything
while a program is being compiled
Compiling uses more memory than interpreting as the whole program must be loaded for
translation while interpreters only take some lines saving memory
Compiled programs need to be recompiled in different operating systems which take a huge
time and chunks of memory while interpreted programs can be used on any OS with a
suitable interpreter as they are in its original source code.
The execution of an interpreted program is much slower as each statement has to be
translated before execution, where are compiled programs can be executed after being
compiled.
Source code must be available in interpreted programs while in compiled programs source
code is not present which can lead to software copyright infringements in interpreted
software.
Linkers
A linker is Used to combine different modules of object code into one executable code
program.
The programs modules can be written separately by different programmers than combined
saving time.
Linkers can be used to save Ram by compiling separately, but Variables might be different
between different modules leading to errors and documentations needs to be written and
read after completion which can be time taking
Device Drivers
A Device Driver is a small program that enables the OS and Software installed to
communicate with a hardware device.
Any Instructions sent by a software is understood by the device driver and then it tells the
hardware what it needs to do
If the appropriate driver is not installed, the computer is unable to send/get data from
hardware, and features might be lost as they might only be on that driver.
Drivers from one OS do not work with Other OS’s
Operating systems
An OS interacts with software, device drivers and hardware to manage a computer’s
resources
The OS Acts as an interface between the user and the computer as well as supplying
important utilities for managing the computer
Major Functions
1 allocates RAM for each program and make sure data doesn’t spread from
program to the other program’s memory allocation
2 manages data from input and output by communicating with the device
driver.
3 manages the storing and retrieving of files from storage devices
4 display error messages to the user if it requires human intervention
5 must handle the security of the system, by managing access rights for each
user and must verify all users when they are logging in.
6 it has to shut down the computer safely when it needs to be closed.
Utility software
is a type of system software that is needs to help maintain a computer system by
managing files and their associated storage devices.
Data Compression
Modifying data so that it occupies less storage on a disk.
Types of Compression
Lossless Compression – a type of file compression without the loss of a single bit of
data, due to the compression software removing repeated bits and replacing with an
indicator showing what is repeated and how many times. Mostly done in
documents, spreadsheets
Lossy Compression- a type of file compression where data bits are permanently
deleted that are unnecessary as they might be repeated again and again. Mostly
don’t in Pictures and videos where data might be same for many frames.
Formatting
Formatting is the process of erasing files that are present on new or existing storage devices
for another purpose.
Types of Formatting
Low level formatting
Usually done by manufacturers, it prepares the structure of the disk by
dividing the disk into cylinders and then dividing the tracks into sectors.
Data is erased permanently
If done repeatedly it would shorten the lifespan of the disk
High Level formatting
It deletes the Pointers that are in the file allocation table which are used to
tell the OS where to look for files.
The files are retrievable
It can help remove viruses
Backup
Backup software is a program that allows files to be copied and kept in a safe location so
that if data is lost due to crashing/corruption, the data can be replaced back
Types of Backups
Full Backup
Back-ups the whole drive.
Incremental Backup
Backups are done incrementally, where only files after a specific date and
time Is backed up.
Differential backup
Backups only data that has been changed since the full back-up
File-Copying
File Copying is creating a duplicate of an existing file, the copy will have the same content as
the original
File Copying software allows users to have more control over which files they are copied
Deleting
Delete utility software deletes the pointers that tell the OS where to find the files.
They can delete files permanently
Some files can’t be deleted normally as they are locked by the operating system, this can be
overcome using deleting utilities.
Antivirus software
A software that detects and removes malware
It will either remove the malware or will quarantine and then ask the user what they want to
do with it.
Methods of detection
Signature Based
The Virus signatures (sequence of bytes) are kept in a database, the
software checks all files for these sequences and when discovered the file is
deleted or quarantined
Only Capable of dealing with viruses that are known
Heuristic-based
Compares source code to that of known viruses and flags threats if a specific
percentage match
Can generate false positives
Behavioural based
Looks for abnormal Behaviour in files
Can generate false positives
Custom vs off the shelf Software
Custom software-Software that is specifically written for a particular client to meet the
organisations
Off the shelf software is software that is readily available in the market and it made for a
general purpose
Types of Interfaces
Command Line Interface
users communicate with the computer by typing commands into the interface
Graphical User Interface
Uses windows, icons, menus, and pointers (WIMP) to carry out command
Dialogue Interface
Users communicate with the device using their voice, speech recognition software is
used to convert spoken words that are recognised from a database of commands.
Gesture Based Interface
The system interprets human gestures and coverts them into commands by
searching through gestures stored in a database.
Sensors
Light
Used to measure the amount of Light
send a reading signal in the form of electrical energy
Ultraviolet
Used to measure the amount of ultraviolet
send a reading signal in the form of electrical energy
Temperature
Used to measure temperature
send a reading signal in the form of electrical energy
Pressure
Used to measure pressure/force on a given area
send a reading signal in the form of electrical energy
Sound
Used to measure sound in an area
send a reading signal in the form of electrical energy
Humidity/Moisture
Used to measure the humidity content of the air or moisture content of the soil
send a reading signal in the form of electrical energy
Infrared
Used to measure the infrared radiation from a heat source
send a reading signal in the form of electrical energy
Oxygen
Used to measure the level of oxygen
Carbon dioxide
Used to measure the level of Co2
Ph
Used to measure the level of acidity
Turbidity
Used to measure the cloudiness of water
Reed switch
Not a sensor but is used to measure rainfall
Electromagnetic field /induction loop
Used to measure the change in the earth’s magnetic field due to the presence of
ferromagnetic materials
Ultrasonic
Used to measure the distance of an object by sending out a sound wave, then
receiving it back due to reflection from the object.
Proximity
Used to detect the presence of an object by sending out sound waves, infrared or
magnetic fields and getting back its reflective wave
Touch
Used in detection devices to measure fluid levels between two conductors one of
which is liquid, separated by an insulator
Calibration
It is the process of testing and changing the device to make sure that readings taken are
accurate and correct
Importance of calibration
The accuracy of a sensor decreases due to constant use and exposure to the environment,
regular Calibration helps maintain accuracy
Types of Calibration
One Point Calibration
A Calibration done on sensors that give readings that have a constant offset from
the ideal values when measuring values. The sensor can be calibrated based on a
single reading to bring it to the ideal reference
Process of calibration
1 Take a reading with the sensor
2 Compare the reading with a device which is said to have the ideal reading
3 Subtract the readings to get the offset value
4 Add the offset value to every reading to get the calibrated value
Two Point Calibration
A calibration done on sensors that give readings that don’t have a constant offset
due to it suffering from an offset and difference in sensitivity. It uses two points in a
curve to calibrate
Process of calibration
1 Take two different readings with your sensors
2 Compare the readings with an ideal sensor
3 Calculate the range value of the sensor reading
4 Calculate the range value of ideal sensor
5 Use formula to calculate correct value
( Correct Value= ((CSv-SLv)*IRv)/SRv)+ILv)
Multipoint Calibration
A calibration similar to other calibration methods but is performed using many
different readings, at least 3.
Monitoring technologies
Uses of Monitoring Technologies
Weather Stations
Sensors used
Temperature Sensor to measure Ambient Temperature
Pressure sensor to measure air pressure
Humidity sensor to measure air humidity
Light sensor to measure the hours of sunlight
Reed switch to measure the rainfall by sending a signal each time a bucket is
tipped due to rainwater filling it
Readings/data from the sensor are fed into the computer and then stored in the
form of tables, spreadsheets etc. so that it can be later processed with calculations
to show information.
The readings from the sensors are converted from analogue to digital signals using
an ADC ( analogue to digital converter) so that the readings can be understood by
the computer
Monitoring water Pollution
Sensors used
Temperature sensor to measure the temperature of the water
PH sensor to measure the acidity of the water
O2 sensor to measure the amount of Oxygen in the water
CO2 sensor to measure the amount of carbon dioxide in the water
Turbidity sensor to measure the cloudiness of the water
Two Way of Comparing
Comparing readings to check if water is at expected levels
o Normal Readings would be deeper in a water body, so one sensor is
kept deeper than the other sensor
o Readings are compared to see if Water is at normal readings or not
Comparing readings affected by pollution (farms and industrial pollution)
o One sensor is kept upstream while the other is downstream after
the site that is said to cause pollution
o Both readings are compared to see if pollution is caused by the site
Both input data into the computer after passing through an ADC and this data is
stored, and also compared to bring in a conclusion.
Environmental Monitoring
How it is done.
Sound sensors to measure the sound pollution in a city
Monitory soil quality using PH sensors to measure acidity, moisture to
measure soil moisture content, temperature sensor to measure
temperature of soil
Temperature sensors can be used to measure temperature and warn people
about hight outside temperatures
Ultra violet sensors can be used to measure ultra violet levels and warn
governments to warn people to wear skin protection if in excessive amounts
Advantages and disadvantages of monitoring technologies
Pros
Readings taken tend to be more accurate
Can take readings anytime of the day or night
Readings are taken at regular intervals
Can Take multiple readings at a time
Graphs can be automatically drawn with great accuracy
Cons
Computers and sensors can be expensive to buy and maintain
Sensors need to calibrated as they loose accuracy after some periods of time
Computers can’t explain results while humans can
Control Technologies
Actuators
A device used in control systems to give an output to another device that can regulate a
variable.
Prevention of malware
Software
A firewall
to filter traffic and prevent malware from entering the system and the
network
Anti-Virus
To scan for Malicious code and remove the file and program
Anti-malware
It can guard against the entry of malware and can also remove malware that
has managed to get in
Physical
Develop anti-malware policies in companies
All employees should be educated about malware and its risks
Operating systems and software should be kept up to date, and unused software
and old versions should be removed
Strong passwords should be used
Users should always log off from their computers after using it
Emails should be read with suspicions and no attachments should be downloaded
without verifying the source of the email
Digital Divide
Digital divide is the gap between people who have access to Technology and can use ICT
skills, and people who have no access or have access but can’t use due to lack of ICT skills.
Inability of certain groups to access software and hardware
Backward Chaining
A Chaining System where the system tries to find a solution by repeatedly splitting the goal
into sub-goals
It explores the system to find a rule where the THEN part of the rule matches a required
goal, if the IF part is known to be true, then it is added to the lost if goals.
It is known as goal-driven
It starts with us having a goal to achieve or to prove it
Forward Chaining
A Chaining system where the system tries to find a solution by taking all data entered into
the system to determine which rules are selected and used.
It explored the system to find rules where the IF part of the rule is true, the THEN part of
that rule is used to add new facts, this is done until a solution is found.
It is known as data driven
It starts with us having data entered by the user, with no definitive goal.
Simulation Scenarios
Pilot Training
Military and Civilian aircrafts are very expensive to build, maintain and run for each
take off and landing, on top of this, crashing in real life can pose a risk to society and
the pilot
Simulations can help pilots get used to flying a plane by using a simulator
It has to recreate the sensation of flying and must be life like, with ways such as
components of the cockpit are made life like, movement of the plane is made using
hydraulics, etc.
Learning to drive a car
They replicate the driving experiences of a car
Safer for drivers to learn without causing damage to the car or other drivers
Can be used to monitor behaviour, performance and attention span of a driver
Disaster Planning
Simulations are carried out to plan for Natural disasters
Town planners can build structures to withstand disasters
Emergency services can also coordinate efficient responses
They can anticipate likelihoods and severity of natural disasters
Mathematicians have made equations for each disaster, these equations can be
input into models and simulated
Nuclear Science Research
Operators of Nuclear reactors can be trained using simulators which are made to
recreate a reactor control room
Computer simulations can run each scenario and prepare personnel and safety
measures for future reactors
Rearranging control rods and flow of coolant can be costly and pose safety risks so
simulations are done by research scientists using supercomputers
Database
Database Types
Flat-File
This Type of Databases store all the data in one Table with Redundant Data present
in it
Relational
This Type of Database stores data in separate linked tables so that data redundancy
is reduced
It has tables with Primary keys and foreign keys to link them together
Advantages and Disadvantages
Relational
Pros
o Saves Storage Spaces are duplicated data is not present
o When Data in one field changes, related linked fields automatically
change
o Editing Data is easy as less data must be entered
Cons
o Designing a relational Database takes time
o Setting up relationships would be much harder as the number of
Tables increase
o Setting up complex queries can be difficult
Flat-File
Pros
o Doesn’t need Planning as data isn’t organised
o Relationships aren’t required as one Table has all the data
o No Expertise is required when setting up a flat-file database
Cons
o Uses a lot of storage as data is duplicated
o Each Record has to be edited when one piece of data is changed
o Data is harder to edit as there are many fields needed to be gone
through
Relationships
One to one
A relationship where one record in one table is linked to only one record in another
table
The linked fields must contain the same data
One to Many
A relationship where one record in one table is linked to many records on another
table
The link uses a foreign key and Primary key, which uses referential integrity.
Many to Many
It is a conceptual relationship, where many fields from one table are linked with
many fields on another table
Many Database Software do many to many relationships using two one to many
relationships and a link table
Key Fields
Primary Key
Is a Field/Fields in a table that enables a record to be uniquely identified
Used in databases as indexes
Foreign key
It is a Field in a table, that have the same data as the primary key in another table
It is used to link the tables
Compound key
It is a primary key, that uses two or more fields to help a field be uniquely identified.
Referential Integrity
Referential integrity forces relationships to be consistent and avoids redundancy by forcing the data
in the foreign key to always have data in the primary key
Normalisation
The process of structuring data in a database into a formal method that is structured
correctly with no or less redundant data.
Normal Forms
Un-Normalised
It is a single Table that contains redundant data with non-atomic data
First Normalised
Rules for 1NF
Has Atomic Data
Data in Fields are not repeating
Has a unique key for each record, by primary or compound key
Second Normalised
non-key attributes are related to part of a compound key, those fields need to be
removed and put to a new table and linked back. [So, this normalisation only applied
to Tables with Compound keys]
Third Normalised
Non-key attributes that are related to another non-key attributes are removed and
linked back
Access Types
Sequential Access
An Access type where Records are found by going through all the records till the
record is found.
Indexed Sequential Access
An Access type where Records are indexed in a particular order, this index is used to
find the point on the disk to search for the Record, instead of going through from
the start.
Direct Access (Random Access)
Each Record has a unique Key, the computer uses this key to go to find the File in
the disk.