Assignment 2 Emerging

You might also like

Download as pdf or txt
Download as pdf or txt
You are on page 1of 3

GRANBY COLLEGES OF SCIENCE & TECHNOLOGY

Naic, Cavite, Philippines


Tel: (046) 412-0437
Email: granby_philippines@yahoo.com.ph

NAME: NIKKOLAI ALDRICH V. GONZALES DATE: FEBRUARY 21, 2023

Assignment No. 2

1. What is Data Science?

Data science is the study of data with the goal of gaining important business insights.
It is a multidisciplinary method for analyzing massive volumes of data that integrates ideas
and techniques from the domains of mathematics, statistics, artificial intelligence, and
computer engineering. Data scientists can ask and receive answers to questions like what
occurred, why it occurred, what will occur, and what can be done with the outcomes thanks
to this study.

2. Explain the Data Processing Cycle

As an establishment or organization they cannot benefit from data in its raw form.
Data processing is the process of taking raw data and turning it into information that can be
used. An organization's team of data scientists and data engineers often performs it in a step-
by-step manner. The unprocessed data is gathered, sorted, processed, examined, and stored
before being provided in a legible way.

There are 6 main steps in Data Processing Cycle; Collection, Preparation, Input, Data
processing, Output. The first step is Collection, collection of raw data is the first step of the
data processing cycle. The type of raw data collected has a huge impact on the output
produced. Second is preparation or data cleaning is the process of sorting and filtering the
raw data to remove unnecessary and inaccurate data. Raw data is checked for errors,
duplication, miscalculations or missing data, and transformed into a suitable form for further
analysis and processing. This is done to ensure that only the highest quality data is fed into
the processing unit. The third process is input where raw data is converted into machine
readable form and fed into the processing unit. This can be in the form of data entry through
a keyboard, scanner or any other input source. Data Processing, in this step, the raw data is
GRANBY COLLEGES OF SCIENCE & TECHNOLOGY
Naic, Cavite, Philippines
Tel: (046) 412-0437
Email: granby_philippines@yahoo.com.ph

subjected to various data processing methods using machine learning and artificial
intelligence algorithms to generate a desirable output. Output is the next process where the
data is finally transmitted and displayed to the user in a readable form like graphs, tables,
vector files, audio, video, documents, etc. This output can be stored and further processed in
the next data processing cycle. The last step of the data processing cycle is storage, where
data and metadata are stored for further use. This allows for quick access and retrieval of
information whenever needed, and also allows it to be used as input in the next data
processing cycle directly.

3. Explain the Data Value Chain

The procedure of transforming unprocessed data into something valuable is referred


to as the "data value chain." In the end, businesses employ data value chains to unearth
massive amounts of information dispersed throughout their operations and make it
accessible and helpful to the parts of the business that need this intelligence.
The process of creating and using data, from initially determining a need for data to its eventual
usage and potential redistribution, is described by the data value chain.
The four main steps of the data value chain are Collection and grading, Access,
sharing and retention, Analytics and solutions.
Collection and grading are activities related to data collection and grading can take
place anywhere in an organization. Records, web clicks, purchase orders, financial
transactions, inventories, logistics and data collected by Internet of Things (IoT) devices are
routinely collected through daily operations and activities.
Access, sharing and retention, in this segment of the data value chain is needed to make
data accessible and usable. That includes the interface or platform that connects datasets with
end users such as data scientists, processes and protocols for how data will be accessed and
stored, and decisions on the importance of data sharing with third parties. If data sharing with
third parties is important, the compatibility or interoperability of platforms can become an
issue. Central to interoperability is the choice of an Application Programming Interface (API) to
allow for data transmission, use, access management and tracking.
GRANBY COLLEGES OF SCIENCE & TECHNOLOGY
Naic, Cavite, Philippines
Tel: (046) 412-0437
Email: granby_philippines@yahoo.com.ph

In Analytics and solutions, as more data is made available and accessible, it is expected that
a growing number of AI consulting firms providing specialized insights to individual economic
sectors or functions will enter the marketplace. AI and data analytics consulting firms will be
positioned to use datasets from multiple sources to create, new algorithms, teach artificial
intelligence algorithms and generate insights to clients, embed algorithms in existing functions
or processes, and use ongoing access to streaming data to offer new possibilities for machine
learning and automated decision systems to operate.

4. What is Big Data?

Big Data is a body of information that is enormous in volume and is always expanding
exponentially. No typical data management systems can effectively store or process this
data because of its magnitude and complexity. Big data is a type of data that is extremely
large.

You might also like