Challenge of Machine Learning

You might also like

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 2

Challenge

Data volume:
Deep learning and machine learning refers to a set of highly intensive computational models. One
typical example is fully connected multi-layer neural networks, where tons of network parameters need
to be estimated properly. The basis to achieve this goal is the availability of huge amount of data. In
fact, while there are no hard guidelines about the minimum number of training documents, a general
rule of thumb is to have at least about 10 the number of samples as parameters in the network. This is
also one of the reasons why deep learning is so successful in domains where huge amount of data can
be easily collected (e.g. computervision, speech, natural language). However, health care and is a
different domain; in fact, we only have approximately 7.5 billion people all over the world (as per
September 2016), with a great part not having access to primary health care. Consequently, we
cannot get as many patients as we want to train a comprehensive deep learning model expecially
Moreover, understanding diseases and their variability is much more complicated than other tasks,
such as image or speech recognition. Consequently, from a big data perspective, the amount of medical
data that is needed to train an effective and robust deep learning model would be much more
comparing with other media.

• Temporality: The diseases such as Pneumonia are always progressing and changing over time in a
non deterministic way which maybe cause complications . However, many existing deep learning
models, including those already proposed in the medical domain, assume static vector-based inputs,
which cannot handle the time factor in a natural way. Designing deep learning approaches that can
handle temporal health care data is an important aspect that will require the development of novel
solutions.For example,even the treatment some people with pneumonia, especially those in high-risk
groups, may experience complications that are hardly to be predicted such as ,bacteria which are cause
by the Pneumonia may spread from the lungs to other organs, potentially causing organ
failure,difficulty breathing ,fluid accumulation around the lungs,lung abscess and so on.

• Domain complexity: Different from other application domains (e.g.image and speech analysis), the
problems in biomedicine and health care are more complicated. The diseases which is Pneumonia are
highly heterogeneous and for most of the diseases there is still no complete knowledge on their causes
and how they progress. Moreover, the number of patients is usually limited in a practical clinical
scenario and we cannot ask for as many patients as we want.As we know that,many germs can cause
pneumonia. The most common are bacteria and viruses in the air we breathe.Some of the viruses that
cause colds and the flu can cause pneumonia. Viruses are the most common cause of pneumonia in
children younger than 5 years. Viral pneumonia is usually mild. But in some cases it can become very
serious. Coronavirus 2019 (COVID-19) may cause pneumonia, which can become severe.

•Medical Data Representation and Transformation :Deep learning algorithms can make the most
effective observations and predictions with the appropriate type and quantity of data. Currently, real-
world medical data are in unstructured formats such as sequences (time series, audio and video signals,
DNA, and so on), trees (XML documents, parse trees, RNA, etc), text data (symptoms description,
tumor description, medical records), or combinations of any of these formats . Unfortunately, the core
of deep learning and machine learning technique can only process numeric input data as eventually it is
broken down to strings of zeros and ones for computing system. Some qualitative data for example the
X-ray chest are not easily converted into a usable format and processing can sometimes become
complicated. Humans can easily process and make meaning of these data and when there is a
simultaneous change, for example, in intensity and quantity, it can easily be understood and adjustment
can be made with regard to the changes; an example is temperature and light. The representation of
similar processes and conditions in deep learning and requires a lot of encoding and thoughtful
mathematical expressions in data representation and transformation.

•Hardware Requirements for Medical Big Data


Deep learning and machine learning solution requires large training data to function effectively if the
study of the machine learning for pneumonia detection is no just stay in research level and fully utilize
in public heath sector. Usually real-world medical data are very large and are constantly increasing. To
implement tasks and create models, the computing machine needs to be equipped with sufficient
processing power. To handle such requirements, data scientists and engineers developed multicore high
performing GPUs and similar processing units as the regular central processing unit is impractical to
handle large-scale deep learning and machine learning tasks . These GPUs are expensive, consume lot
of power, and are not readily available for common use or in medical institutes and hospitals where the
data are captured and generated. The challenge is that industry level deep learning systems use high-
end data centers, which are not available in medical institutions, whereas, deployment after training is
done on smart devices, such as laptops, smart wearable devices, embedded computers, and other
mobile devices, which have small and inefficient processing units. The larger the deep learning
architecture, the bigger the computing requirement needed to accomplish training. Deploying deeping
learning solution to the real world thus becomes a costly and processor consuming situation.

Handling Biomedical Data Stream


Another challenge with deep learning is dealing with fast moving and streaming data if the study of the
machine learning for pneumonia detection is no just stay in research level and fully utilize in public
heath sector. There is a rapid change in the health care industry with huge volume of health care data
emanating at a rapid rate. The benefit of this is that the medical practitioners can leverage on these with
the support of deep learning model to diagnose and deliver health care services for different
pathological conditions. These data can be found in real-time biomedical signals from many sources,
including blood glucose monitoring, brain activity, blood pressure and oxygen saturation level,
biomedical imaging from ultrasound, electrography, MRI, in thousands of terabytes for insight into
medical conditions. Unstructured data format of useful patient records in the form of clinical text
contains useful patterns and genomic data describing relationship between various genetic markers,
disease conditions, and mutations. Physiological sensing data from and Electroencephalography are
important signals that are acquired from different parts of the body. It is important for deep learning to
be able to make meaning of large volumes of continuous input data that change with time and also take
into consideration when previous data become obsolete. In electrocardiography continuous activity
learning framework for streaming videos by intricately tying together deep hybrid feature models and
active learning was proposed. In another architecture, a streaming hardware accelerator was proposed
for incremental feature learning with denoising autoencoder . Although, some deep learning
architecture variants have tried to proffer techniques for working around this situation, there are
unresolved challenges regarding effective analyses of fast moving, large-scale streaming data in terms
of memory consumption, feature selection, missing data, and computational complexity.

You might also like