Welcome to Scribd!
Academic Documents
Professional Documents
Culture Documents
Hobbies & Crafts Documents
Personal Growth Documents
Machine Learning Technique Aktu Quantum PDF
AI-enhanced title
pixloy) fe plel oy pngla) Ry ple 278) 1 union of the regions R,, R, covers all the space, we have siptarde [ploy Leyes (275) P= los) | (play|s)- peng hen ponds. 2.8) 11 Thus the probability af error is minimized if isthe region of space winds Then Fe, becomes region where the reverse it 2-8L(CSIT-Sem- Regression & Bayesian Learning 42, Inaclassifiation task with M clases, yn unknown pattern, represented by the feature vector xis assigned to class if ph, |2)> Hajlov Jet ‘Guede. | Consider the Bayesian el Aistributed classes, where fier for the uniformly fa + relay ay] Posey) = ‘muuition ata elbebl © muuttion sults for some values for @ and b ‘Typical easosare presentedin the Fig. 28.1. Pel) aa = ae a wig 281. QueBTT | Detine Bayes classifier. Explain how classification is done by using Bayes classifier. Anawer 1. ABayes classifier iea simple probabilistic classifier based on applying Bayes theorem (from Bayesian statistics) with strong (Naivel independence assimptions‘Machine Learning Techniques 29L CT Sem 2 ANaive Baye clasier assumes that the presence (or ab prea ature of dass is unrelated to the presence fon fy other feature Depending on the precne nature of tho probity mode azine ined ey een in asuportaed oe eee 4 Inmany practia applications, parameter estination for mats ipe even pfantosn tong na can work with the Naive Bayes model without believing in Bs maa probability or using any Bayesian methods. oe “An aulvantage ofthe Nave Bayes classifiers that ire ammunt of taining data 1o estimate the parameters (aecas ‘arlances ofthe variables) necessary for cleetcation, 6 The perceptron bears a certain relationship to elas classifier known as the Bayes classifier, up a lamiel ptr enc bean ap ‘When the environment is Gaussian, the Bayes class Fhe th nv the Bayes classifier reduces toa In the Bayes lanier, or Bayes bypotheis texting proce minimize the average tisk, denoted by 8: For a te-daseprelem, ‘represented by classes Cy and Co, the average risk ie defined Be GA] PialCpterCaRfRearoae souk | Rete f RerOe where the various terms are defined as follows: , = Prior probability that the observation vector x ie drawn from subspace H, with =1,2,and P, +P, C, = Cost of deciding in favour of class C, represented by subypace H, sthen class Ci true, with fj 1,2 P, (iC,) Conditional probability density function of therandam vectarX Pig 26.2a)depict able digram representation ofthe Bayes eassiir ‘Tn important points in this Block dlagram are twofold ‘The data processing in designing the Bayes classifiers confined entirely tothe computation of the likelihood ratio nx by. Thiscomputation incompletely invariant to the valves assigned to the per probailities and involved inthe decision-making proces These quantities merely afect the values of the threshold x. From acomputational point of view, we find it more convenient to ‘work with logsrithan of the likelihood ratio rather than the Telibood ratio tee 01 (CstTson 8) inet tipeiaiandier vector . | Likelihood Asin x to class 6) apne, | he [|i = MEE es ean o #@ aiess ion, a Mpa eS | ratio FE“ comparator be 08 88> lag ae iets ane o mt ‘Fig. 29.1, Two equivalent implementations of the Bayes dassifir : Neh Likeond rat text (5) Lagabod ratio et ‘Gur Hoy] Discuss Bayes eassifier using some example in detail, awe ‘Bayes canter: Rafer Q.28, Page 2, Unit 2 Forexample: 1. Let Dhe etesning sto taresand the nscale. ach feature is reprosented by an timensional atribste vector ater sy ct) doping measurements made on the esture rom ‘atte cetpoctve fy Ayn 4 ‘Suppace that there ae clases, C, Cyr Cy Gama feta the Shsaier wil pede tha tong to the Sass vig he hhet foster probit, ndnned oo. That laser predetsthatX longs ln fan aly HC R0> pl ND for 157m fet ‘Toon we maximize AC, Tela Cfo wich, 0 emai read the mas posterior hypothe By Bape hearer, ic) qs = BEIGE es, only POC| C) PIC) nowd to be ‘hatimed. IF tho elas prior proebitis ae not known thes 38 Commonly assweed that the classes are equally Uikely 1 HC =p = aC, and therefore peX|C) rained Ober Pax|6)G) is maximized. ion data sets with many ateributes ville extromely expensive 1% To reduce computation in evsating 2X16) acs conditional independence is made 8. As pO) is constant for al clas the comptation fPXIC) tho asumption ofMe retngues canny ig Lat sat ameeey hx preume hat the vas ofthe atts women Tsar one another, vento cla abel ote a? ts, piX1C) Tlpealca Ploy Cy) *P UXq| Cale... x PGE, 1C,) ses Th pects [CPs Py Gon IG) are easy ting eran Ol ry ig sete fete codbed hehe the senken 4 tna vad An tempat pC) we om Tras isoprene mur of ayo ae eae thease didely [ono sae dase ib 1. haeotoerssaned igen ised ott etemdtshareu Gisasandtbon wilt ‘liter eviaden seeBn0dby, EB ef] cae oO Tas otha 1) tg ‘i Theres eneetocmpute the mean and the standard devin tthe ale af tribute yf tring a of clas C, Te Sales reused testinte 16) vit For example, let X = (35, Rs. 40,000) where A, and A, are the sets age and ince respctv Lat he lat abel Sebarecompater ‘Vid. The associated clase label for X’ wya-oorey = as abe fi Xe yd, bus compet Latssarpe hat apa ate eran na three it stecotimoonvsieedatribute Suppose thet from tho training set, we find that customer in. whe Acer are 012 urs, ntr words HE ee and thi clase, we have ye5B and = 12 Inari t pret Prodi the class abel of XC) pICis eval lass; The cant edt predict hat the cas label of is POX|CD PC)» ‘The predicted reat ca IC pC for 1s jm. 04, ee abe the ean, for which X16) MC) # 2-121 (CSITSom5) ah a Regression & Bayesian Learning FHT] Let blue, green, and red be three clamses of objects with peor probabicon given by Phi) = 4 green) «1, eed) = Vt erative three types of objects pencils, pens, ane paper Lat the irre J er gual pbs ot tese objects given follow Use Bersmcen) 8 Pepenireen) an rune 6 Bees Ptpenircen= 73 Prey Pigenbivey= Wo Pepaperfb) = 13 Pee us, Plpeneed) iS) Papaperred)= 1 Tver | AeperBayes rule: igen ree) aren) Mereentenci = Epecil area) Rr) + PpendiZ Be) Pitoe + Pipencl ed) Pied) 1a 1 ite a" 2*a"a*e a Pipenciv be) Fae) ‘Pibtueipenell = “Pipencl/ green) P(green) + Plpencil bi Pius) + Pipa ed Pee) ay 244 wos oat = 0.5050 ed) Pleed) Pheaa rod Prod) + Pipencii/Bive) Pius) + pon grees) green) at ad 6X4 wo, = “tan ~o5 “8 ‘Since, Pgreen/penll has the highest value therefore pencil belongs to lass green. Pipeu 1) Pigrees) Pigroon/pen) = pircay green) Pigreen) + Ppen/ blue) Piblue) + pen red) Pred)aca Kerrie Boe PISL(CSAT Sema y 14 L(CSIT-SEmD) Regression & Bayesian Learning Pipe/ boo) Se pala ret + Roa 7 Neer a | MBN sare dais ment emmen gen ivr edad we Nae eeTcae AS sence Cth ot wh to predic sha pba © ae oe arb estislowee aon Pe ttt Tell are een ume he etre ‘Pipen/ green) P(green) + Pipen’ blue) 7 ri Pu) ge ol Pod) . 4,4 2 ica a og75 ~ o376 ~ 0%? § os Sine Mephesto, pn gen aol Bon igreenpaper) = -——_Pibaper/ green) Pigresn)___ 7 psp oe an papa) fos secanaee Pius Pge Pon Bos raters Prbtutpapee = Since, Predipaper asthe highest 409 Since. highest valu therefore, paper belongs to BEER] expinin Naive Bayes class paper! be) Pie) green) Pigreen) + Pipapee/ Blac) ue) + paper! red) Pir) ___ sper re Pei) Fipaper/gren) Pron’ « pope Has) Pibtae)« Pipapen’ ed Pied) “ym oo ig. 2.231. TheJaring carve for Nsw Bayes learn, 4, Atcoming Boolean variables the parameters are ‘O= AC = tre, 0, = PUX,= true} = true, og" PAX =trve | C= Fale) 15. Naive Bayes models can be viewed as Bayesian networks in which each [Xchas Ca he ole part and Clas no parents 5. ANaive Bayer model with gaussian POX, |C) inequivalent to amixture ‘fgauesiane with diagonal eovariance matrices, 1. While isture ofgaussans are used or dnsty estimation in continous tomains, Naive Beyos models wed in dacrte and mixed domains 8. Naive Bayes modes allow fr very ofcient inference of marginal and conditional distributions 8. Naive Bayes learning has no dificlty with noiy data and can give ore appopriata probable predtions, GaeRAT, | Consider a tworsiats (Tasty or non-Tasty) problem with the following training data, Use Naive Bayes classifier to classify the pattorn = "Cook = Asa, Health-Status = Bad, Culsine = Continental”.Machine Learning Techniques 301 (00en gression & Bayesinn Learning Cook | Health Status | Cuisine ‘asa Bed | tadian 2.6 aoa ‘a Good Continental eo 0N Sa = i Linstead tno = 0x 24344 20 Sia (Good | Indian rarer Usha Bad Indian - “Therefore the pritonis tasty, Usha Bd Continental { SE] Pxptain EM algorithm with steps Sie Bed “Continental | No lhe _ Si Good Continental Yoo Anawer Usha Good Tian Yea] 1 The Egcatin eintion seri ent — 1 Bat ream ietond estimate for nde parameters when the Ushe (oe Continental No Eetaincompleteor has misang dt posto some variate 2 EMchogees random value for the ising data pints ad eatin aaa prow act of daa 5, These new vals are then recursively used to estimate a eltr ret ye | wo Yes |e Lew Tne] ‘+ These are the two asc ep othe FM agri mn} 2 fo fae [2 [a [indin [a [a] * Mtmeeniey’ seta [s]ow [et Jemamal spo] * Rabntybanlecere maa a — it |i ‘Thén for those given parameter values, estimate the value of vente £18 |i. Tnitjslize the mean 1, eeceea yal as Heat ea ‘ho mixing velit sitet ‘by random values, (or other values) re [ve Fis[oe] Comparten lesa ‘Asha [26 [0 | Bed | 2/6 [8/8 [indian [ale | 1/4 ii Again estimate all the parameters using the current values Sean far| Gm as female poe]. Gonpali etin See EE 2 'v. Put some convergence criterion. Ey 1 PeEeimiaramecmensiomne ie F (orifall the parameters converge to some values) then stoP, wo] a0Machine Leaening Techniques QaeHAE | Describe the usage, advantages and disadvantages o EM algorithm. “Anewer Usage of EM algorithm 1. Tteanbeuced te il the missing data ina sample. 2 Itcanbo used asthe bass of unsupervised learning of clusters 3. can be used forthe purpose of ectimating the parameters of Hidden ‘Markov Model (MD. 4 Teean be used for discovering the value of latent variables. Advantages of EM algorithm are: 2. Teisalwaye guaranteed that likelihood willinrease with ench iterate, 2 TheE-step and Motep are often pretty eagy fr many problems in terms of implementation. 3, Solutions tothe M-steps often exit in the clased form, Disadvantages of EM algorithm are: 1. Tehae slow convergence 2 Temakes convergence to the acl optima only. 3. It requires both the probablitics, forward and backward (numerical ‘optimization requires only forward probability), ‘QueAG.| Write a short note on Bayestan network. or Explain Bayesian networkby takingan example. How isthe Bayesian network powerful representation for uncertainty knowledge ? aa ‘A Hayesian network is directed acyclic graph in which esch node i ‘notated with quantitative probability information, The specifcationis as fellows: Ast of random variables makes up the nodes ofthe network Variables may be discrete or continuous A mt of directed links or arrows connect pirsof nodes, there is om from to node i eaid Wo be parent oly Each node o, han & conditional probability distribution ‘Ps |parents tba quantifies the effect of parvatson the sade The graph hax na directed eyees and ene ina directed cycle DAG 217 L CSITSem, 'SIT-Sem-5) CSHT-Sems) 2-181 (Cs Regression & Bayesian Learning ‘A Bayesian network provides complote description of the domain, ‘very entry i the full ont probability dstihtion canbe called ftom the information in the network 4 Bayesian networks provide a concise way to xepresent conditional {ndapondence relationships nthe domain [A Bayesian network soften exponentially smaller than tefl oint ‘strut, For example 1. Suppose we wantto determin the posbiliy of ras geting wet or dey digo tothe oearrence of diffrent seasons. 2. ‘Tho weather has three states: Sunny, Cloudy, an Reiny. There are two poesblities forthe grass: Weto Dey. ‘8. Thesprinkler canbe on or off. Iie rainy, the grat gets wet but ifitis funy, we ean rake grass wet by pouring Water from a sprinkler. “4. Suppose that the gras i wet. This ral be contributed by one ofthe {wo Toscana: Fry itis reining Secondly, the sprinklers are turned 5. Using the Baye's rule, we can deduce the most contributing factor towards the wot grass a Conon Sprinkder Baia Wet go “Fig 2463. Bayesian network possesses the following merits in uncertainty knowledge representation. 1L- Bayesian network can conveniently handle incomplete data. 2 Bayesian network can learn the essual relation of variales. In dat nalyes, astal relation ishelpfl fr field knowledge understanding, t ‘analeo easily lead to precise prediction even under much interference 3. ‘The combination of bayesian network and bayesian statistics can take full advantage of fold knowledge and information from data, 4. ‘Thecombination of bayesian network and other models can effectively void averting problem.