Download as pdf or txt
Download as pdf or txt
You are on page 1of 3

DATA MINING AND DATA WAREHOUSING

[As per Choice Based Credit System (CBCS) scheme]


(Effective from the academic year 2017 - 2018)
SEMESTER – VI

Subject Code : 17CS651 IA Marks :40


Number of Lecture Hours/Week : 3 Exam Marks :60
Total Number of Lecture Hours : 40 Exam Hours :03
Course outcome
CO1: Understands data mining problems and implement the data warehouse
CO2:Demonstrate the association rules for a given data pattern.
CO3: Discuss between classification and clustering solution..
Assignment
1. Explain with diagram, a three-tier data warehouse architecture. (10 Marks Jan 2019,July
2019) CO1
2. A) What is a Data warehouse and what are its four key feature? (4 Marks July 2019)
CO1
B) Differentiate between ODS and data warehouse. (06 Marks Jan 2019) CO1
3. Explain OLAP operations with examples. (10 Marks Jan 2019,July 2019,2018) CO1
4. With neat diagram, explain implementation of operational data Store. (10 Marks July
2018) CO1
5. a) Define:1) Dimensions’ 11) Measures 111) Fact tables. (03 Marks Jan 2019) CO1
b) Compare OLTP and OLAP Systems. (06 Marks July 2019,2018,jan 2018) CO1
6. Explain the following terms with examples: 1) Snow Flake schema ii) Fact Constellation
schema iii) Star schema (10 Marks july 2019)
7. A)What is metadata in data warehouse? What it contains? (03 Marks Jan 2019) CO1
B) Present five major characteristics from Codd's rule.(5 Marks jan 2018) CO1
8. Describe ROLAP, MOLAP, HOLAP. (10 Marks jan 2019) CO1
9. Explain indexing OLAP data: Bitmap index and join index with example. (10 Marks jan
2019) CO1
10. What is Data Mining? With a neat diagram, Explain the KDD process in Data Mining.
(10 Marks jul 2019) CO1
11. Define Data preprocessing. Mention the steps involved in it. Explain any 2 steps in
detail.(10 Marks June 2019) CO1
12. Describe the various types of attributes and data sets. (10 Marks June 2019) CO1
13. For the following vectors X and Y. Calculate cosine similarity where X = {3 2 0 5 0 0 0
2 0 0}, Y{1 0 0 0 0 0 0 1 0 2} (10 Marks June 2019) CO1
14. What is Frequent Itemset Generation? Explain Frequent Itemset Generation using Apriori
principle.(10 marks jan 2018) 08 Marks jan 2019, jul 2019) CO2
15. A) Explain the following terms with example : i) Rule generations; ii) Computational
complexity.(6 Marks Jul 2019) CO2
b) Generate frequent itemset for the given data with support - 50%.( 5 Marks jul 2019)
CO2
TID 100 200 300 400
items {1,3,4} {2,3,5} {1,2,3,5} {2,5}
16. For the following transaction data set : i) Construct an FP tree ii) Generate the list of
frequent itemset. Ordered by their corresponding suffixes. ( 9 Marks jul 2019) CO2
TID 1 2 3 4 5 6 7 8 9 10

Items {a,b} {b,c,d} {a,c,d,e} {a,d,e} {a,b,c} {a,b,c,d} {a} {a,b,c} {a,b,d} {b,c,e}

17. Briefly explain the candidate generation procedure 𝐹𝑘−1 × 𝐹𝑘−1 Merging strategy.(7
Marks jul 2019) CO2
18. Explain alternative methods for generating frequent itemsets. (08 Marks jan 2018) CO2
19. Explain frequent itemset generation in FP-growth algorithm. (10 Marks jul 2018) CO2
20. Explain how decision tree induction algorithm works. Give example. ( 8 Marks jul
2019) CO3
21. List and explain the different characteristics of decision tree induction. (8 Marks jul
2019) CO3
22. Describe the nearest neighbour classification techniques. (8 Marks jul 2019) CO3
23. With neat block diagram, explain general approach to solve classification problem.(08
Marks jan 2019) CO3
24. Explain how to build a decision tree using Hunt’s algorithm. (08 Marks jan 2019) CO3
25. Explain rule based classifiers with illustration. (08 Marks jan 2019) CO3
26. Explain K-nearest neighbor classification algorithm with example. (08 Marks jan 2019)
27. Explain how predictive accuracy of classification method is estimated.(10 marks Jan
2018)
28. What are Bayesian classifiers? State Baye’s theorem and explain how Baye’s theorem ts
used in the Naive Bayestan classifier with example. (10 Marks jul 2018, jan 2018) (8
Marks jul 2019) CO3
29. What is Cluster analysis? Describe the different types of clustering. (8 Marks jun
2019,jan 2019,jul 2018, jan 2018) CO3
30. Explain K - means clustering Algorithm with example .(12 Marks jun 2019,jul
2018,Jan 2018) CO3
31. Explain Graph based clustering algorithms with Example. .(12 Marks jun 2019,jul
2018,Jan 2018) CO3
32. Explain Divisive hierarchical clustering With example.(12 Marks jun 2019,jul
2018,Jan 2018) CO3
33. What are the basic approaches used for generating agglomerative hierarchical
clustering? (8 Marks jun 2019,jan 2019) CO3
34. Explain DBSCAN algorithm, with example. (8 Marks jun 2019,jan 2019) CO3
35. Briefly explain BIRCH scalable clustering algorithm. (08 Marks jan 2019) CO3
36. Explain Different types of Cluster Evaluation with examples. (10 Marks) CO3

You might also like